Back to Blog

Essay / 2026-04-30

Why Small Open-Source Models Often Win in Business

Most companies reach for the biggest, most capable model they can pay for. For real business workflows, that's usually the wrong instinct.

When companies start working with AI, the default instinct is to reach for the biggest, most capable model they can afford. The frontier model from OpenAI, or Anthropic, or Google. The thing they read about in the news.

For most business workflows, that's the wrong instinct, and it costs you on three different fronts at once.

This is a follow-up to the flagship private AI piece. The argument there was that responsible AI for companies means controlling the model and the data. The question this post answers is: once you've accepted that, which model should you actually pick?

The short answer for most business cases is: a small, open-source one.

What businesses actually need from a model

Most casual ChatGPT use is open-ended. People ask it about random things in their life, write essays, have free-form conversations. For that, you genuinely do need a huge model with a lot of general world knowledge.

Business automations are almost never like this.

In business, you usually want AI to:

  • follow clear instructions
  • look at a piece of data (an email, a document, a database row)
  • produce a specific, structured output (a category, an extraction, a draft, a routing decision)

This is what's called instruction following. It is not "have a conversation about anything"; it is "do this specific thing reliably, every time, on data I give you". Classification, extraction, routing, drafting against a template, summarising against a known corpus, parsing messy inputs into clean ones. That kind of work.

Modern small open-source models (Google's smaller Gemma family, Qwen's smaller variants, and others) are excellent at instruction following, and that's most of what businesses actually need.

The three boring advantages

Cost

Smaller means cheaper to run, sometimes by an order of magnitude. At the volumes a real business workflow generates (thousands of calls a day, sometimes much more), the difference between a small open-source model and a frontier API is the difference between a line item that doesn't need a meeting and one that does.

Speed

Fewer parameters means fewer calculations per response, so the output comes back much faster. For interactive tools (anything where a person is waiting on the answer) latency is part of the user experience, and the small-model gap there is large.

Hardware

Smaller models run on hardware that's actually available without a procurement project. A small model can sit on a single GPU you rent in a private cloud, or even on a workstation under someone's desk. Big frontier models cannot, which immediately forces you back into renting them as a service from somebody else, with all the loss of control that comes with that.

These three benefits compound. Cheaper plus faster plus easier-to-host means you can deploy more workflows, more places, with less ceremony.

The quieter benefit: small models tend to hallucinate less

This is the one most people don't expect, but the intuition is straightforward.

A huge model carries a vast amount of internal world knowledge that it can confabulate from when it doesn't actually know the answer. The phrase "halfway plausible nonsense" exists for a reason: a frontier model with no grounding will often invent something convincing rather than say it doesn't know. A small model has less of that internal knowledge to draw on, so it is more inclined to follow what's actually in front of it and less inclined to invent.

For business workflows, this is exactly the behaviour you want, because you are giving the model the context it needs anyway: the document to extract from, the policy to apply, the data to classify. You don't need it to know things from its training data; you need it to read carefully and follow instructions. Small models do this more obediently, with less risk of confidently making things up.

When the big model is still the right answer

Sometimes you do need the big one. Truly open-ended reasoning, very long context, broad world knowledge, agentic work that has to flexibly handle whatever comes. For those cases the gap is real and the frontier still matters.

But "most of the time you need the frontier" is the kind of folk wisdom that mostly survives because nobody questions it. For instruction-following business workflows, which are the bulk of what real automations actually do, the smaller model is often the better choice on every dimension.

What this changes about how you should choose a model

The biggest shift, once you take this seriously, is that the model becomes a deployment question rather than a vendor question.

"Which open-source model fits this workflow, and where do we run it?" replaces "which provider do we sign with?"

That changes a lot of downstream things. Cost, latency, and reliability become engineering decisions you can actually act on, rather than line items you negotiate with someone else's account manager. "Will this still work in six months" is an answer you control rather than a hope. Vendor risk doesn't disappear (you still have hardware, hosting, and dependencies), but the catastrophic single point of failure (one provider, one model, one set of terms) is gone.

The trade-off is that someone has to know how to deploy and run these models. That's the practical work, and it's covered separately in the deployment piece.

The point I want to leave you with is the upstream choice. If you've accepted that you want control over the model, smaller open-source models often make that easier, cheaper, and more reliable than the conventional wisdom assumes. The biggest model on the leaderboard is rarely the right answer for the workflow in front of you.