Skip to the content.

Open source

One of the two large strategic camps in the contemporary AI model ecosystem; the term as it is used in 2026 carries some specific gotchas that the older free-software vocabulary did not.


In one sentence

An open-source AI model is one whose weights — the trained parameter values — are published publicly so that anyone can download them, run them on their own hardware, modify them, and (usually) redistribute them; the term has accumulated some controversy because not every model labeled “open” actually meets the older free-software standards.

Why this term exists

The original “open source” movement, going back to the late 1990s, was about source code — the human-readable instructions that compile into running software. The user got the recipe, could read it, modify it, and rebuild the program from scratch.

Modern AI models break that model in interesting ways. The “source” of a language model is, arguably:

  1. The training-time code (the algorithm that does the training)
  2. The training data (the trillions of tokens fed in)
  3. The trained weights (the resulting parameter blob)

Releasing the weights is the thing most labs mean when they say “open source.” That is genuinely useful — you can run the model, modify it via fine-tuning, and use it commercially (often). But the training data is almost never released, and the training code is sometimes withheld. So the recipe is incomplete.

This has produced an ongoing argument in the AI community about whether models like Llama, Gemma, and Qwen are really open source, or just open weights. The older free-software purists say the latter is a marketing-friendly relabel that misses the point. The pragmatists argue that open weights provide most of the practical benefits to most users.

What it actually does — concretely

When a model is published as open-weights, you typically get:

What you do not always get:

That asymmetry is the modern open-source AI debate in one paragraph.

Working example

Three open-weights models are the workhorses of the local-AI ecosystem in mid-2026:

All three live on this MacBook’s hard drive and can run without sending a single token to a cloud provider. None of the three released their full training data.

Why this matters in a teaching context

For a BBA or MBA classroom, the open-source AI conversation is genuinely strategic. Three angles worth surfacing:

  1. Vendor independence. An organization that builds critical workflows around a closed-source frontier model is at the mercy of price changes, terms-of-service updates, and API deprecations. Open weights are a hedge — you can always run the version you have, on hardware you control, indefinitely.

  2. The Llama bet. Meta’s strategy of releasing capable models open-weights is one of the clearest pieces of competitive AI positioning in the past three years. Worth working through with students: why would Meta give away models that cost hundreds of millions of dollars to train? (Hint: commoditize your complement.)

  3. The compliance puzzle. Many regulated industries — healthcare, finance, defense — cannot use cloud-hosted closed models with sensitive data. Open-weights models on local hardware are the only viable path. This drives substantial enterprise adoption that does not show up in the trade press.

Trade-offs


Related entries: Closed source, Ollama, Parameters.

Return to Dictionary All Entries (A–Z)