The open source AI landscape in 2026
Open source AI sounds simple until you look closely. A practical survey of the major models, the licensing reality, the geopolitics, and what it all means if you are trying to make real decisions.

Everyone agrees open source AI is important. Nobody agrees on what it means
"Open source" is one of those terms that sounds settled until you start asking questions. In software, it has a clear definition backed by decades of legal precedent. In AI, it's a marketing claim that ranges from genuinely open to "we published the weights and wrote a blog post."
This matters because the decision to use an open model isn't just a technical one. It's a legal, operational, and sometimes geopolitical one. And the coverage tends to collapse all of that into a binary: open good, closed bad. Or worse: open cheap, closed expensive.
Neither framing is useful if you're actually trying to decide whether to build on Llama or call the Claude API.
This is the companion piece to our commercial AI landscape survey. Same approach: no hype, no jargon tourism, no "top 10 open source models that will DISRUPT your workflow." A clear look at who the players are, what "open" actually means in each case, and what it means for companies making real decisions with real budgets.
What "open" actually means (and doesn't)
In October 2024, the Open Source Initiative published its first formal definition of open source AI. The requirements: you must be able to use, study, modify, and share the system. That means releasing the source code used for training and inference, full documentation of how data was processed, model weights, and architectural documentation.
Here's the part that sparked the debate: the definition does not require releasing the training data itself. Only specifications of how it was gathered and filtered. Critics argue this makes the definition toothless. Supporters argue that requiring full data release would make open source AI practically impossible due to licensing, privacy, and scale constraints.
By the OSI's own standard, very few models actually qualify. Pythia from EleutherAI, OLMo from AI2, Google's T5. The models most people call "open source" are technically something else.
Open weight vs. open source
The distinction that matters most for enterprise decisions:
Open weight means the model publisher releases the trained weights so you can run the model yourself. You get the end product but not the recipe. You can't reproduce the training run, meaningfully audit the model's biases, or understand why it behaves the way it does in edge cases. Meta's Llama falls here. So does DeepSeek.
Open source (by the OSI definition) means you get weights, training code, data specifications, and documentation. You can reproduce, audit, and modify the full pipeline. Mistral's models ship under Apache 2.0 with training code. OpenAI's GPT-OSS does the same. Alibaba's Qwen family uses Apache 2.0.
Why does this matter for your company? Three reasons.
First, licensing. Llama's license prohibits commercial use by companies with more than 700 million monthly active users without Meta's permission. It also prohibits using Llama outputs to train competing models. If you're a mid-market company, the MAU cap probably doesn't affect you today. But building core infrastructure on a license with conditions you don't control is a decision worth making consciously.
Second, reproducibility. If you can't retrain or meaningfully modify a model, you're dependent on the publisher for future versions. That's a different kind of vendor lock-in than paying for an API, but it's vendor lock-in nonetheless.
Third, auditability. If your industry has compliance requirements around how AI systems reach decisions, "we downloaded the weights from Meta" is a different answer than "here's our full training pipeline and data provenance."
The major models
Five organizations are producing the open models that matter. Here's where each stands.
Meta (Llama 4)
Meta released Llama 4 in April 2025 with three variants, all using mixture-of-experts architecture:
Scout has 17 billion active parameters across 16 experts (109 billion total). Its standout feature is a 10 million token context window, up from 128,000 in Llama 3. It runs on a single H100 GPU with quantization.
Maverick has 17 billion active parameters across 128 experts (400 billion total) with a 1 million token context window. Meta claims it outperforms GPT-4o and Gemini 2.0 Flash on coding, reasoning, and multilingual tasks.
Behemoth is the flagship: 288 billion active parameters, roughly 2 trillion total. It was still in training as of the announcement, but early benchmarks showed it outperforming GPT-4.5 and Claude Sonnet 3.7 on STEM tasks.
All three are natively multimodal (text and image input) and support 12 languages. The architecture shift to mixture-of-experts is significant because it means only a fraction of the model's parameters activate for any given query, which dramatically reduces the compute required to run inference.
Licensing: Meta Llama 3 Community License. Not open source by the OSI definition. Commercial use permitted with the 700 million MAU restriction. Attribution required.
The practical read: Llama 4 is the model most enterprises evaluate first because the Meta name provides organizational comfort. The models are genuinely capable, the community is enormous, and the fine-tuning ecosystem is mature. The licensing restrictions are worth reading carefully, but they won't affect most mid-market companies.
Mistral
The French lab released Mistral 3 in December 2025:
Mistral Large 3 is a mixture-of-experts model with 41 billion active parameters and 675 billion total. It's Mistral's most capable model to date and competitive with the frontier commercial offerings on most benchmarks.
Ministral 3 covers the smaller end: dense models at 3, 7, and 14 billion parameters for edge deployment and cost-sensitive workloads.
Devstral Small 2 is a 24 billion parameter coding model. Mistral claims it outperforms Qwen 3 Coder on code generation tasks.
Licensing: Apache 2.0 across the board. Both base and instruction-tuned versions. This is genuinely open source with no commercial restrictions.
The practical read: Mistral is the clearest "open source" play in the market. Apache 2.0 licensing means no surprises. The models are strong, and the European origin matters for companies with EU data residency requirements. If licensing clarity is a priority, Mistral is the safest bet.
DeepSeek
DeepSeek, out of China, made waves in January 2025 with R1, a reasoning model that claimed training costs of $5.6 million. That number needs context: it covers GPU costs for the pre-training run only, excluding R&D, hardware total cost of ownership, prior research iterations, and team costs. The actual investment is substantially higher. But even with caveats, the efficiency gains are real and put competitive pressure on the entire market.
DeepSeek uses mixture-of-experts architecture and prices aggressively. Their latest model lists at $0.28 per million input tokens with cache hits as low as $0.028. They halved prices in late 2025.
The controversies are real and varied:
OpenAI has claimed DeepSeek used ChatGPT outputs to train its models, which would violate OpenAI's terms of use. The White House and FBI have investigated whether DeepSeek obtained restricted NVIDIA chips through third parties in Singapore. Security researchers found R1 lacks basic guardrails against harmful prompts.
And the one that matters most for enterprise decisions: all DeepSeek user data is stored on servers in China. Chinese law requires companies to share data with state agencies upon request.
Government actions have been swift. NASA and the US Navy banned DeepSeek from all systems and devices. Congressional offices banned staff use. The "No DeepSeek on Government Devices Act" was introduced with bipartisan support in both the House and Senate. Italy removed it entirely from app stores. Australia, Taiwan, the Czech Republic, and the Netherlands imposed government device bans.
Licensing: Open weights, not open source.
The practical read: DeepSeek is technically impressive and the pricing is genuinely disruptive. But for US and European mid-market companies, the data sovereignty issue alone should give you pause. If your data touches anything regulated or proprietary, the risk calculus is straightforward. Run it locally with no data leaving your infrastructure, or don't use it.
Alibaba (Qwen)
Alibaba's Qwen3 family, released in April 2025, is the most comprehensive model lineup from any single provider:
Dense models ranging from 600 million to 32 billion parameters. Sparse models at 30 billion (3 billion active) and 235 billion (22 billion active). Trained on 36 trillion tokens across 119 languages and dialects.
In January 2026, Alibaba released Qwen3-Max with stronger agentic capabilities, including integrated tool use for web search, data extraction, and code execution.
The models support multimodal processing across text, audio, and vision. The latest omni model processes all three simultaneously.
Licensing: Apache 2.0. Genuinely open source.
The practical read: Qwen punches above its name recognition in the West. The model quality is competitive with Llama 4 and Mistral Large 3 across most benchmarks. The multilingual support (119 languages) is unmatched. The Apache 2.0 licensing is clean. The main friction for Western enterprises is the same as DeepSeek but less severe: Alibaba is a Chinese company, but the Apache 2.0 license means you're downloading and running the model yourself, not sending data to Chinese servers. That's a meaningful difference.
OpenAI (GPT-OSS)
The surprise entry. In August 2025, OpenAI released two open models:
gpt-oss-120b achieves near-parity with OpenAI's o4-mini on reasoning benchmarks. It runs on a single 80GB GPU.
gpt-oss-20b delivers results comparable to o3-mini and runs on edge devices with 16GB of memory.
Both are trained using reinforcement learning and techniques from OpenAI's frontier models. Both use Apache 2.0 licensing.
These models are not available through OpenAI's API or in ChatGPT. Weights are downloadable from Hugging Face only. This is OpenAI hedging its bets, and it was likely a competitive response to Llama 4 and DeepSeek's momentum.
The practical read: The models are capable and the licensing is clean. The strategic question is whether OpenAI will sustain this effort or treat it as a one-time competitive move. For companies already in the OpenAI ecosystem, these models offer a way to run workloads locally without switching providers. For everyone else, the Mistral and Qwen families offer more sustained investment and broader model selection.
The geopolitics
If you're making enterprise decisions about open AI models, the geopolitics are no longer background reading. They're part of the evaluation.
Export controls aren't working the way anyone planned
US export controls on advanced AI chips were supposed to slow Chinese AI development. DeepSeek's R1 demonstrated the opposite: constraints on compute forced efficiency innovations that produced competitive models on a fraction of the budget.
In January 2026, the Trump administration published new regulations permitting the sale of advanced AI chips (NVIDIA H200, AMD MI325X) to China, loosening restrictions that had been in place since 2022. Critics argue this undercuts years of strategic policy. Supporters argue the controls weren't working anyway and were mostly hurting US chipmakers.
The FBI is still investigating whether DeepSeek obtained restricted NVIDIA chips through third-party channels in Singapore before the policy change. The enforcement challenge is real: controlling the flow of commodity hardware across international supply chains is a different problem than controlling the export of nuclear materials.
For mid-market companies, the practical implication is this: Chinese open models will continue to improve regardless of US policy. The quality gap between Chinese and Western open models is narrowing. The question isn't whether to pay attention to them. It's how to evaluate the data sovereignty and compliance implications.
The national security debate
The NTIA's analysis of open weight models acknowledged the tension: making powerful models widely available could enable misuse, but it also diversifies AI development, decentralizes market control, and enables users to run models without sharing data with third parties.
The Biden administration's position was that the government "should not restrict the wide availability of model weights." The Trump administration has not reversed this stance. No definitive restrictions on open weight model distribution have been enacted.
The arguments in favor of keeping models open are substantive: open models expand R&D participation, prevent monopolistic control of AI capabilities, and improve security through public scrutiny. The arguments against are also substantive: adversaries can incorporate open models into military systems, and accountability mechanisms for misuse are unclear.
For enterprises, the practical takeaway: the regulatory environment for open AI models is unsettled. Building on open models is legal and will almost certainly remain legal. But if you're in a regulated industry, document your model selection rationale and be prepared to articulate why you chose the models you did.
The real economics
The pitch for open models sounds compelling: no per-token charges, full control over your data, ability to fine-tune for your use case. The reality is more complicated.
Where open models win on cost
A 2026 Lenovo analysis found that running models on owned infrastructure yields up to an 18x cost advantage per million tokens compared to commercial APIs. For high-utilization workloads, on-premises infrastructure hits breakeven in under four months.
That's a real number. If your company processes millions of tokens daily across well-defined workloads, the math can favor self-hosting significantly.
Where open models lose on cost
The number that doesn't make the pitch deck: AI engineers who can deploy, fine-tune, and maintain open models command $300,000 to $500,000 in annual compensation. Maintenance and vulnerability patching adds 15 to 30 percent in annual operational costs. At enterprise scale, core AI infrastructure runs $6 to $12 million annually including multi-region deployment, GPU clusters, and specialized team.
For most mid-market companies, the talent cost alone exceeds the API spend they're trying to avoid.
The honest math
The enterprise adoption numbers reflect this reality. Roughly 50 percent of companies rely solely on commercial models. 30 percent use a mix of commercial and open source. 20 percent have gone fully open source.
The companies going fully open source tend to be larger enterprises with existing ML teams, strict data residency requirements, or highly specialized workloads that justify the infrastructure investment. If three of those four conditions don't describe your company, commercial APIs are probably the more cost-effective choice.
The hybrid approach is where most mid-market companies end up: commercial APIs for general-purpose workloads, open models for specific tasks where fine-tuning or data control provides a meaningful advantage. One enterprise fintech platform fine-tuned Llama 3 70B on financial terminology for regulatory document summarization while using commercial APIs for everything else. That kind of targeted deployment is where the economics of open models actually work.
What's changing
The landscape is shifting in ways that affect the open-vs-commercial calculus. Four things to track.
The performance gap is narrowing
Two years ago, open models were clearly behind frontier commercial models on most tasks. That gap has compressed significantly. Llama 4 Maverick claims parity with GPT-4o. Mistral Large 3 is competitive with top-tier commercial offerings. Qwen3-Max benchmarks alongside the best commercial reasoning models.
The gap hasn't closed entirely at the frontier. The most capable commercial models (Claude Opus, GPT-5, Gemini Ultra) still lead on complex reasoning, nuanced instruction following, and agentic tasks. But for the majority of enterprise use cases, the performance difference between a top open model and a commercial API is smaller than the difference between either one and the model you were using eighteen months ago.
Mixture-of-experts is the new default
Every major open model released in 2025 used mixture-of-experts architecture. The idea: instead of activating all parameters for every query, route each query to a subset of specialized "expert" sub-networks.
The practical impact is dramatic. Llama 4 Scout has 109 billion total parameters but only activates 17 billion per query. This means you get the quality of a large model with the inference cost of a small one. It's the single biggest reason open models are becoming viable on reasonable hardware.
For enterprises evaluating self-hosting, MoE changes the infrastructure math. Models that would have required multi-GPU clusters two years ago now run on a single high-end GPU. That doesn't eliminate the talent and operational costs, but it significantly reduces the hardware barrier.
Multimodal is table stakes
Llama 4, Qwen3-Omni, and Mistral's latest releases all handle text and images natively. Qwen processes text, audio, and vision simultaneously. This isn't experimental anymore. Multimodal capability is a baseline feature.
For enterprises, this means the use cases that required separate models for text and vision (document processing with images, visual inspection combined with reporting, multimodal customer support) can now be handled by a single open model. The integration simplification alone can justify the evaluation.
Hugging Face is the infrastructure layer
Hugging Face has become the default platform for discovering, downloading, and deploying open models. The numbers: over 2 million models hosted, 500,000 datasets, 18 million monthly visitors, 5 million registered users, and more than 10,000 companies using the platform including Intel, Pfizer, Bloomberg, and eBay.
Their enterprise hub supports private model hosting, SOC 2 compliance, SSO, audit logs, and regional data storage. Over 2,000 organizations use the enterprise tier for secure deployment.
If you're evaluating open models, Hugging Face is where you'll start. Understanding their enterprise features is worth an hour of your team's time.
What this means for your company
The same principle applies here as with the commercial landscape: start with the problem, not the model.
Don't default to open source for cost savings. The per-token math looks attractive until you factor in infrastructure, talent, and operational overhead. For most mid-market companies processing fewer than a few million tokens daily, commercial APIs are cheaper when you account for total cost of ownership. The exceptions are specific and you'll know if you're one of them: strict data residency requirements, specialized fine-tuning needs, or existing ML teams with spare capacity.
Read the license before you build on it. "Open" is not a single thing. Llama's 700 million MAU restriction might not matter today, but it might matter after an acquisition. DeepSeek's data sovereignty implications are immediate. Apache 2.0 (Mistral, Qwen, GPT-OSS) is the cleanest option if licensing risk is a concern. Spend thirty minutes reading the actual license. It's shorter than most terms of service.
The hybrid approach is the practical one. Use commercial APIs as your default. Deploy open models for workloads where you have a specific, defensible reason: fine-tuning on proprietary data, air-gapped environments, regulatory requirements that prohibit sending data to third parties. This is how the 30 percent of companies running hybrid setups operate, and it's the configuration that balances capability with operational sanity.
If you're going to self-host, pick one model family and go deep. The temptation is to evaluate everything. The productive move is to pick Llama or Mistral or Qwen based on your licensing preferences and build real competence with that family. You'll learn more from deploying one model in production than from benchmarking five in a sandbox.
Don't ignore the geopolitics. If you're evaluating DeepSeek or Qwen, understand where your data goes and what jurisdiction governs it. For DeepSeek, any use that involves sending data to their API puts that data on Chinese servers subject to Chinese law. For Qwen, the Apache 2.0 license means you download and run the model locally, which is a fundamentally different risk profile. These are not equivalent "Chinese AI models." The deployment model matters more than the country of origin.
The bottom line
Open source AI is maturing fast. The models are genuinely capable. The performance gap with commercial offerings has narrowed to the point where it doesn't matter for most enterprise tasks. The mixture-of-experts revolution means the hardware requirements are falling in parallel. And the licensing landscape is clearer than it was a year ago, if you take the time to actually read it.
But "open" isn't free, and it isn't simple. The talent costs are real. The operational burden is real. The geopolitical complexity is real. And the label "open source" itself is doing less work than it appears to, because most of the models that use it don't meet the formal definition.
The companies getting this right aren't the ones who picked a side in the open-vs-closed debate. They're the ones who asked a simpler question: what problem are we solving, and which tool solves it with the least total friction? Sometimes that answer is an API call. Sometimes it's a model running on your own hardware. Usually it's both.
That's not a technology decision. It's a clarity decision. And it's the same one that separates companies that are moving from companies that are still reading blog posts about moving.
Get The Pepper Report
The short list of what we're writing and what we're reading.