News

Scaleups

Ai2 Releases Olmo 3 With Complete Model Development Transparency

The Allen Institute for AI launched Olmo 3, a family of fully open-source language models that provides unprecedented access to the entire development process, from training data to final weights.

The Allen Institute for AI launched Olmo 3, a family of fully open-source language models that provides unprecedented access to the entire development process, from training data to final weights.

The Allen Institute for AI launched Olmo 3, a family of fully open-source language models that provides unprecedented access to the entire development process, from training data to final weights.

NewDecoded

Published Nov 22, 2025

Nov 22, 2025

4 min read

Image from Ai2

Revolutionary Approach to Open AI

On November 20, 2025, the Allen Institute for AI (Ai2) introduced Olmo 3, a significant departure from typical open-source AI releases. Instead of simply sharing final model weights, Ai2 provides access to what they call the "model flow": every dataset, checkpoint, training stage, and code dependency used throughout development. The release includes four model variants at 7B and 32B parameter scales: Olmo 3-Base for foundational tasks, Olmo 3-Think for advanced reasoning, Olmo 3-Instruct for conversational applications, and Olmo 3-RL Zero for reinforcement learning research.

Competitive Performance at Lower Cost

Olmo 3-Think (32B) achieves 96.1% on the MATH benchmark and 91.4% on HumanEvalPlus coding tests, positioning it as what Ai2 calls the strongest fully open thinking model available. The base models demonstrate strong capabilities in programming, reading comprehension, and mathematical problem solving while maintaining performance across extended context lengths of up to 65,000 tokens. Notably, Ai2 reports that Olmo 3 requires roughly 6x fewer training tokens than comparable models and is 2.5x more efficient to train than Meta's Llama 3.1 based on GPU-hours per token.

Complete Data Pipeline Released

The training infrastructure includes Dolma 3, a 9.3-trillion-token corpus encompassing web data, scientific PDFs, code, and mathematical content. Ai2 also released Dolci, a specialized post-training data suite for reasoning and tool use. All datasets come with extensive decontamination and quality filtering, and the team provides open-source tools for data processing, including utilities for deduplication, cleaning, and test set removal. Training was conducted on up to 1,024 H100 GPUs, achieving 7,700 tokens per device per second for the 7B model.

Traceability and Customization Tools

Integration with OlmoTrace allows researchers to connect model outputs directly back to specific training data in real time. The release includes checkpoints from every major training milestone, enabling developers to fork the model at any point in its development. Teams can study capability emergence over time, run ablations on specific stages, or customize training for domain-specific needs. All components are released under the Apache 2.0 license and are available through Hugging Face, the Ai2 Playground, and OpenRouter.

Industry Reception

Noah Smith, Ai2's senior director of NLP research, emphasized that many customers from regulated enterprises to research institutions prioritize transparency and data control. The organization positions the 32B scale as a sweet spot that balances strong performance with accessibility for fine-tuning and deployment on modest hardware. Omdia analyst Mark Beccue noted that Ai2's approach demonstrates sustained commitment to genuine openness, particularly relevant as agentic AI capabilities increasingly require sophisticated reasoning models.

Decoded Take

Decoded Take

Decoded Take

Olmo 3 arrives at a critical inflection point for open-source AI. While companies like Meta and OpenAI have faced criticism for limited transparency (OpenAI particularly over hiding raw reasoning tokens in their models), Ai2's comprehensive release challenges the industry's definition of "open."

By exposing the complete development pipeline, Ai2 addresses growing enterprise demands for auditability and control, especially in regulated sectors. The timing is strategic: as reasoning models and agentic AI gain prominence, organizations need to understand not just what models do, but why. The 32B parameter ceiling reflects a pragmatic choice favoring accessibility over raw scale, targeting researchers and businesses that lack hyperscale infrastructure. This positions Olmo 3 as infrastructure for the next wave of specialized AI development, where customization and trust matter more than benchmark supremacy.

The release effectively raises the bar for what "open source" should mean in AI, potentially pressuring competitors to increase transparency or risk losing credibility with developers who increasingly demand more than downloadable weights

Share this article

Related Articles

Related Articles

Related Articles