The Allen Institute for AI (Ai2) just lately launched what it calls its strongest family of models yet, Olmo 3. However the firm stored iterating on the fashions, increasing its reinforcement studying (RL) runs, to create Olmo 3.1.
The brand new Olmo 3.1 fashions focus on effectivity, transparency, and management for enterprises.
Ai2 up to date two of the three variations of Olmo 2: Olmo 3.1 Suppose 32B, the flagship mannequin optimized for superior analysis, and Olmo 3.1 Instruct 32B, designed for instruction-following, multi-turn dialogue, and gear use.
Olmo 3 has a 3rd model, Olmo 3-Base for programming, comprehension, and math. It additionally works nicely for proceed fine-tuning.
Ai2 stated that to improve Olmo 3 Suppose 32B to Olmo 3.1, its researchers prolonged its finest RL run with an extended coaching schedule.
“After the authentic Olmo 3 launch, we resumed our RL coaching run for Olmo 3 32B Suppose, coaching for a further 21 days on 224 GPUs with additional epochs over our Dolci-Suppose-RL dataset,” Ai2 stated in a blog post. “This yielded Olmo 3.1 32B Suppose, which brings substantial features throughout math, reasoning, and instruction-following benchmarks: enhancements of 5+ factors on AIME, 4+ factors on ZebraLogic, 4+ factors on IFEval, and 20+ factors on IFBench, alongside stronger efficiency on coding and complicated multi-step duties.”
To get to Olmo 3.1 Instruct, Ai2 stated its researchers utilized the recipe behind the smaller Instruct dimension, 7B, to the bigger mannequin.
Olmo 3.1 Instruct 32B is “optimized for chat, instrument use, & multi-turn dialogue—making it a way more performant sibling of Olmo 3 Instruct 7B and prepared for real-world functions,” Ai2 stated in a post on X.
For now, the new checkpoints are accessible on the Ai2 Playground or Hugging Face, with API entry coming quickly.
Higher efficiency on benchmarks
The Olmo 3.1 fashions carried out nicely on benchmark exams, predictably beating the Olmo 3 fashions.
Olmo 3.1 Suppose outperformed Qwen 3 32B fashions in the AIME 2025 benchmark and carried out shut to Gemma 27B.

Olmo 3.1 Instruct carried out strongly towards its open-source friends, even beating fashions like Gemma 3 on the Math benchmark.

“As for Olmo 3.1 32B Instruct, it’s a larger-scale instruction-tuned mannequin constructed for chat, instrument use, and multi-turn dialogue. Olmo 3.1 32B Instruct is our most succesful absolutely open chat mannequin to date and — in our evaluations — the strongest absolutely open 32B-scale instruct mannequin,” the firm stated.
Ai2 additionally upgraded its RL-Zero 7B fashions for math and coding. The corporate stated on X that each fashions benefited from longer and extra steady coaching runs.
Dedication to transparency and open supply
Ai2 beforehand advised VentureBeat that it designed the Olmo 3 household of fashions to provide enterprises and analysis labs extra management and understanding of the information and coaching that went into the mannequin.
Organizations might add to the mannequin’s information combine and retrain it to additionally study from what’s been added.
This has lengthy been a dedication for Ai2, which additionally provides a tool called OlmoTrace that tracks how LLM outputs match its coaching information.
“Collectively, Olmo 3.1 Suppose 32B and Olmo 3.1 Instruct 32B present that openness and efficiency can advance collectively. By extending the identical mannequin circulation, we proceed to enhance capabilities whereas retaining end-to-end transparency over information, code, and coaching selections,” Ai2 stated.
Disclaimer: This article is sourced from external platforms. OverBeta has not independently verified the information. Readers are advised to verify details before relying on them.