212

Olmo 3

Team Olmo
Allyson Ettinger
Amanda Bertsch
Bailey Kuehl
David Graham
David Heineman
Dirk Groeneveld
Faeze Brahman
Finbarr Timbers
Hamish Ivison
Jacob Morrison
Jake Poznanski
Kyle Lo
Luca Soldaini
Matt Jordan
Mayee Chen
Michael Noukhovitch
Nathan Lambert
Pete Walsh
Pradeep Dasigi
Robert Berry
Saumya Malik
Saurabh Shah
Scott Geng
Shane Arora
Shashank Gupta
Taira Anderson
Teng Xiao
Tyler Murray
Tyler Romero
Victoria Graf
Akari Asai
Akshita Bhagia
Alexander Wettig
Alisa Liu
Aman Rangapur
Chloe Anastasiades
Costa Huang
Dustin Schwenk
Harsh Trivedi
Ian Magnusson
Jaron Lochner
Jiacheng Liu
Lester James V. Miranda
Maarten Sap
Malia Morgan
Michael Schmitz
Michal Guerquin
Michael Wilson
Regan Huff
Ronan Le Bras
Rui Xin
Rulin Shao
Sam Skjonsberg
Shannon Zejiang Shen
Shuyue Stella Li
Tucker Wilde
Valentina Pyatkin
Will Merrill
Yapei Chang
Yuling Gu
Zhiyuan Zeng
Ashish Sabharwal
Luke Zettlemoyer
Pang Wei Koh
Ali Farhadi
Noah A. Smith
Hannaneh Hajishirzi
Main:56 Pages
35 Figures
Bibliography:1 Pages
51 Tables
Appendix:60 Pages
Abstract

We introduce Olmo 3, a family of state-of-the-art, fully-open language models at the 7B and 32B parameter scales. Olmo 3 model construction targets long-context reasoning, function calling, coding, instruction following, general chat, and knowledge recall. This release includes the entire model flow, i.e., the full lifecycle of the family of models, including every stage, checkpoint, data point, and dependency used to build it. Our flagship model, Olmo 3 Think 32B, is the strongest fully-open thinking model released to-date.

View on arXiv
Comments on this paper