Today we release OLMo 2 32B, the most capable and largest model in the OLMo 2 family, scaling up the OLMo 2 training recipe used for our 7B and 13B models released in November. It is trained up to 6T tokens and post-trained using Tulu 3.1. OLMo 2 32B is the first fully-open model (all data, code, weights, and details are freely available) to outperform GPT3.5-Turbo and GPT-4o mini on a suite of popular, multi-skill academic benchmarks. It is comparable to the leading open-weight models while requiring only a fraction of training compute. For example, OLMo 2 32B takes only one third of the cost of training Qwen 2.5 32B while reaching similar performance. The OLMo 2 family of models—now available in 7B, 13B, and 32B parameter sizes, all can be finetuned on a single H100 GPU node, and all models are available on the Ai2 playground.
Notes: first table here: https://allenai.org/blog/olmo2-32B
Size Notes: "It is trained up to 6T tokens and post-trained using Tulu 3.1." "OLMo 2 32B is trained for 1.5 epochs, up to 6T tokens" Pretraining Stage 1 (OLMo-Mix-1124) 6T tokens ( = 1.5 epochs) Pretraining Stage 2 (Dolmino-Mix-1124) 100B tokens (3 runs) 300B tokens (1 run) merged Post-training (Tulu 3 SFT OLMo mix) SFT + DPO + PPO (preference mix)
Notes: 32B