More OLMo! More performance! More details!
We applied Tulu post-training to OLMo 2 as well, so you can get strong model performance AND see what your model was actually trained on.
Reposted from Kyle Lo @ ICLR 2025
kicking off 2025 with our OLMo 2 tech report while payin homage to the sequelest of sequels ๐Ÿซก

๐Ÿš— 2 OLMo 2 Furious ๐Ÿ”ฅ is everythin we learned since OLMo 1, with deep dives into:

๐Ÿš– stable pretrain recipe
๐Ÿš” lr anneal ๐Ÿค data curricula ๐Ÿค soups
๐Ÿš˜ tulu post-train recipe
๐Ÿšœ compute infra setup

๐Ÿ‘‡๐Ÿงต

Comments