More OLMo! More performance! More details!
We applied Tulu post-training to OLMo 2 as well, so you can get strong model performance AND see what your model was actually trained on.
We applied Tulu post-training to OLMo 2 as well, so you can get strong model performance AND see what your model was actually trained on.
Reposted from
Kyle Lo @ ICLR 2025
kicking off 2025 with our OLMo 2 tech report while payin homage to the sequelest of sequels ๐ซก
๐ 2 OLMo 2 Furious ๐ฅ is everythin we learned since OLMo 1, with deep dives into:
๐ stable pretrain recipe
๐ lr anneal ๐ค data curricula ๐ค soups
๐ tulu post-train recipe
๐ compute infra setup
๐๐งต
๐ 2 OLMo 2 Furious ๐ฅ is everythin we learned since OLMo 1, with deep dives into:
๐ stable pretrain recipe
๐ lr anneal ๐ค data curricula ๐ค soups
๐ tulu post-train recipe
๐ compute infra setup
๐๐งต
Comments