Super exiting stuff. I hope this will lead to smaller, more capable models!
Reposted from Clem Delangue 🤗
Just 10 days after o1's public debut, we’re thrilled to unveil the open-source version of the technique behind its success: scaling test-time compute

By giving models more "time to think," Llama 1B outperforms Llama 8B in math—beating a model 8x its size. The full recipe is open-source!

Comments