New blogpost: "Training as we know it might end".
It was originally a panorama of the new methods of synthetic generation but the stakes are now much higher and I openly wonder if model training is not soon going to change forever. https://vintagedata.org/blog/posts/training-as-we-know-it-will-end
It was originally a panorama of the new methods of synthetic generation but the stakes are now much higher and I openly wonder if model training is not soon going to change forever. https://vintagedata.org/blog/posts/training-as-we-know-it-will-end
Comments
I see that "The physics of language models" is one key source for this line of thought.
Is another source, perhaps, recent papers showing that a small number of excellent examples of a heuristic can have huge benefits in training data? I unfortunately forget the citations.
https://arxiv.org/pdf/2505.03335
Maybe we'll find out...
I agree on data selection being v important. Not too sure everything is gonna be RL gyms.