If I have time I'll put together a more detailed thread tomorrow, but for now, I think this new paper about limitations of Chain-of-Thought models could be quite important. Worth a look if you're interested in these sorts of things.
https://ml-site.cdn-apple.com/papers/the-illusion-of-thinking.pdf
https://ml-site.cdn-apple.com/papers/the-illusion-of-thinking.pdf
Comments
You can prompt an LLM to sort of behave in a CoT like fashion, but there are also models being explicitly designed around working that way period which is what CoT is usually referring to.
And Iβm also not a fan of following up a scientific journal article with a magazine articleβ¦ appreciate your thoughts/thread
Whether large volume of language (many words and phrases) or large items of language (big words), the idea is easier than SP = stochastic parrot, as description of "AI" tools (with little intelligence) π»π¦
https://en.wikipedia.org/wiki/Stochastic_parrot
Does a machine have cognition? I think we should keep the process of AI in mind in order to define it.
The real issue is what's meant by "modelling of reasoning going on."
I wish the AI companies published all details about how their models work, so that researchers could make theoretical analysis of their principles, instead of this black box prodding.