New blog post: why are we using LLMs as calculators? Mostly because we want to use all the software we write as calculators, and also because the end-goal is not calculators but "AGI". https://vickiboykis.com/2024/11/09/why-are-we-using-llms-as-calculators/
Comments
One who wasn't directly related to ML/AI asked why can't it do math well while computers are good at calculation. We went into tokenization, how models are trained & them working by sampling possible outputs
Before I started programming, I inadvertently applied for a programming-heavy engineering role. They asked me to count the 0s and 1s in an arbitrary number, eg 10110. I was unfamiliar with binary or string conversation tricks at the time...
I've seen various LLM plugins that let you access a Python interpreter. If we just gave these programs access to a calculator app and prompted it...
Is this cheating? Will it be indistinguishable from reason if we trivialize it's ability to do basic math?
That sounds to me like a high bar to clear for a "lookup" oriented process. Can't we just give the LLM access to basic calculation?
But LLMs don't distinguish between that. "Reasoning" is honestly too broad a term to describe the problem.