I think one of the reasons that I like Rust, and why I've been so reluctant to really use LLMs at all comes from a core tenet in building safety critical systems:
When it really matters, it is better for a tool to say "I don't know", give up, and to make it clear it is doing so, than to guess wrong
When it really matters, it is better for a tool to say "I don't know", give up, and to make it clear it is doing so, than to guess wrong
Comments
1) Common questions about popular tools. If google couldn't find it right away, taking 5 seconds to ask the LLM before going to Discord or SO can't hurt, so long as the answer is verifiable.
They have limited context. It's better to know when you can't rely on your tool, than to think it is good when it isn't.
It's not easy to push that in either direction.
I'd rather my compiler say "abiguous, rejected", or my avionics say "Fault, inoperative" when they aren't sure.
Software development is relatively much more relaxed in time.
I know folks who work on diagnostics work VERY hard to avoid misleading ones.
It has generated working code, but when it doesn't, you have to fight it, and be very clear in your instructions.
Maybe LLMs work for other people, but I can't shake the feeling that creeping doubt in the back of my mind about the quality of what it is doing, would drain my productivity.
That might just be me, but 🤷
I think a lot of the efficiency issues will be addressed in time.
Provenance of training data might get fixed at some point, I doubt any time soon though.
I think treating statistical tools the same as deterministic tools is always a mistake.
library you’re not familiar with. I need to play around with running some locally. If I had one that I could feed an example directory and docs to then it just uses that as reference that be perfect
I just worry about how they are marketed, and how people expect to use them in their workflows.
This limit was hit during the incident, and operators thought it had leveled off.
If you have some time to spare, I'd highly recommend to watch this 3 1/2 hour explainer from Andrey Karpathy (filled with many details few know probably about modern LLMs and how they get built) ->