Au contraire—here's one paper detailing automation to solve the fact tracing problem in LLMs, that even predates the release of ChatGPT:
https://arxiv.org/pdf/2205.11482
Disclaimer, I haven't read this, just showing that this is a studiable problem. We don't have to guess.
https://arxiv.org/pdf/2205.11482
Disclaimer, I haven't read this, just showing that this is a studiable problem. We don't have to guess.
Comments
Authority laundering via LLM is def a problem and we will need to develop social conventions to combat it.
Easy to verify if you just re-run the completion. It comes up with something else. This is high-variance.
In OP's example, this is an easy way to show a student the bot is "lying".
If synthetic data overly skews the proportion of factuality in training data, that's a problem, and will defeat this little heuristic.