claim
Large language models are trained to always continue a sequence, and the training objective rewards producing probable continuations, which results in the absence of a built-in mechanism for stopping generation or outputting an 'uncertain' token when the model lacks knowledge.
Authors
Sources
- Hallucination Causes: Why Language Models Fabricate Facts mbrenndoerfer.com via serper
Referenced by nodes (1)
- Large Language Models concept