claim
Training data issues cause hallucinations because web corpora contain factual errors, misinformation, and knowledge imbalances that the next-token prediction objective cannot distinguish from accurate content, leading the model to learn errors with the same confidence as truths.
Authors
Sources
- Hallucination Causes: Why Language Models Fabricate Facts mbrenndoerfer.com via serper
Referenced by nodes (1)
- hallucination concept