claim
Large language models trained on supervised finetuning data learn the style of confident, well-structured prose because human annotators tend to produce such responses when demonstrating ideal answers.
Authors
Sources
- Hallucination Causes: Why Language Models Fabricate Facts mbrenndoerfer.com via serper
Referenced by nodes (2)
- Large Language Models concept
- supervised fine-tuning concept