claim
Supervised finetuning datasets suffer from a selection bias where annotators write detailed, authoritative responses for topics they know well and shorter, hedged responses for topics where they are less confident, which reinforces model overconfidence on well-known topics.
Authors
Sources
- Hallucination Causes: Why Language Models Fabricate Facts mbrenndoerfer.com via serper
Referenced by nodes (1)
- supervised fine-tuning concept