claim
Inconsistent annotator knowledge in long-running annotation projects leads to conflicting 'correct' responses within Supervised Fine-Tuning (SFT) datasets, which degrades model calibration and causes inconsistent outputs at inference time.
Authors
Sources
- Hallucination Causes: Why Language Models Fabricate Facts mbrenndoerfer.com via serper
Referenced by nodes (1)
- supervised fine-tuning concept