reference
The paper 'Unfamiliar Finetuning Examples Control How Language Models Hallucinate' by Kang et al. (2024) investigates the impact of finetuning examples on hallucination behavior.
Authors
Sources
- Awesome-Hallucination-Detection-and-Mitigation - GitHub github.com via serper
Referenced by nodes (3)
- hallucination concept
- Language Model concept
- fine-tuning concept