claim
Research presented at EMNLP 2025 found that alignment-tuned Large Language Models produce more faithful explanations than base models, and that faithfulness and plausibility are positively correlated.
Authors
Sources
- EdinburghNLP/awesome-hallucination-detection - GitHub github.com via serper
Referenced by nodes (2)
- Large Language Models concept
- faithfulness concept