reference
Li et al. (2023) created 'HaluEval', a large-scale benchmark for evaluating hallucinations in Large Language Models.
Authors
Sources
- Re-evaluating Hallucination Detection in LLMs - arXiv arxiv.org via serper
- The Synergy of Symbolic and Connectionist AI in LLM ... arxiv.org via serper
- A Knowledge Graph-Based Hallucination Benchmark for Evaluating ... arxiv.org via serper
- Awesome-Hallucination-Detection-and-Mitigation - GitHub github.com via serper
Referenced by nodes (3)
- Large Language Models concept
- hallucination concept
- HaluEval concept