claim
HaluEval is a collection of generated and human-annotated hallucinated samples used for evaluating the performance of large language models in recognizing hallucinations.
Authors
Sources
- EdinburghNLP/awesome-hallucination-detection - GitHub github.com via serper
- A Knowledge Graph-Based Hallucination Benchmark for Evaluating ... arxiv.org via serper
Referenced by nodes (3)
- Large Language Models concept
- hallucination concept
- HaluEval concept