reference
The HalluLens benchmark is a comprehensive hallucination benchmark for Large Language Models that incorporates both new extrinsic and existing intrinsic evaluation tasks and is built upon a clear taxonomy.
Authors
Sources
- [PDF] HalluLens: LLM Hallucination Benchmark | Semantic Scholar www.semanticscholar.org via serper
Referenced by nodes (1)
- Large Language Models concept