reference
The HalluQA benchmark evaluates AI models using human annotations of intrinsic and extrinsic hallucinated spans and factuality, alongside metrics such as ROUGE-1/2/L, BERTScore, textual entailment, QA-based consistency, and Spearman correlation with human scores.
Authors
Sources
- EdinburghNLP/awesome-hallucination-detection - GitHub github.com via serper
Referenced by nodes (1)
- BERTScore concept