Relations (1)

related 3.91 — strongly supporting 14 facts

Justification not yet generated — showing supporting facts

Facts (14)

Sources
EdinburghNLP/awesome-hallucination-detection - GitHub github.com GitHub 4 facts
claimKF1, BLEU, ROUGE, chrF, METEOR, BERTScore, BARTScore, BLEURT, and average length are metrics used for evaluating AI systems.
claimHallucination detection metrics measure either the degree of hallucination in generated responses relative to given knowledge or their overlap with gold faithful responses, including Critic, Q² (F1, NLI), BERTScore, F1, BLEU, and ROUGE.
referenceThe QuestEval metric is used for testing consistency, coherence, fluency, and relevance in AI-generated text, alongside other metrics like ROUGE, BLUE, METEOR, BERTScore, SummaQA, and QAGS.
referenceThe 'Survey of Hallucination in Natural Language Generation' classifies metrics into statistical metrics (such as ROUGE, BLEU, PARENT, and Knowledge F1) and model-based metrics.
Survey and analysis of hallucinations in large language models frontiersin.org Frontiers 3 facts
claimTraditional automatic metrics like BLEU, ROUGE, and METEOR are inadequate for assessing factual consistency in large language models, according to Maynez et al. (2020).
claimTraditional lexical metrics like BLEU or ROUGE fail to capture semantic grounding in AI systems.
claimAutomatic metrics such as BLEU or ROUGE fail to capture factual consistency and reliability in Large Language Models, according to Maynez et al. (2020).
Detecting and Evaluating Medical Hallucinations in Large Vision ... arxiv.org arXiv 2 facts
claimThe BLEU metric accounts for significant length differences between generated text and ground truth, making it more versatile than ROUGE, but it remains a weak measure of factual correctness.
claimBertScore mitigates some shortcomings of ROUGE and BLEU but does not intuitively reflect factual accuracy or the degree of hallucination in medical texts.
A Comprehensive Benchmark and Evaluation Framework for Multi ... arxiv.org arXiv 2 facts
claimThe evaluation of medical agents has evolved from linguistic metrics like BLEU and ROUGE to action-oriented benchmarks such as MedAgentBench and MedAgentBoard.
claimTraditional n-gram metrics like ROUGE and BLEU are insufficient for capturing the clinical validity of generated text in medical LLMs.
Unknown source 1 fact
claimBLEU, ROUGE, and METEOR are traditional automatic metrics used for evaluating text generation.
Practices, opportunities and challenges in the fusion of knowledge ... frontiersin.org Frontiers 1 fact
claimCurrent evaluation metrics like BLEU (Papineni et al., 2002) and ROUGE (Lin, 2004) mainly measure surface text similarity and fail to effectively capture the semantic consistency between generated text and knowledge graph content.
A framework to assess clinical safety and hallucination rates of LLMs ... nature.com Nature 1 fact
claimAutomated metrics like ROUGE, BLEU, and BERT-score, which are designed to compare model-generated text with expert-written examples, have significant limitations in healthcare because they focus on surface-level textual similarity rather than semantic nuances, contextual dependencies, and domain-specific knowledge.