claim
LLM-as-Judge methods offer a more reliable alternative for factual evaluation in question-answering tasks because they show strong agreement with human judgments.
Authors
Sources
- Re-evaluating Hallucination Detection in LLMs - arXiv arxiv.org via serper
Referenced by nodes (2)
- Question Answering concept
- LLM-as-a-judge concept