procedure
The Hallucination Metric from the DeepEval package estimates the likelihood of hallucination as the degree to which an LLM response contradicts or disagrees with the provided context, as assessed by an LLM (specifically GPT-4o-mini in the Cleanlab study).
Authors
Sources
- Benchmarking Hallucination Detection Methods in RAG - Cleanlab cleanlab.ai via serper
Referenced by nodes (2)
- Cleanlab entity
- gpt-4o-mini concept