reference
A previous study benchmarking alternative hallucination detection techniques, including DeepEval, G-Eval, and RAGAS, found that TLM (Trustworthy Language Model) evaluation models detect incorrect RAG responses with higher precision and recall.
Authors
Sources
- Real-Time Evaluation Models for RAG: Who Detects Hallucinations ... cleanlab.ai via serper
Referenced by nodes (3)
- RAGAS concept
- Trustworthy Language Model concept
- TLM concept