claim
Self-Evaluation, also known as Self-Reflection or LLM as a judge, is a technique where an LLM is asked to evaluate its own generated answer and rate its confidence on a 1-5 Likert scale.
Authors
Sources
- Benchmarking Hallucination Detection Methods in RAG - Cleanlab cleanlab.ai via serper
Referenced by nodes (2)
- LLM-as-a-judge concept
- Likert scale concept