claim
Trustworthy Language Model (TLM) scores the trustworthiness of LLM responses using a combination of self-reflection, consistency across multiple sampled responses, and probabilistic measures.
Authors
Sources
- Benchmarking Hallucination Detection Methods in RAG - Cleanlab cleanlab.ai via serper
Referenced by nodes (1)
- Trustworthy Language Model concept