claim
Trustworthy Language Model (TLM) can identify scenarios where LLMs made reasoning or factuality errors, identified multiple contradictory yet plausible responses, or were given atypical prompts relative to their original training data.
Authors
Sources
- Benchmarking Hallucination Detection Methods in RAG - Cleanlab cleanlab.ai via serper
Referenced by nodes (1)
- Trustworthy Language Model concept