claim
Current Large Language Models struggle most to detect hallucinated content that is semantically close to the truth.
Authors
Sources
- MedHallu: Benchmark for Medical LLM Hallucination Detection www.emergentmind.com via serper
Referenced by nodes (1)
- Large Language Models concept