reference
Gaurang Sriramanan et al. (2024) developed 'LLM-Check', a method for investigating the detection of hallucinations in large language models, published in Advances in Neural Information Processing Systems, volume 37.
Authors
Sources
- Re-evaluating Hallucination Detection in LLMs - arXiv arxiv.org via serper
- Empowering GraphRAG with Knowledge Filtering and Integration arxiv.org via serper
- Awesome-Hallucination-Detection-and-Mitigation - GitHub github.com via serper
Referenced by nodes (3)
- Large Language Models concept
- hallucination detection concept
- Advances in Neural Information Processing Systems entity