claim
Evaluation of medical capabilities in existing Large Vision-Language Models (LVLMs) is unreliable because it relies on outdated benchmarks that suffer from data leakage during pre-training.
Authors
Sources
- Detecting and Evaluating Medical Hallucinations in Large Vision ... arxiv.org via serper
Referenced by nodes (1)
- Large Vision-Language Models concept