claim
The authors of 'Re-evaluating Hallucination Detection in LLMs' assert that ROUGE misaligns with the requirements for evaluating hallucination detection in Large Language Models.
The authors of 'Re-evaluating Hallucination Detection in LLMs' assert that ROUGE misaligns with the requirements for evaluating hallucination detection in Large Language Models.