Relations (1)
related 2.32 — strongly supporting 4 facts
The Hallucinations Leaderboard is a platform specifically designed to evaluate Large Language Models by measuring their reliability and tendency to generate hallucinated content, as described in [1], [2], and [3]. Furthermore, the leaderboard utilizes standardized evaluation frameworks to perform zero-shot and few-shot testing on these models, as noted in [4].
Facts (4)
Sources
The Hallucinations Leaderboard, an Open Effort to Measure ... huggingface.co 4 facts
claimThe Hallucinations Leaderboard is an open project designed to measure and address hallucinations in LLMs, aiming to provide insights into model generalization, limitations, and tendencies to generate hallucinated content.
claimThe Hallucinations Leaderboard is a platform designed to evaluate large language models against benchmarks specifically created to assess hallucination-related issues using in-context learning.
procedureThe Hallucinations Leaderboard utilizes the EleutherAI Language Model Evaluation Harness to perform zero-shot and few-shot evaluations of large language models via in-context learning.
claimThe Hallucinations Leaderboard evaluates Large Language Models (LLMs) on their ability to handle various types of hallucinations to provide researchers and developers with insights into model reliability and efficiency.