formula
BLEU (Bilingual Evaluation Understudy) is a metric used to evaluate text quality in large language models integrated with knowledge graphs by comparing generated text to human-written reference texts, calculated as BLEU = BP * exp(sum(w_n * log(p_n))), where BP is the brevity penalty, w_n are weights, and p_n are precision scores for n-grams.

Authors

Sources

Referenced by nodes (3)