SOTAVerified

nlg evaluation

Evaluate the generated text by NLG (Natural Language Generation) systems, like large language models

Papers

Showing 3140 of 71 papers

TitleStatusHype
ImaginE: An Imagination-Based Automatic Evaluation Metric for Natural Language Generation0
ImaginE: An Imagination-Based Automatic Evaluation Metric for Natural Language Generation0
Language Model Augmented Relevance Score0
Large Language Models Are Active Critics in NLG Evaluation0
LLM-based NLG Evaluation: Current Status and Challenges0
A Survey of Natural Language Generation0
MIPE: A Metric Independent Pipeline for Effective Code-Mixed NLG Evaluation0
A Tutorial on Evaluation Metrics used in Natural Language Generation0
Ev2R: Evaluating Evidence Retrieval in Automated Fact-Checking0
Agreement is overrated: A plea for correlation to assess human evaluation reliability0
Show:102550
← PrevPage 4 of 8Next →

No leaderboard results yet.