SOTAVerified

nlg evaluation

Evaluate the generated text by NLG (Natural Language Generation) systems, like large language models

Papers

Showing 5171 of 71 papers

TitleStatusHype
A Survey of Natural Language Generation0
ImaginE: An Imagination-Based Automatic Evaluation Metric for Natural Language Generation0
Compression, Transduction, and Creation: A Unified Framework for Evaluating Natural Language GenerationCode1
Perturbation CheckLists for Evaluating NLG Evaluation MetricsCode0
Language Model Augmented Relevance Score0
All That's `Human' Is Not Gold: Evaluating Human Evaluation of Generated Text0
MIPE: A Metric Independent Pipeline for Effective Code-Mixed NLG Evaluation0
All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text0
Active Evaluation: Efficient NLG Evaluation with Few Pairwise Comparisons0
ImaginE: An Imagination-Based Automatic Evaluation Metric for Natural Language Generation0
A Tutorial on Evaluation Metrics used in Natural Language Generation0
A Study of Automatic Metrics for the Evaluation of Natural Language ExplanationsCode0
Evaluation rules! On the use of grammars and rule-based systems for NLG evaluation0
A Survey of Evaluation Metrics Used for NLG Systems0
Evaluation of Text Generation: A Survey0
Agreement is overrated: A plea for correlation to assess human evaluation reliability0
The use of rating and Likert scales in Natural Language Generation human evaluation tasks: A review and some recommendations0
Treat the system like a human student: Automatic naturalness evaluation of generated text without reference texts0
Why We Need New Evaluation Metrics for NLGCode0
Survey of the State of the Art in Natural Language Generation: Core tasks, applications and evaluation0
A Snapshot of NLG Evaluation Practices 2005 - 20140
Show:102550
← PrevPage 3 of 3Next →

No leaderboard results yet.