SOTAVerified

Towards a Better Evaluation of Metrics for Machine Translation

2020-11-01WMT (EMNLP) 2020Unverified0· sign in to hype

Peter Stanchev, Weiyue Wang, Hermann Ney

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

An important aspect of machine translation is its evaluation, which can be achieved through the use of a variety of metrics. To compare these metrics, the workshop on statistical machine translation annually evaluates metrics based on their correlation with human judgement. Over the years, methods for measuring correlation with humans have changed, but little research has been performed on what the optimal methods for acquiring human scores are and how human correlation can be measured. In this work, the methods for evaluating metrics at both system- and segment-level are analyzed in detail and their shortcomings are pointed out.

Tasks

Reproductions