SOTAVerified

Machine Translation

Machine translation is the task of translating a sentence in a source language to a different target language.

Approaches for machine translation can range from rule-based to statistical to neural-based. More recently, encoder-decoder attention-based architectures like BERT have attained major improvements in machine translation.

One of the most popular datasets used to benchmark machine translation systems is the WMT family of datasets. Some of the most commonly used evaluation metrics for machine translation systems include BLEU, METEOR, NIST, and others.

( Image credit: Google seq2seq )

Papers

Showing 12511275 of 10752 papers

TitleStatusHype
A Study in Improving BLEU Reference Coverage with Diverse Automatic ParaphrasingCode0
Fast and Simple Mixture of Softmaxes with BPE and Hybrid-LightRNN for Language GenerationCode0
From Gameplay to Symbolic Reasoning: Learning SAT Solver Heuristics in the Style of Alpha(Go) ZeroCode0
Evaluating Structural Generalization in Neural Machine TranslationCode0
Evaluating Sequence-to-Sequence Models for Handwritten Text RecognitionCode0
Evaluating the morphological competence of Machine Translation SystemsCode0
Evaluating Pronominal Anaphora in Machine Translation: An Evaluation Measure and a Test SuiteCode0
An Empirical Study on the Robustness of Massively Multilingual Neural Machine TranslationCode0
Evaluating Rewards for Question Generation ModelsCode0
Evaluating the Morphosyntactic Well-formedness of Generated TextsCode0
Evaluating Machine Translation Models for English-Hindi Language Pairs: A Comparative AnalysisCode0
An Empirical Study of Translation Hypothesis Ensembling with Large Language ModelsCode0
A Call for Clarity in Reporting BLEU ScoresCode0
Evaluating Optimal Reference TranslationsCode0
Evaluation of Chinese-English Machine Translation of Emotion-Loaded Microblog Texts: A Human Annotated Dataset for the Quality Assessment of Emotion TranslationCode0
Estimating post-editing effort: a study on human judgements, task-based and reference-based metrics of MT qualityCode0
Evaluating Automatic Metrics with Incremental Machine Translation SystemsCode0
Escaping the sentence-level paradigm in machine translationCode0
Evaluating bilingual word embeddings on the long tailCode0
Equalizing Gender Biases in Neural Machine Translation with Word Embeddings TechniquesCode0
Error Analysis of Cross-lingual Tagging and ParsingCode0
Entity Projection via Machine Translation for Cross-Lingual NERCode0
Evaluating Gender Bias in German Machine TranslationCode0
Evaluation of Google Translate for Mandarin Chinese translation using sentiment and semantic analysisCode0
Advancing Neural Network Performance through Emergence-Promoting Initialization SchemeCode0
Show:102550
← PrevPage 51 of 431Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Transformer Cycle (Rev)BLEU score35.14Unverified
2Noisy back-translationBLEU score35Unverified
3Transformer+Rep(Uni)BLEU score33.89Unverified
4T5-11BBLEU score32.1Unverified
5BiBERTBLEU score31.26Unverified
6Transformer + R-DropBLEU score30.91Unverified
7Bi-SimCutBLEU score30.78Unverified
8BERT-fused NMTBLEU score30.75Unverified
9Data Diversification - TransformerBLEU score30.7Unverified
10SimCutBLEU score30.56Unverified
#ModelMetricClaimedVerifiedStatus
1Transformer+BT (ADMIN init)BLEU score46.4Unverified
2Noisy back-translationBLEU score45.6Unverified
3mRASP+Fine-TuneBLEU score44.3Unverified
4Transformer + R-DropBLEU score43.95Unverified
5Transformer (ADMIN init)BLEU score43.8Unverified
6AdminBLEU score43.8Unverified
7BERT-fused NMTBLEU score43.78Unverified
8MUSE(Paralllel Multi-scale Attention)BLEU score43.5Unverified
9T5BLEU score43.4Unverified
10Local Joint Self-attentionBLEU score43.3Unverified
#ModelMetricClaimedVerifiedStatus
1PiNMTBLEU score40.43Unverified
2BiBERTBLEU score38.61Unverified
3Bi-SimCutBLEU score38.37Unverified
4Cutoff + Relaxed Attention + LMBLEU score37.96Unverified
5DRDABLEU score37.95Unverified
6Transformer + R-Drop + CutoffBLEU score37.9Unverified
7SimCutBLEU score37.81Unverified
8Cutoff+KneeBLEU score37.78Unverified
9CutoffBLEU score37.6Unverified
10CipherDAugBLEU score37.53Unverified
#ModelMetricClaimedVerifiedStatus
1HWTSC-Teacher-SimScore19.97Unverified
2MS-COMET-22Score19.89Unverified
3MS-COMET-QE-22Score19.76Unverified
4KG-BERTScoreScore17.28Unverified
5metricx_xl_DA_2019Score17.17Unverified
6COMET-QEScore16.8Unverified
7COMET-22Score16.31Unverified
8UniTE-srcScore15.68Unverified
9UniTE-refScore15.38Unverified
10metricx_xxl_DA_2019Score15.24Unverified