SOTAVerified

Machine Translation

Machine translation is the task of translating a sentence in a source language to a different target language.

Approaches for machine translation can range from rule-based to statistical to neural-based. More recently, encoder-decoder attention-based architectures like BERT have attained major improvements in machine translation.

One of the most popular datasets used to benchmark machine translation systems is the WMT family of datasets. Some of the most commonly used evaluation metrics for machine translation systems include BLEU, METEOR, NIST, and others.

( Image credit: Google seq2seq )

Papers

Showing 84018450 of 10752 papers

TitleStatusHype
Multi-level Evaluation for Machine Translation0
MT Tuning on RED: A Dependency-Based Evaluation Metric0
The Edinburgh/JHU Phrase-based Machine Translation Systems for WMT 20150
Edinburgh's Syntax-Based Systems at WMT 20150
Morphological Segmentation and OPUS for Finnish-English Machine Translation0
Montreal Neural Machine Translation Systems for WMT'150
An Investigation of Machine Translation Evaluation Metrics in Cross-lingual Question Answering0
Drem: The AFRL Submission to the WMT15 Tuning Task0
Machine Translation Evaluation using Recurrent Neural NetworksCode0
Strategy-Based Technology for Estimating MT Quality0
UAlacant word-level machine translation quality estimation system at WMT 20150
LORIA System for the WMT15 Quality Estimation Shared Task0
UdS-Sant: English--German Hybrid Machine Translation System0
Discontinuous Statistical Machine Translation with Target-Side Dependency Syntax0
ListNet-based MT Rescoring0
BEER 1.1: ILLC UvA submission to metrics and tuning taskCode0
LIMSI@WMT'15 : Translation Task0
DFKI's experimental hybrid MT system for WMT 20150
Abu-MaTran at WMT 2015 Translation Task: Morphological Segmentation and Web Crawling0
LeBLEU: N-gram-based Translation Evaluation Score for Morphologically Complex Languages0
Sheffield Systems for the Finnish-English WMT Translation Task0
Dependency Analysis of Scrambled References for Better Evaluation of Japanese Translation0
SHEF-NN: Translation Quality Estimation with Neural Networks0
The KIT-LIMSI Translation System for WMT 20150
Data Selection With Fewer Words0
Data enhancement and selection strategies for the word-level Quality Estimation0
The Karlsruhe Institute of Technology Translation Systems for the WMT 20150
Investigations on Phrase-based Decoding with Recurrent Neural Network Language and Translation Models0
The University of Illinois submission to the WMT 2015 Shared Translation Task0
CUNI in WMT15: Chimera Strikes Again0
VERTa: a Linguistically-motivated Metric at the WMT15 Metrics Task0
Improving evaluation and optimization of MT systems against MEANT0
USHEF and USAAR-USHEF participation in the WMT15 QE shared task0
How do Humans Evaluate Machine Translation0
Hierarchical Machine Translation With Discontinuous Phrases0
Connotation in Translation0
The AFRL-MITLL WMT15 System: There's More than One Way to Decode It!0
GF Wide-coverage English-Finnish MT system for WMT 20150
Semantic Tuples for Evaluation of Image to Sentence Generation0
Statistical Machine Translation with Automatic Identification of Translationese0
USAAR-SAPE: An English--Spanish Statistical Automatic Post-Editing System0
Reading metrics for estimating task efficiency with MT output0
Pronoun Translation and Prediction with or without Coreference Links0
Pronoun-Focused MT and Cross-Lingual Pronoun Prediction: Findings of the 2015 DiscoMT Shared Task on Pronoun Translation0
Proceedings of the Second Workshop on Discourse in Machine Translation0
Predicting Pronoun Translation Using Syntactic, Morphological and Contextual Features from Parallel Data0
Predicting Pronouns across Languages with Continuous Word Spaces0
Part-of-Speech Driven Cross-Lingual Pronoun Prediction with Feed-Forward Neural Networks0
A Proposal for a Coherence Corpus in Machine Translation0
Exploration of Inter- and Intralingual Variation of Discourse Phenomena0
Show:102550
← PrevPage 169 of 216Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Transformer Cycle (Rev)BLEU score35.14Unverified
2Noisy back-translationBLEU score35Unverified
3Transformer+Rep(Uni)BLEU score33.89Unverified
4T5-11BBLEU score32.1Unverified
5BiBERTBLEU score31.26Unverified
6Transformer + R-DropBLEU score30.91Unverified
7Bi-SimCutBLEU score30.78Unverified
8BERT-fused NMTBLEU score30.75Unverified
9Data Diversification - TransformerBLEU score30.7Unverified
10SimCutBLEU score30.56Unverified
#ModelMetricClaimedVerifiedStatus
1Transformer+BT (ADMIN init)BLEU score46.4Unverified
2Noisy back-translationBLEU score45.6Unverified
3mRASP+Fine-TuneBLEU score44.3Unverified
4Transformer + R-DropBLEU score43.95Unverified
5Transformer (ADMIN init)BLEU score43.8Unverified
6AdminBLEU score43.8Unverified
7BERT-fused NMTBLEU score43.78Unverified
8MUSE(Paralllel Multi-scale Attention)BLEU score43.5Unverified
9T5BLEU score43.4Unverified
10Local Joint Self-attentionBLEU score43.3Unverified
#ModelMetricClaimedVerifiedStatus
1PiNMTBLEU score40.43Unverified
2BiBERTBLEU score38.61Unverified
3Bi-SimCutBLEU score38.37Unverified
4Cutoff + Relaxed Attention + LMBLEU score37.96Unverified
5DRDABLEU score37.95Unverified
6Transformer + R-Drop + CutoffBLEU score37.9Unverified
7SimCutBLEU score37.81Unverified
8Cutoff+KneeBLEU score37.78Unverified
9CutoffBLEU score37.6Unverified
10CipherDAugBLEU score37.53Unverified
#ModelMetricClaimedVerifiedStatus
1HWTSC-Teacher-SimScore19.97Unverified
2MS-COMET-22Score19.89Unverified
3MS-COMET-QE-22Score19.76Unverified
4KG-BERTScoreScore17.28Unverified
5metricx_xl_DA_2019Score17.17Unverified
6COMET-QEScore16.8Unverified
7COMET-22Score16.31Unverified
8UniTE-srcScore15.68Unverified
9UniTE-refScore15.38Unverified
10metricx_xxl_DA_2019Score15.24Unverified