SOTAVerified

Machine Translation

Machine translation is the task of translating a sentence in a source language to a different target language.

Approaches for machine translation can range from rule-based to statistical to neural-based. More recently, encoder-decoder attention-based architectures like BERT have attained major improvements in machine translation.

One of the most popular datasets used to benchmark machine translation systems is the WMT family of datasets. Some of the most commonly used evaluation metrics for machine translation systems include BLEU, METEOR, NIST, and others.

( Image credit: Google seq2seq )

Papers

Showing 1045110500 of 10752 papers

TitleStatusHype
UPM system for WMT 20120
The CMU-Avenue French-English Translation System0
Using Categorial Grammar to Label Translation Rules0
Constructing Parallel Corpora for Six Indian Languages via Crowdsourcing0
SPEDE: Probabilistic Edit Distance Metrics for MT Evaluation0
Comparing human perceptions of post-editing effort with post-editing operations0
DFKI's SMT System for WMT 20120
LORIA System for the WMT12 Quality Estimation Shared Task0
Proceedings of the Seventh Workshop on Statistical Machine Translation0
PRHLT Submission to the WMT12 Quality Estimation Task0
Match without a Referee: Evaluating MT Adequacy without Reference Translations0
DCU-Symantec Submission for the WMT 2012 Quality Estimation Task0
QCRI at WMT12: Experiments in Spanish-English and German-English Machine Translation of News Text0
Selecting Data for English-to-Czech Machine Translation0
Probes in a Taxonomy of Factored Phrase-Based Models0
Evaluating the Learning Curve of Domain Adaptive Statistical Machine Translation Systems0
Leave-One-Out Phrase Model Training for Large-Scale Deployment0
CCG Syntactic Reordering Models for Phrase-based Machine Translation0
On Hierarchical Re-ordering and Permutation Parsing for Phrase-based Decoding0
Data Issues of the Multilingual Translation Matrix0
Formemes in English-Czech Deep Syntactic MT0
Kriya - The SFU System for Translation Task at WMT-120
Findings of the 2012 Workshop on Statistical Machine Translation0
Linguistic Features for Quality Estimation0
LIMSI @ WMT120
Quality estimation for Machine Translation output using linguistic analysis and decoding features0
PROMT DeepHybrid system for WMT12 shared translation task0
Semantic Textual Similarity for MT evaluation0
Class error rates for evaluation of machine translation output0
Phrase Model Training for Statistical Machine Translation with Word Lattices of Preprocessing Alternatives0
Improving AMBER, an MT Evaluation Metric0
GHKM Rule Extraction and Scope-3 Parsing in Moses0
Combining Quality Prediction and System Selection for Improved Automatic Translation Output0
Joshua 4.0: Packing, PRO, and Paraphrases0
Analysing the Effect of Out-of-Domain Data on SMT Systems0
Joint WMT 2012 Submission of the QUAERO Project0
Fully Automatic Semantic MT Evaluation0
Morpheme- and POS-based IBM1 and language model scores for translation quality estimation0
LIUM's SMT Machine Translation Systems for WMT 20120
Black Box Features for the WMT 2012 Quality Estimation Shared Task0
Quality Estimation: an experimental study using unsupervised similarity measures0
Machine Learning for Hybrid Machine Translation0
Direct Error Rate Minimization for Statistical Machine Translation0
Review of Hypothesis Alignment Algorithms for MT System Combination via Confusion Network Decoding0
Non-Linear Models for Confidence Estimation0
Regression with Phrase Indicators for Estimating MT Quality0
DEPFIX: A System for Automatic Correction of Czech MT Outputs0
Putting Human Assessments of Machine Translation Systems in Order0
Optimization Strategies for Online Large-Margin Learning in Machine Translation0
PREFER: Using a Graph-Based Approach to Generate Paraphrases for Language Learning0
Show:102550
← PrevPage 210 of 216Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Transformer Cycle (Rev)BLEU score35.14Unverified
2Noisy back-translationBLEU score35Unverified
3Transformer+Rep(Uni)BLEU score33.89Unverified
4T5-11BBLEU score32.1Unverified
5BiBERTBLEU score31.26Unverified
6Transformer + R-DropBLEU score30.91Unverified
7Bi-SimCutBLEU score30.78Unverified
8BERT-fused NMTBLEU score30.75Unverified
9Data Diversification - TransformerBLEU score30.7Unverified
10SimCutBLEU score30.56Unverified
#ModelMetricClaimedVerifiedStatus
1Transformer+BT (ADMIN init)BLEU score46.4Unverified
2Noisy back-translationBLEU score45.6Unverified
3mRASP+Fine-TuneBLEU score44.3Unverified
4Transformer + R-DropBLEU score43.95Unverified
5Transformer (ADMIN init)BLEU score43.8Unverified
6AdminBLEU score43.8Unverified
7BERT-fused NMTBLEU score43.78Unverified
8MUSE(Paralllel Multi-scale Attention)BLEU score43.5Unverified
9T5BLEU score43.4Unverified
10Local Joint Self-attentionBLEU score43.3Unverified
#ModelMetricClaimedVerifiedStatus
1PiNMTBLEU score40.43Unverified
2BiBERTBLEU score38.61Unverified
3Bi-SimCutBLEU score38.37Unverified
4Cutoff + Relaxed Attention + LMBLEU score37.96Unverified
5DRDABLEU score37.95Unverified
6Transformer + R-Drop + CutoffBLEU score37.9Unverified
7SimCutBLEU score37.81Unverified
8Cutoff+KneeBLEU score37.78Unverified
9CutoffBLEU score37.6Unverified
10CipherDAugBLEU score37.53Unverified
#ModelMetricClaimedVerifiedStatus
1HWTSC-Teacher-SimScore19.97Unverified
2MS-COMET-22Score19.89Unverified
3MS-COMET-QE-22Score19.76Unverified
4KG-BERTScoreScore17.28Unverified
5metricx_xl_DA_2019Score17.17Unverified
6COMET-QEScore16.8Unverified
7COMET-22Score16.31Unverified
8UniTE-srcScore15.68Unverified
9UniTE-refScore15.38Unverified
10metricx_xxl_DA_2019Score15.24Unverified