SOTAVerified

Parallel sentences mining with transfer learning in an unsupervised setting

2021-06-01NAACL 2021Unverified0· sign in to hype

Yu Sun, Shaolin Zhu, Feng Yifan, Chenggang Mi

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

The quality and quantity of parallel sentences are known as very important training data for constructing neural machine translation (NMT) systems. However, these resources are not available for many low-resource language pairs. Many existing methods need strong supervision are not suitable. Although several attempts at developing unsupervised models, they ignore the language-invariant between languages. In this paper, we propose an approach based on transfer learning to mine parallel sentences in the unsupervised setting.With the help of bilingual corpora of rich-resource language pairs, we can mine parallel sentences without bilingual supervision of low-resource language pairs. Experiments show that our approach improves the performance of mined parallel sentences compared with previous methods. In particular, we achieve excellent results at two real-world low-resource language pairs.

Tasks

Reproductions