Revisiting the Weaknesses of Reinforcement Learning for Neural Machine Translation
2021-06-16NAACL 2021Code Available1· sign in to hype
Samuel Kiegeland, Julia Kreutzer
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/samuki/reinforce-joeyOfficialIn paperpytorch★ 27
Abstract
Policy gradient algorithms have found wide adoption in NLP, but have recently become subject to criticism, doubting their suitability for NMT. Choshen et al. (2020) identify multiple weaknesses and suspect that their success is determined by the shape of output distributions rather than the reward. In this paper, we revisit these claims and study them under a wider range of configurations. Our experiments on in-domain and cross-domain adaptation reveal the importance of exploration and reward scaling, and provide empirical counter-evidence to these claims.