Neural Sequence Model Training via α-divergence Minimization
2017-06-30Code Available0· sign in to hype
Sotetsu Koyamada, Yuta Kikuchi, Atsunori Kanemura, Shin-ichi Maeda, Shin Ishii
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/sotetsuk/alpha-dimt-icmlwsOfficialIn paperpytorch★ 0
Abstract
We propose a new neural sequence model training method in which the objective function is defined by -divergence. We demonstrate that the objective function generalizes the maximum-likelihood (ML)-based and reinforcement learning (RL)-based objective functions as special cases (i.e., ML corresponds to 0 and RL to 1). We also show that the gradient of the objective function can be considered a mixture of ML- and RL-based objective gradients. The experimental results of a machine translation task show that minimizing the objective function with > 0 outperforms 0, which corresponds to ML-based methods.