Sequence Modeling with Unconstrained Generation Order
2019-11-01NeurIPS 2019Code Available0· sign in to hype
Dmitrii Emelianenko, Elena Voita, Pavel Serdyukov
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/TIXFeniks/neurips2019_intrusOfficialIn papertf★ 0
Abstract
The dominant approach to sequence generation is to produce a sequence in some predefined order, e.g. left to right. In contrast, we propose a more general model that can generate the output sequence by inserting tokens in any arbitrary order. Our model learns decoding order as a result of its training procedure. Our experiments show that this model is superior to fixed order models on a number of sequence generation tasks, such as Machine Translation, Image-to-LaTeX and Image Captioning.