SOTAVerified

CUNI System for the WMT18 Multimodal Translation Task

2018-10-01WS 2018Unverified0· sign in to hype

Jind{\v{r}}ich Helcl, Jind{\v{r}}ich Libovick{\'y}, Du{\v{s}}an Vari{\v{s}}

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We present our submission to the WMT18 Multimodal Translation Task. The main feature of our submission is applying a self-attentive network instead of a recurrent neural network. We evaluate two methods of incorporating the visual features in the model: first, we include the image representation as another input to the network; second, we train the model to predict the visual features and use it as an auxiliary objective. For our submission, we acquired both textual and multimodal additional data. Both of the proposed methods yield significant improvements over recurrent networks and self-attentive textual baselines.

Tasks

Reproductions