SOTAVerified

Integrating Encyclopedic Knowledge into Neural Language Models

2016-12-01IWSLT 2016Unverified0· sign in to hype

Yang Zhang, Jan Niehues, Alexander Waibel

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Neural models have recently shown big improvements in the performance of phrase-based machine translation. Recurrent language models, in particular, have been a great success due to their ability to model arbitrary long context. In this work, we integrate global semantic information extracted from large encyclopedic sources into neural network language models. We integrate semantic word classes extracted from Wikipedia and sentence level topic information into a recurrent neural network-based language model. The new resulting models exhibit great potential in alleviating data sparsity problems with the additional knowledge provided. This approach of integrating global information is not restricted to language modeling but can also be easily applied to any model that profits from context or further data resources, e.g. neural machine translation. Using this model has improved rescoring quality of a state-of-the-art phrase-based translation system by 0.84 BLEU points. We performed experiments on two language pairs.

Tasks

Reproductions