Generating News Headlines with Recurrent Neural Networks
2015-12-05Code Available0· sign in to hype
Konstantin Lopyrev
Code Available — Be the first to reproduce this paper.
ReproduceCode
Abstract
We describe an application of an encoder-decoder recurrent neural network with LSTM units and attention to generating headlines from the text of news articles. We find that the model is quite effective at concisely paraphrasing news articles. Furthermore, we study how the neural network decides which input words to pay attention to, and specifically we identify the function of the different neurons in a simplified attention mechanism. Interestingly, our simplified attention mechanism performs better that the more complex attention mechanism on a held out set of articles.