SOTAVerified

LSTM based Conversation Models

2016-03-31Code Available0· sign in to hype

Yi Luan, Yangfeng Ji, Mari Ostendorf

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In this paper, we present a conversational model that incorporates both context and participant role for two-party conversations. Different architectures are explored for integrating participant role and context information into a Long Short-term Memory (LSTM) language model. The conversational model can function as a language model or a language generation model. Experiments on the Ubuntu Dialog Corpus show that our model can capture multiple turn interaction between participants. The proposed method outperforms a traditional LSTM model as measured by language model perplexity and response ranking. Generated responses show characteristic differences between the two participant roles.

Tasks

Reproductions