SOTAVerified

Zero Pronoun Resolution with Attention-based Neural Network

2018-08-01COLING 2018Code Available0· sign in to hype

Qingyu Yin, Yu Zhang, Wei-Nan Zhang, Ting Liu, William Yang Wang

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Recent neural network methods for zero pronoun resolution explore multiple models for generating representation vectors for zero pronouns and their candidate antecedents. Typically, contextual information is utilized to encode the zero pronouns since they are simply gaps that contain no actual content. To better utilize contexts of the zero pronouns, we here introduce the self-attention mechanism for encoding zero pronouns. With the help of the multiple hops of attention, our model is able to focus on some informative parts of the associated texts and therefore produces an efficient way of encoding the zero pronouns. In addition, an attention-based recurrent neural network is proposed for encoding candidate antecedents by their contents. Experiment results are encouraging: our proposed attention-based model gains the best performance on the Chinese portion of the OntoNotes corpus, substantially surpasses existing Chinese zero pronoun resolution baseline systems.

Tasks

Reproductions