SOTAVerified

Language Modeling with Graph Temporal Convolutional Networks

2019-05-01ICLR 2019Unverified0· sign in to hype

Hongyin Luo, Yichen Li, Jie Fu, James Glass

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Recently, there have been some attempts to use non-recurrent neural models for language modeling. However, a noticeable performance gap still remains. We propose a non-recurrent neural language model, dubbed graph temporal convolutional network (GTCN), that relies on graph neural network blocks and convolution operations. While the standard recurrent neural network language models encode sentences sequentially without modeling higher-level structural information, our model regards sentences as graphs and processes input words within a message propagation framework, aiming to learn better syntactic information by inferring skip-word connections. Specifically, the graph network blocks operate in parallel and learn the underlying graph structures in sentences without any additional annotation pertaining to structure knowledge. Experiments demonstrate that the model without recurrence can achieve comparable perplexity results in language modeling tasks and successfully learn syntactic information.

Tasks

Reproductions