SOTAVerified

Language Production Dynamics with Recurrent Neural Networks

2018-07-01WS 2018Unverified0· sign in to hype

Jes{\'u}s Calvillo, Matthew Crocker

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We present an analysis of the internal mechanism of the recurrent neural model of sentence production presented by Calvillo et al. (2016). The results show clear patterns of computation related to each layer in the network allowing to infer an algorithmic account, where the semantics activates the semantically related words, then each word generated at each time step activates syntactic and semantic constraints on possible continuations, while the recurrence preserves information through time. We propose that such insights could generalize to other models with similar architecture, including some used in computational linguistics for language modeling, machine translation and image caption generation.

Tasks

Reproductions