Exposing the Functionalities of Neurons for Gated Recurrent Unit Based Sequence-to-Sequence Model
2023-03-27Unverified0· sign in to hype
Yi-Ting Lee, Da-Yi Wu, Chih-Chun Yang, Shou-De Lin
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
The goal of this paper is to report certain scientific discoveries about a Seq2Seq model. It is known that analyzing the behavior of RNN-based models at the neuron level is considered a more challenging task than analyzing a DNN or CNN models due to their recursive mechanism in nature. This paper aims to provide neuron-level analysis to explain why a vanilla GRU-based Seq2Seq model without attention can achieve token-positioning. We found four different types of neurons: storing, counting, triggering, and outputting and further uncover the mechanism for these neurons to work together in order to produce the right token in the right position.