SOTAVerified

Syntax Encoding with Application in Authorship Attribution

2018-10-01EMNLP 2018Unverified0· sign in to hype

Richong Zhang, Zhiyuan Hu, Hongyu Guo, Yongyi Mao

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We propose a novel strategy to encode the syntax parse tree of sentence into a learnable distributed representation. The proposed syntax encoding scheme is provably information-lossless. In specific, an embedding vector is constructed for each word in the sentence, encoding the path in the syntax tree corresponding to the word. The one-to-one correspondence between these ``syntax-embedding'' vectors and the words (hence their embedding vectors) in the sentence makes it easy to integrate such a representation with all word-level NLP models. We empirically show the benefits of the syntax embeddings on the Authorship Attribution domain, where our approach improves upon the prior art and achieves new performance records on five benchmarking data sets.

Tasks

Reproductions