SOTAVerified

Specializing Word Vectors by Spectral Decomposition on Heterogeneously Twisted Graphs

2020-12-01COLING 2020Code Available0· sign in to hype

Yuanhang Ren, Ye Du

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Traditional word vectors, such as word2vec and glove, have a well-known inclination to conflate the semantic similarity with other semantic relations. A retrofitting procedure may be needed to solve this issue. In this work, we propose a new retrofitting method called Heterogeneously Retrofitted Spectral Word Embedding. It heterogeneously twists the similarity matrix of word pairs with lexical constraints. A new set of word vectors is generated by a spectral decomposition of the similarity matrix, which has a linear algebraic analytic form. Our method has a competitive performance compared with the state-of-the-art retrofitting method such as AR (CITATION). In addition, since our embedding has a clear linear algebraic relationship with the similarity matrix, we carefully study the contribution of each component in our model. Last but not least, our method is very efficient to execute.

Tasks

Reproductions