SOTAVerified

All-but-the-Top: Simple and Effective Postprocessing for Word Representations

2017-02-05ICLR 2018Code Available0· sign in to hype

Jiaqi Mu, Suma Bhat, Pramod Viswanath

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Real-valued word representations have transformed NLP applications; popular examples are word2vec and GloVe, recognized for their ability to capture linguistic regularities. In this paper, we demonstrate a very simple, and yet counter-intuitive, postprocessing technique -- eliminate the common mean vector and a few top dominating directions from the word vectors -- that renders off-the-shelf representations even stronger. The postprocessing is empirically validated on a variety of lexical-level intrinsic tasks (word similarity, concept categorization, word analogy) and sentence-level tasks (semantic textural similarity and text classification) on multiple datasets and with a variety of representation methods and hyperparameter choices in multiple languages; in each case, the processed representations are consistently better than the original ones.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
MRGRU-RNN-WORD2VECAccuracy78.26Unverified
SST-5 Fine-grained classificationGRU-RNN-WORD2VECAccuracy45.02Unverified

Reproductions