Explaining Word Embeddings via Disentangled Representation
2020-12-01Asian Chapter of the Association for Computational LinguisticsUnverified0· sign in to hype
Keng-Te Liao, Cheng-Syuan Lee, Zhong-Yu Huang, Shou-De Lin
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Disentangled representations have attracted increasing attention recently. However, how to transfer the desired properties of disentanglement to word representations is unclear. In this work, we propose to transform typical dense word vectors into disentangled embeddings featuring improved interpretability via encoding polysemous semantics separately. We also found the modular structure of our disentangled word embeddings helps generate more efficient and effective features for natural language processing tasks.