SOTAVerified

Integrating Semantic Knowledge into Lexical Embeddings Based on Information Content Measurement

2017-04-01EACL 2017Code Available0· sign in to hype

Hsin-Yang Wang, Wei-Yun Ma

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Distributional word representations are widely used in NLP tasks. These representations are based on an assumption that words with a similar context tend to have a similar meaning. To improve the quality of the context-based embeddings, many researches have explored how to make full use of existing lexical resources. In this paper, we argue that while we incorporate the prior knowledge with context-based embeddings, words with different occurrences should be treated differently. Therefore, we propose to rely on the measurement of information content to control the degree of applying prior knowledge into context-based embeddings - different words would have different learning rates when adjusting their embeddings. In the result, we demonstrate that our embeddings get significant improvements on two different tasks: Word Similarity and Analogical Reasoning.

Tasks

Reproductions