SOTAVerified

On Learning Word Embeddings From Linguistically Augmented Text Corpora

2019-05-01WS 2019Unverified0· sign in to hype

Amila Silva, Chathurika Amarathunga

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Word embedding is a technique in Natural Language Processing (NLP) to map words into vector space representations. Since it has boosted the performance of many NLP downstream tasks, the task of learning word embeddings has been addressing significantly. Nevertheless, most of the underlying word embedding methods such as word2vec and GloVe fail to produce high-quality embeddings if the text corpus is small and sparse. This paper proposes a method to generate effective word embeddings from limited data. Through experiments, we show that our proposed model outperforms existing works for the classical word similarity task and for a domain-specific application.

Tasks

Reproductions