SOTAVerified

Unsupervised Learning of Entailment-Vector Word Embeddings

2018-01-01ICLR 2018Unverified0· sign in to hype

James Henderson

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Entailment vectors are a principled way to encode in a vector what information is known and what is unknown. They are designed to model relations where one vector should include all the information in another vector, called entailment. This paper investigates the unsupervised learning of entailment vectors for the semantics of words. Using simple entailment-based models of the semantics of words in text (distributional semantics), we induce entailment-vector word embeddings which outperform the best previous results for predicting entailment between words, in unsupervised and semi-supervised experiments on hyponymy.

Tasks

Reproductions