SOTAVerified

What do you mean, BERT? Assessing BERT as a Distributional Semantics Model

2019-11-13Unverified0· sign in to hype

Timothee Mickus, Denis Paperno, Mathieu Constant, Kees Van Deemter

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Contextualized word embeddings, i.e. vector representations for words in context, are naturally seen as an extension of previous noncontextual distributional semantic models. In this work, we focus on BERT, a deep neural network that produces contextualized embeddings and has set the state-of-the-art in several semantic tasks, and study the semantic coherence of its embedding space. While showing a tendency towards coherence, BERT does not fully live up to the natural expectations for a semantic vector space. In particular, we find that the position of the sentence in which a word occurs, while having no meaning correlates, leaves a noticeable trace on the word embeddings and disturbs similarity relationships.

Tasks

Reproductions