SOTAVerified

Sense Vocabulary Compression through the Semantic Knowledge of WordNet for Neural Word Sense Disambiguation

2019-05-14GWC 2019Code Available0· sign in to hype

Loïc Vial, Benjamin Lecouteux, Didier Schwab

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In this article, we tackle the issue of the limited quantity of manually sense annotated corpora for the task of word sense disambiguation, by exploiting the semantic relationships between senses such as synonymy, hypernymy and hyponymy, in order to compress the sense vocabulary of Princeton WordNet, and thus reduce the number of different sense tags that must be observed to disambiguate all words of the lexical database. We propose two different methods that greatly reduces the size of neural WSD models, with the benefit of improving their coverage without additional training data, and without impacting their precision. In addition to our method, we present a WSD system which relies on pre-trained BERT word vectors in order to achieve results that significantly outperform the state of the art on all WSD evaluation tasks.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
SemEval 2007 Task 17SemCor+WNGC, hypernymsF173.4Unverified
SemEval 2007 Task 7SemCor+WNGC, hypernymsF190.4Unverified
SemEval 2013 Task 12SemCor+WNGC, hypernymsF178.7Unverified
SemEval 2015 Task 13SemCor+WNGC, hypernymsF182.6Unverified
Senseval-2SemCor+WNGC, hypernymsF179.7Unverified
SensEval 3 Task 1SemCor+WNGC, hypernymsF177.8Unverified
Supervised:SemCor+WNGC, hypernymsSenseval 279.7Unverified

Reproductions