SOTAVerified

Put It Back: Entity Typing with Language Model Enhancement

2018-10-01EMNLP 2018Code Available0· sign in to hype

Ji Xin, Hao Zhu, Xu Han, Zhiyuan Liu, Maosong Sun

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Entity typing aims to classify semantic types of an entity mention in a specific context. Most existing models obtain training data using distant supervision, and inevitably suffer from the problem of noisy labels. To address this issue, we propose entity typing with language model enhancement. It utilizes a language model to measure the compatibility between context sentences and labels, and thereby automatically focuses more on context-dependent labels. Experiments on benchmark datasets demonstrate that our method is capable of enhancing the entity typing model with information from the language model, and significantly outperforms the state-of-the-art baseline. Code and data for this paper can be found from https://github.com/thunlp/LME.

Tasks

Reproductions