SOTAVerified

Towards Robust Named Entity Recognition for Historic German

2019-06-18WS 2019Code Available0· sign in to hype

Stefan Schweter, Johannes Baiter

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Recent advances in language modeling using deep neural networks have shown that these models learn representations, that vary with the network depth from morphology to semantic relationships like co-reference. We apply pre-trained language models to low-resource named entity recognition for Historic German. We show on a series of experiments that character-based pre-trained language models do not run into trouble when faced with low-resource datasets. Our pre-trained character-based language models improve upon classical CRF-based methods and previous work on Bi-LSTMs by boosting F1 score performance by up to 6%. Our pre-trained language and NER models are publicly available under https://github.com/stefan-it/historic-ner .

Tasks

Reproductions