SOTAVerified

Improving Cuneiform Language Identification with BERT

2019-06-01WS 2019Unverified0· sign in to hype

Gabriel Bernier-Colborne, Cyril Goutte, Serge L{\'e}ger

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We describe the systems developed by the National Research Council Canada for the Cuneiform Language Identification (CLI) shared task at the 2019 VarDial evaluation campaign. We compare a state-of-the-art baseline relying on character n-grams and a traditional statistical classifier, a voting ensemble of classifiers, and a deep learning approach using a Transformer network. We describe how these systems were trained, and analyze the impact of some preprocessing and model estimation decisions. The deep neural network achieved 77\% accuracy on the test data, which turned out to be the best performance at the CLI evaluation, establishing a new state-of-the-art for cuneiform language identification.

Tasks

Reproductions