CogALex-VI Shared Task: Transrelation - A Robust Multilingual Language Model for Multilingual Relation Identification
2020-12-12Code Available0· sign in to hype
Lennart Wachowiak, Christian Lang, Barbara Heinisch, Dagmar Gromann
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/Text2TCS/TransrelationOfficialnone★ 2
Abstract
We describe our submission to the CogALex-VI shared task on the identification of multilingual paradigmatic relations building on XLM-RoBERTa (XLM-R), a robustly optimized and multilingual BERT model. In spite of several experiments with data augmentation, data addition and ensemble methods with a Siamese Triple Net, Translrelation, the XLM-R model with a linear classifier adapted to this specific task, performed best in testing and achieved the best results in the final evaluation of the shared task, even for a previously unseen language.