SOTAVerified

Multidomain Pretrained Language Models for Green NLP

2021-04-01EACL (AdaptNLP) 2021Code Available0· sign in to hype

Antonis Maronikolakis, Hinrich Schütze

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

When tackling a task in a given domain, it has been shown that adapting a model to the domain using raw text data before training on the supervised task improves performance versus solely training on the task. The downside is that a lot of domain data is required and if we want to tackle tasks in n domains, we require n models each adapted on domain data before task learning. Storing and using these models separately can be prohibitive for low-end devices. In this paper we show that domain adaptation can be generalised to cover multiple domains. Specifically, a single model can be trained across various domains at the same time with minimal drop in performance, even when we use less data and resources. Thus, instead of training multiple models, we can train a single multidomain model saving on computational resources and training time.

Tasks

Reproductions