SOTAVerified

To Share or not to Share: Predicting Sets of Sources for Model Transfer Learning

2021-04-16EMNLP 2021Code Available0· sign in to hype

Lukas Lange, Jannik Strötgen, Heike Adel, Dietrich Klakow

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In low-resource settings, model transfer can help to overcome a lack of labeled data for many tasks and domains. However, predicting useful transfer sources is a challenging problem, as even the most similar sources might lead to unexpected negative transfer results. Thus, ranking methods based on task and text similarity -- as suggested in prior work -- may not be sufficient to identify promising sources. To tackle this problem, we propose a new approach to automatically determine which and how many sources should be exploited. For this, we study the effects of model transfer on sequence labeling across various domains and tasks and show that our methods based on model similarity and support vector machines are able to predict promising sources, resulting in performance increases of up to 24 F1 points.

Tasks

Reproductions