SOTAVerified

Improving Zero-Shot Multilingual Text Generation via Iterative Distillation

2022-10-01COLING 2022Unverified0· sign in to hype

Ernie Chang, Alex Marin, Vera Demberg

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

The demand for multilingual dialogue systems often requires a costly labeling process, where human translators derive utterances in low resource languages from resource rich language annotation. To this end, we explore leveraging the inductive biases for target languages learned by numerous pretrained teacher models by transferring them to student models via sequence-level knowledge distillation. By assuming no target language text, the both the teacher and student models need to learn from the target distribution in a few/zero-shot manner. On the MultiATIS++ benchmark, we explore the effectiveness of our proposed technique to derive the multilingual text for 6 languages, using only the monolingual English data and the pretrained models. We show that training on the synthetic multilingual generation outputs yields close performance to training on human annotations in both slot F1 and intent accuracy; the synthetic text also scores high in naturalness and correctness based on human evaluation.

Tasks

Reproductions