SOTAVerified

LLMs Are Zero-Shot Context-Aware Simultaneous Translators

2024-06-19Code Available1· sign in to hype

Roman Koshkin, Katsuhito Sudoh, Satoshi Nakamura

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

The advent of transformers has fueled progress in machine translation. More recently large language models (LLMs) have come to the spotlight thanks to their generality and strong performance in a wide range of language tasks, including translation. Here we show that open-source LLMs perform on par with or better than some state-of-the-art baselines in simultaneous machine translation (SiMT) tasks, zero-shot. We also demonstrate that injection of minimal background information, which is easy with an LLM, brings further performance gains, especially on challenging technical subject-matter. This highlights LLMs' potential for building next generation of massively multilingual, context-aware and terminologically accurate SiMT systems that require no resource-intensive training or fine-tuning.

Tasks

Reproductions