Adaptive Machine Translation with Large Language Models

Yasmin Moslem, Rejwanul Haque, John D. Kelleher, Andy Way

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

23 Citations (Scopus)

Abstract

Consistency is a key requirement of high-quality translation. It is especially important to adhere to pre-approved terminology and adapt to corrected translations in domain-specific projects. Machine translation (MT) has achieved significant progress in the area of domain adaptation. However, real-time adaptation remains challenging. Large-scale language models (LLMs) have recently shown interesting capabilities of in-context learning, where they learn to replicate certain input-output text generation patterns, without further fine-tuning. By feeding an LLM at inference time with a prompt that consists of a list of translation pairs, it can then simulate the domain and style characteristics. This work aims to investigate how we can utilize in-context learning to improve real-time adaptive MT. Our extensive experiments show promising results at translation time. For example, LLMs can adapt to a set of in-domain sentence pairs and/or terminology while translating a new sentence. We observe that the translation quality with few-shot in-context learning can surpass that of strong encoder-decoder MT systems, especially for high-resource languages. Moreover, we investigate whether we can combine MT from strong encoder-decoder models with fuzzy matches, which can further improve translation quality, especially for less supported languages. We conduct our experiments across five diverse language pairs, namely English-to-Arabic (EN-AR), English-to-Chinese (EN-ZH), English-to-French (EN-FR), English-to-Kinyarwanda (EN-RW), and English-to-Spanish (EN-ES).

Original languageEnglish
Title of host publicationProceedings of the 24th Annual Conference of the European Association for Machine Translation, EAMT 2023
EditorsMary Nurminen, Mary Nurminen, Judith Brenner, Maarit Koponen, Sirkku Latomaa, Mikhail Mikhailov, Frederike Schierl, Tharindu Ranasinghe, Eva Vanmassenhove, Sergi Alvarez Vidal, Nora Aranberri, Mara Nunziatini, Carla Parra Escartin, Mikel Forcada, Maja Popovic, Carolina Scarton, Helena Moniz
PublisherEuropean Association for Machine Translation
Pages227-237
Number of pages11
ISBN (Electronic)9789520329471
Publication statusPublished - 2023
Event24th Annual Conference of the European Association for Machine Translation, EAMT 2023 - Tampere, Finland
Duration: 12 Jun 202315 Jun 2023

Publication series

NameProceedings of the 24th Annual Conference of the European Association for Machine Translation, EAMT 2023

Conference

Conference24th Annual Conference of the European Association for Machine Translation, EAMT 2023
Country/TerritoryFinland
CityTampere
Period12/06/202315/06/2023

Fingerprint

Dive into the research topics of 'Adaptive Machine Translation with Large Language Models'. Together they form a unique fingerprint.

Cite this