TY - GEN
T1 - Adaptive Machine Translation with Large Language Models
AU - Moslem, Yasmin
AU - Haque, Rejwanul
AU - Kelleher, John D.
AU - Way, Andy
N1 - Publisher Copyright:
© 2023 The authors. This article is licensed under a Creative Commons 4.0 licence, no derivative works, attribution, CC-BY-ND.
PY - 2023
Y1 - 2023
N2 - Consistency is a key requirement of high-quality translation. It is especially important to adhere to pre-approved terminology and adapt to corrected translations in domain-specific projects. Machine translation (MT) has achieved significant progress in the area of domain adaptation. However, real-time adaptation remains challenging. Large-scale language models (LLMs) have recently shown interesting capabilities of in-context learning, where they learn to replicate certain input-output text generation patterns, without further fine-tuning. By feeding an LLM at inference time with a prompt that consists of a list of translation pairs, it can then simulate the domain and style characteristics. This work aims to investigate how we can utilize in-context learning to improve real-time adaptive MT. Our extensive experiments show promising results at translation time. For example, LLMs can adapt to a set of in-domain sentence pairs and/or terminology while translating a new sentence. We observe that the translation quality with few-shot in-context learning can surpass that of strong encoder-decoder MT systems, especially for high-resource languages. Moreover, we investigate whether we can combine MT from strong encoder-decoder models with fuzzy matches, which can further improve translation quality, especially for less supported languages. We conduct our experiments across five diverse language pairs, namely English-to-Arabic (EN-AR), English-to-Chinese (EN-ZH), English-to-French (EN-FR), English-to-Kinyarwanda (EN-RW), and English-to-Spanish (EN-ES).
AB - Consistency is a key requirement of high-quality translation. It is especially important to adhere to pre-approved terminology and adapt to corrected translations in domain-specific projects. Machine translation (MT) has achieved significant progress in the area of domain adaptation. However, real-time adaptation remains challenging. Large-scale language models (LLMs) have recently shown interesting capabilities of in-context learning, where they learn to replicate certain input-output text generation patterns, without further fine-tuning. By feeding an LLM at inference time with a prompt that consists of a list of translation pairs, it can then simulate the domain and style characteristics. This work aims to investigate how we can utilize in-context learning to improve real-time adaptive MT. Our extensive experiments show promising results at translation time. For example, LLMs can adapt to a set of in-domain sentence pairs and/or terminology while translating a new sentence. We observe that the translation quality with few-shot in-context learning can surpass that of strong encoder-decoder MT systems, especially for high-resource languages. Moreover, we investigate whether we can combine MT from strong encoder-decoder models with fuzzy matches, which can further improve translation quality, especially for less supported languages. We conduct our experiments across five diverse language pairs, namely English-to-Arabic (EN-AR), English-to-Chinese (EN-ZH), English-to-French (EN-FR), English-to-Kinyarwanda (EN-RW), and English-to-Spanish (EN-ES).
UR - http://www.scopus.com/inward/record.url?scp=85172366127&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85172366127
T3 - Proceedings of the 24th Annual Conference of the European Association for Machine Translation, EAMT 2023
SP - 227
EP - 237
BT - Proceedings of the 24th Annual Conference of the European Association for Machine Translation, EAMT 2023
A2 - Nurminen, Mary
A2 - Nurminen, Mary
A2 - Brenner, Judith
A2 - Koponen, Maarit
A2 - Latomaa, Sirkku
A2 - Mikhailov, Mikhail
A2 - Schierl, Frederike
A2 - Ranasinghe, Tharindu
A2 - Vanmassenhove, Eva
A2 - Vidal, Sergi Alvarez
A2 - Aranberri, Nora
A2 - Nunziatini, Mara
A2 - Escartin, Carla Parra
A2 - Forcada, Mikel
A2 - Popovic, Maja
A2 - Scarton, Carolina
A2 - Moniz, Helena
PB - European Association for Machine Translation
T2 - 24th Annual Conference of the European Association for Machine Translation, EAMT 2023
Y2 - 12 June 2023 through 15 June 2023
ER -