Incremental Domain Adaptation for Neural Machine Translation in Low-Resource Settings

Marimuthu Kalimuthu, Michael Barz, Daniel Sonntag

In: Proceedings of the Fourth Arabic Natural Language Processing Workshop. Arabic Natural Language Processing Workshop (WANLP-2019) 4th located at ACL August 1-2 Florence Italy Pages 1-10 Association for Computational Linguistics 8/2019.


We study the problem of incremental domain adaptation of a generic neural machine translation model with limited resources (e.g., budget and time) for human translations or model training. In this paper, we propose a novel query strategy for selecting {``}unlabeled{''} samples from a new domain based on sentence embeddings for Arabic. We accelerate the fine-tuning process of the generic model to the target domain. Specifically, our approach estimates the informativeness of instances from the target domain by comparing the distance of their sentence embeddings to embeddings from the generic domain. We perform machine translation experiments (Ar-to-En direction) for comparing a random sampling baseline with our new approach, similar to active learning, using two small update sets for simulating the work of human translators. For the prescribed setting we can save more than 50{\%} of the annotation costs without loss in quality, demonstrating the effectiveness of our approach.

W19-4601.pdf (pdf, 787 KB)

German Research Center for Artificial Intelligence
Deutsches Forschungszentrum für Künstliche Intelligenz