arXiv Analytics

Sign in

arXiv:2005.04816 [cs.CL]AbstractReferencesReviewsResources

Leveraging Monolingual Data with Self-Supervision for Multilingual Neural Machine Translation

Aditya Siddhant, Ankur Bapna, Yuan Cao, Orhan Firat, Mia Chen, Sneha Kudugunta, Naveen Arivazhagan, Yonghui Wu

Published 2020-05-11Version 1

Over the last few years two promising research directions in low-resource neural machine translation (NMT) have emerged. The first focuses on utilizing high-resource languages to improve the quality of low-resource languages via multilingual NMT. The second direction employs monolingual data with self-supervision to pre-train translation models, followed by fine-tuning on small amounts of supervised data. In this work, we join these two lines of research and demonstrate the efficacy of monolingual data with self-supervision in multilingual NMT. We offer three major results: (i) Using monolingual data significantly boosts the translation quality of low-resource languages in multilingual models. (ii) Self-supervision improves zero-shot translation quality in multilingual models. (iii) Leveraging monolingual data with self-supervision provides a viable path towards adding new languages to multilingual models, getting up to 33 BLEU on ro-en translation without any parallel data or back-translation.

Related articles: Most relevant | Search more
arXiv:2010.02523 [cs.CL] (Published 2020-10-06)
Multi-task Learning for Multilingual Neural Machine Translation
arXiv:1908.09324 [cs.CL] (Published 2019-08-25)
Multilingual Neural Machine Translation with Language Clustering
arXiv:2207.04906 [cs.CL] (Published 2022-07-11)
High-resource Language-specific Training for Multilingual Neural Machine Translation