arXiv Analytics

Sign in

arXiv:1601.01073 [cs.CL]AbstractReferencesReviewsResources

Multi-Way, Multilingual Neural Machine Translation with a Shared Attention Mechanism

Orhan Firat, Kyunghyun Cho, Yoshua Bengio

Published 2016-01-06Version 1

We propose multi-way, multilingual neural machine translation. The proposed approach enables a single neural translation model to translate between multiple languages, with a number of parameters that grows only linearly with the number of languages. This is made possible by having a single attention mechanism that is shared across all language pairs. We train the proposed multi-way, multilingual model on ten language pairs from WMT'15 simultaneously and observe clear performance improvements over models trained on only one language pair. In particular, we observe that the proposed model significantly improves the translation quality of low-resource language pairs.

Related articles: Most relevant | Search more
arXiv:1910.03467 [cs.CL] (Published 2019-10-07)
Overcoming the Rare Word Problem for Low-Resource Language Pairs in Neural Machine Translation
arXiv:2207.04906 [cs.CL] (Published 2022-07-11)
High-resource Language-specific Training for Multilingual Neural Machine Translation
arXiv:1909.07342 [cs.CL] (Published 2019-09-16)
Multilingual Neural Machine Translation for Zero-Resource Languages