arXiv Analytics

Sign in

arXiv:1904.02461 [cs.CL]AbstractReferencesReviewsResources

ReWE: Regressing Word Embeddings for Regularization of Neural Machine Translation Systems

Inigo Jauregi Unanue, Ehsan Zare Borzeshi, Nazanin Esmaili, Massimo Piccardi

Published 2019-04-04Version 1

Regularization of neural machine translation is still a significant problem, especially in low-resource settings. To mollify this problem, we propose regressing word embeddings (ReWE) as a new regularization technique in a system that is jointly trained to predict the next word in the translation (categorical value) and its word embedding (continuous value). Such a joint training allows the proposed system to learn the distributional properties represented by the word embeddings, empirically improving the generalization to unseen sentences. Experiments over three translation datasets have showed a consistent improvement over a strong baseline, ranging between 0.91 and 2.54 BLEU points, and also a marked improvement over a state-of-the-art system.

Related articles: Most relevant | Search more
arXiv:2210.03696 [cs.CL] (Published 2022-10-07)
NMTSloth: Understanding and Testing Efficiency Degradation of Neural Machine Translation Systems
arXiv:2202.02170 [cs.CL] (Published 2022-02-04)
The Ecological Footprint of Neural Machine Translation Systems
arXiv:2103.06964 [cs.CL] (Published 2021-03-11)
Learning Policies for Multilingual Training of Neural Machine Translation Systems