Joint Dropout: Improving Generalizability in Low-Resource Neural Machine Translation through Phrase Pair Variables

Ali Araabi, Vlad Niculae, Christof Monz


Abstract
Despite the tremendous success of Neural Machine Translation (NMT), its performance on low- resource language pairs still remains subpar, partly due to the limited ability to handle previously unseen inputs, i.e., generalization. In this paper, we propose a method called Joint Dropout, that addresses the challenge of low-resource neural machine translation by substituting phrases with variables, resulting in significant enhancement of compositionality, which is a key aspect of generalization. We observe a substantial improvement in translation quality for language pairs with minimal resources, as seen in BLEU and Direct Assessment scores. Furthermore, we conduct an error analysis, and find Joint Dropout to also enhance generalizability of low-resource NMT in terms of robustness and adaptability across different domains.
Anthology ID:
2023.mtsummit-research.2
Volume:
Proceedings of Machine Translation Summit XIX, Vol. 1: Research Track
Month:
September
Year:
2023
Address:
Macau SAR, China
Editors:
Masao Utiyama, Rui Wang
Venue:
MTSummit
SIG:
Publisher:
Asia-Pacific Association for Machine Translation
Note:
Pages:
12–25
Language:
URL:
https://aclanthology.org/2023.mtsummit-research.2
DOI:
Bibkey:
Cite (ACL):
Ali Araabi, Vlad Niculae, and Christof Monz. 2023. Joint Dropout: Improving Generalizability in Low-Resource Neural Machine Translation through Phrase Pair Variables. In Proceedings of Machine Translation Summit XIX, Vol. 1: Research Track, pages 12–25, Macau SAR, China. Asia-Pacific Association for Machine Translation.
Cite (Informal):
Joint Dropout: Improving Generalizability in Low-Resource Neural Machine Translation through Phrase Pair Variables (Araabi et al., MTSummit 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.mtsummit-research.2.pdf