Neural Machine Translation for Low-resource Languages: A Survey
Citations Over TimeTop 1% of 2022 papers
Abstract
Neural Machine Translation (NMT) has seen tremendous growth in the last ten years since the early 2000s and has already entered a mature phase. While considered the most widely used solution for Machine Translation, its performance on low-resource language pairs remains sub-optimal compared to the high-resource counterparts due to the unavailability of large parallel corpora. Therefore, the implementation of NMT techniques for low-resource language pairs has been receiving the spotlight recently, thus leading to substantial research on this topic. This article presents a detailed survey of research advancements in low-resource language NMT (LRL-NMT) and quantitative analysis to identify the most popular techniques. We provide guidelines to select the possible NMT technique for a given LRL data setting based on our findings. We also present a holistic view of the LRL-NMT research landscape and provide recommendations to enhance the research efforts further.
Related Papers
- → Improving Statistical Machine Translation with Word Class Models(2013)41 cited
- → Language Model Pre-training Method in Machine Translation Based on Named Entity Recognition(2020)14 cited
- → Towards State-of-the-art English-Vietnamese Neural Machine Translation(2017)8 cited
- → INCORPORATION OF RECOVERY AND REPAIR TIME IN THE RELIABILITY MODELLING OF FAULT-TOLERANT SYSTEM(1983)3 cited
- → Incorporation of Recovery and Repair Time in the Reliability Modelling of Fault-Tolerant System(1983)