Multilingual Translation from Denoising Pre-Training
2021pp. 3450–3466
Citations Over TimeTop 10% of 2021 papers
Yuqing Tang, Chau Tran, Xian Li, Peng‐Jen Chen, Naman Goyal, Vishrav Chaudhary, Jiatao Gu, Angela Fan
Abstract
Recent work demonstrates the potential of training one model for multilingual machine translation. In parallel, denoising pretraining using unlabeled monolingual data as a starting point for finetuning bitext machine translation systems has demonstrated strong performance gains. However, little has been explored on the potential to combine denoising pretraining with multilingual machine translation in a single model. In this work, we fill this gap by studying how multilingual translation models can be created through multilingual finetuning.
Related Papers
- → Improving Statistical Machine Translation with Word Class Models(2013)41 cited
- → Language Model Pre-training Method in Machine Translation Based on Named Entity Recognition(2020)14 cited
- → Towards State-of-the-art English-Vietnamese Neural Machine Translation(2017)8 cited
- → English-Japanese Neural Machine Translation with Encoder-Decoder-Reconstructor(2017)2 cited
- → Recurrent Stacking of Layers for Compact Neural Machine Translation Models(2018)2 cited