Improving Machine Translation Quality with Denoising Autoencoder and Pre-Ordering
Abstract
The problems in machine translation are related to the characteristics of a family of languages, especially syntactic divergences between languages. In the translation task, having both source and target languages in the same language family is a luxury that cannot be relied upon. The trained models for the task must overcome such differences either through manual augmentations or automatically inferred capacity built into the model design. In this work, we investigated the impact of multiple methods of differing word orders during translation and further experimented in assimilating the source languages syntax to the target word order using pre-ordering. We focused on the field of extremely low-resource scenarios. We also conducted experiments on practical data augmentation techniques that support the reordering capacity of the models through varying the target objectives, adding the secondary goal of removing noises or reordering broken input sequences. In particular, we propose methods to improve translat on quality with the denoising autoencoder in Neural Machine Translation (NMT) and pre-ordering method in Phrase-based Statistical Machine Translation (PBSMT). The experiments with a number of English-Vietnamese pairs show the improvement in BLEU scores as compared to both the NMT and SMT systems.
To cite this article: T. Hong-Viet, N. Van-Vinh and N. Hoang-Quan, “Improving Machine Translation Quality with Denoising Autoencoder and Pre-Ordering,” in CIT. Journal of Computing and Information Technology, vol. 29, no. 1, pp. 39–56, 2022, doi: 10.20532/cit.2021.1005316.
Keywords
Full Text:
PDFThis work is licensed under a Creative Commons Attribution-NoDerivatives 4.0 International License.