skip to main content
research-article

Low-resource Neural Machine Translation: Methods and Trends

Authors Info & Claims
Published:15 November 2022Publication History
Skip Abstract Section

Abstract

Neural Machine Translation (NMT) brings promising improvements in translation quality, but until recently, these models rely on large-scale parallel corpora. As such corpora only exist on a handful of language pairs, the translation performance is far from the desired effect in the majority of low-resource languages. Thus, developing low-resource language translation techniques is crucial and it has become a popular research field in neural machine translation. In this article, we make an overall review of existing deep learning techniques in low-resource NMT. We first show the research status as well as some widely used low-resource datasets. Then, we categorize the existing methods and show some representative works detailedly. Finally, we summarize the common characters among them and outline the future directions in this field.

REFERENCES

  1. [1] Artetxe Mikel, Labaka Gorka, and Agirre Eneko. 2017. Learning bilingual word embeddings with (almost) no bilingual data. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 451462.Google ScholarGoogle ScholarCross RefCross Ref
  2. [2] Artetxe Mikel, Labaka Gorka, and Agirre Eneko. 2019. An effective approach to unsupervised machine translation. arXiv preprint arXiv:1902.01313 (2019).Google ScholarGoogle Scholar
  3. [3] Artetxe Mikel, Labaka Gorka, and Agirre Eneko. 2020. Translation artifacts in cross-lingual transfer learning. arXiv preprint arXiv:2004.04721 (2020).Google ScholarGoogle Scholar
  4. [4] Artetxe Mikel, Labaka Gorka, Agirre Eneko, and Cho Kyunghyun. 2017. Unsupervised neural machine translation. arXiv preprint arXiv:1710.11041 (2017).Google ScholarGoogle Scholar
  5. [5] Bahdanau Dzmitry, Cho Kyunghyun, and Bengio Yoshua. 2014. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014).Google ScholarGoogle Scholar
  6. [6] Banerjee Satanjeev and Lavie Alon. 2005. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization. 6572.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. [7] Bertoldi Nicola and Federico Marcello. 2009. Domain adaptation for statistical machine translation with monolingual resources. In Proceedings of the Fourth Workshop on Statistical Machine Translation. 182189.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. [8] Bojar Ondřej and Tamchyna Aleš. 2011. Improving translation model by monolingual data. In Proceedings of the 6th Workshop on Statistical Machine Translation. 330336.Google ScholarGoogle Scholar
  9. [9] Brown Peter F., Pietra Stephen A. Della, Pietra Vincent J. Della, and Mercer Robert L.. 1993. The mathematics of statistical machine translation: Parameter estimation. Computat. Ling. 19, 2 (1993), 263311.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. [10] Bugliarello Emanuele and Okazaki Naoaki. 2019. Enhancing machine translation with dependency-aware self-attention. arXiv preprint arXiv:1909.03149 (2019).Google ScholarGoogle Scholar
  11. [11] Carbonell Jaime G., Klein Steve, Miller David, Steinbaum Mike, Grassiany Tomer, and Frey Jochen. 2006. Context-based machine translation. In Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers, 19–28.Google ScholarGoogle Scholar
  12. [12] Caswell Isaac, Chelba Ciprian, and Grangier David. 2019. Tagged back-translation. arXiv preprint arXiv:1906.06442 (2019).Google ScholarGoogle Scholar
  13. [13] Chen Yun, Liu Yang, Cheng Yong, and Li Victor O. K.. 2017. A teacher-student framework for zero-resource neural machine translation. arXiv preprint arXiv:1705.00753 (2017).Google ScholarGoogle Scholar
  14. [14] Chen Yun, Liu Yang, and Li Victor. 2018. Zero-resource neural machine translation with multi-agent communication game. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 32.Google ScholarGoogle ScholarCross RefCross Ref
  15. [15] Cheng Yong. 2019. Joint training for pivot-based neural machine translation. In Joint Training for Neural Machine Translation. Springer, 4154.Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Cheng Yong. 2019. Semi-supervised learning for neural machine translation. In Joint Training for Neural Machine Translation. Springer, 2540.Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Chiang David. 2005. A hierarchical phrase-based model for statistical machine translation. In Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL’05). 263270.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. [18] Chimalamarri Santwana, Sitaram Dinkar, and Jain Ashritha. 2020. Morphological segmentation to improve crosslingual word embeddings for low resource languages. ACM Trans. Asian Low-resour. Lang. Inf. Process. 19, 5 (2020), 115.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. [19] Chronopoulou Alexandra, Stojanovski Dario, and Fraser Alexander. 2020. Reusing a pretrained language model on languages with limited corpora for unsupervised NMT. arXiv preprint arXiv:2009.07610 (2020).Google ScholarGoogle Scholar
  20. [20] Collins Michael, Koehn Philipp, and Kučerová Ivona. 2005. Clause restructuring for statistical machine translation. In Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL’05). 531540.Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. [21] Cotterell Ryan and Kreutzer Julia. 2018. Explaining and generalizing back-translation through wake-sleep. arXiv preprint arXiv:1806.04402 (2018).Google ScholarGoogle Scholar
  22. [22] Currey Anna and Heafield Kenneth. 2019. Incorporating source syntax into transformer-based neural machine translation. In Proceedings of the 4th Conference on Machine Translation (Volume 1: Research Papers). 2433.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Currey Anna, Miceli-Barone Antonio Valerio, and Heafield Kenneth. 2017. Copied monolingual data improves low-resource neural machine translation. In Proceedings of the 2nd Conference on Machine Translation. 148156.Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Devlin Jacob, Chang Ming-Wei, Lee Kenton, and Toutanova Kristina. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).Google ScholarGoogle Scholar
  25. [25] Dou Qing and Knight Kevin. 2012. Large scale decipherment for out-of-domain machine translation. In Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning. 266275.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Edunov Sergey, Ott Myle, Auli Michael, and Grangier David. 2018. Understanding back-translation at scale. arXiv preprint arXiv:1808.09381 (2018).Google ScholarGoogle Scholar
  27. [27] Kholy Ahmed El, Habash Nizar, Leusch Gregor, Matusov Evgeny, and Sawaf Hassan. 2013. Language independent connectivity strength features for phrase pivot statistical machine translation. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 412418.Google ScholarGoogle Scholar
  28. [28] Fadaee Marzieh, Bisazza Arianna, and Monz Christof. 2017. Data augmentation for low-resource neural machine translation. arXiv preprint arXiv:1705.00440 (2017).Google ScholarGoogle Scholar
  29. [29] Fadaee Marzieh and Monz Christof. 2018. Back-translation sampling by targeting difficult words in neural machine translation. arXiv preprint arXiv:1808.09006 (2018).Google ScholarGoogle Scholar
  30. [30] Farajian M. Amin, Turchi Marco, Negri Matteo, and Federico Marcello. 2017. Multi-domain neural machine translation through unsupervised adaptation. In Proceedings of the 2nd Conference on Machine Translation. 127137.Google ScholarGoogle ScholarCross RefCross Ref
  31. [31] Firat Orhan, Sankaran Baskaran, Al-Onaizan Yaser, Vural Fatos T. Yarman, and Cho Kyunghyun. 2016. Zero-resource translation with multi-lingual neural machine translation. arXiv preprint arXiv:1606.04164 (2016).Google ScholarGoogle Scholar
  32. [32] Gal Yarin and Ghahramani Zoubin. 2016. A theoretically grounded application of dropout in recurrent neural networks. Adv. Neural Inf. Process. Syst. 29 (2016), 10191027.Google ScholarGoogle Scholar
  33. [33] Ganin Yaroslav, Ustinova Evgeniya, Ajakan Hana, Germain Pascal, Larochelle Hugo, Laviolette François, Marchand Mario, and Lempitsky Victor. 2016. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17, 1 (2016), 2096–2030.Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. [34] Gao Fei, Zhu Jinhua, Wu Lijun, Xia Yingce, Qin Tao, Cheng Xueqi, Zhou Wengang, and Liu Tie-Yan. 2019. Soft contextual data augmentation for neural machine translation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 55395544.Google ScholarGoogle ScholarCross RefCross Ref
  35. [35] Gibadullin Ilshat, Valeev Aidar, Khusainova Albina, and Khan Adil. 2019. A survey of methods to leverage monolingual data in low-resource neural machine translation. arXiv preprint arXiv:1910.00373 (2019).Google ScholarGoogle Scholar
  36. [36] Gu Jiatao, Hassan Hany, Devlin Jacob, and Li Victor O. K.. 2018. Universal neural machine translation for extremely low resource languages. arXiv preprint arXiv:1802.05368 (2018).Google ScholarGoogle Scholar
  37. [37] Gu Jiatao, Wang Yong, Chen Yun, Cho Kyunghyun, and Li Victor O. K.. 2018. Meta-learning for low-resource neural machine translation. arXiv preprint arXiv:1808.08437 (2018).Google ScholarGoogle Scholar
  38. [38] Gulcehre Caglar, Firat Orhan, Xu Kelvin, Cho Kyunghyun, Barrault Loic, Lin Huei-Chi, Bougares Fethi, Schwenk Holger, and Bengio Yoshua. 2015. On using monolingual corpora in neural machine translation. arXiv preprint arXiv:1503.03535 (2015).Google ScholarGoogle Scholar
  39. [39] Guo Junliang, Tan Xu, He Di, Qin Tao, Xu Linli, and Liu Tie-Yan. 2019. Non-autoregressive neural machine translation with enhanced decoder input. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 37233730.Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. [40] Guzmán Francisco, Chen Peng-Jen, Ott Myle, Pino Juan, Lample Guillaume, Koehn Philipp, Chaudhary Vishrav, and Ranzato Marc’Aurelio. 2019. The FLoRes evaluation datasets for low-resource machine translation: Nepali-English and Sinhala-English. arXiv preprint arXiv:1902.01382 (2019).Google ScholarGoogle Scholar
  41. [41] He Di, Xia Yingce, Qin Tao, Wang Liwei, Yu Nenghai, Liu Tie-Yan, and Ma Wei-Ying. 2016. Dual learning for machine translation. Adv. Neural Inf. Process. Syst. 29 (2016), 820828.Google ScholarGoogle Scholar
  42. [42] Hoang Vu Cong Duy, Koehn Philipp, Haffari Gholamreza, and Cohn Trevor. 2018. Iterative back-translation for neural machine translation. In Proceedings of the 2nd Workshop on Neural Machine Translation and Generation. 1824.Google ScholarGoogle ScholarCross RefCross Ref
  43. [43] Imankulova Aizhan, Dabre Raj, Fujita Atsushi, and Imamura Kenji. 2019. Exploiting out-of-domain parallel data through multilingual transfer learning for low-resource neural machine translation. arXiv preprint arXiv:1907.03060 (2019).Google ScholarGoogle Scholar
  44. [44] Imankulova Aizhan, Sato Takayuki, and Komachi Mamoru. 2019. Filtered pseudo-parallel corpus improves low-resource neural machine translation. ACM Trans. Asian Low-resour. Lang. Inf. Process. 19, 2 (2019), 116.Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. [45] Irvine Ann and Callison-Burch Chris. 2013. Combining bilingual and comparable corpora for low resource machine translation. In Proceedings of the 8th Workshop on Statistical Machine Translation. 262270.Google ScholarGoogle Scholar
  46. [46] Irvine Ann and Callison-Burch Chris. 2014. Hallucinating phrase translations for low resource MT. In Proceedings of the 18th Conference on Computational Natural Language Learning. 160170.Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Irvine Ann and Callison-Burch Chris. 2016. End-to-end statistical machine translation with zero or small parallel texts. Nat. Lang. Eng. 22, 4 (2016), 517.Google ScholarGoogle ScholarCross RefCross Ref
  48. [48] Isozaki Hideki, Sudoh Katsuhito, Tsukada Hajime, and Duh Kevin. 2010. Head finalization: A simple reordering rule for SOV languages. In Proceedings of the Joint 5th Workshop on Statistical Machine Translation and Metrics. 244251.Google ScholarGoogle Scholar
  49. [49] Ji Baijun, Zhang Zhirui, Duan Xiangyu, Zhang Min, Chen Boxing, and Luo Weihua. 2020. Cross-lingual pre-training based transfer for zero-shot neural machine translation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34. 115122.Google ScholarGoogle ScholarCross RefCross Ref
  50. [50] Jitao Xu, Crego Josep M., and Senellart Jean. 2020. Boosting neural machine translation with similar translations. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 15801590.Google ScholarGoogle Scholar
  51. [51] Johnson Melvin, Schuster Mike, Le Quoc V., Krikun Maxim, Wu Yonghui, Chen Zhifeng, Thorat Nikhil, Viégas Fernanda, Wattenberg Martin, Corrado Greg, et al. 2017. Google’s multilingual neural machine translation system: Enabling zero-shot translation. Trans. Assoc. Computat. Ling. 5 (2017), 339351.Google ScholarGoogle ScholarCross RefCross Ref
  52. [52] Karakanta Alina, Dehdari Jon, and Genabith Josef van. 2018. Neural machine translation for low-resource languages without parallel corpora. Mach. Transl. 32, 1 (2018), 167189.Google ScholarGoogle ScholarDigital LibraryDigital Library
  53. [53] Khayrallah Huda, Thompson Brian, Post Matt, and Koehn Philipp. 2020. Simulated multiple reference training improves low-resource machine translation. arXiv preprint arXiv:2004.14524 (2020).Google ScholarGoogle Scholar
  54. [54] Kim Yunsu, Gao Yingbo, and Ney Hermann. 2019. Effective cross-lingual transfer of neural machine translation models without shared vocabularies. arXiv preprint arXiv:1905.05475 (2019).Google ScholarGoogle Scholar
  55. [55] Kim Yunsu, Geng Jiahui, and Ney Hermann. 2019. Improving unsupervised word-by-word translation with language model and denoising autoencoder. arXiv preprint arXiv:1901.01590 (2019).Google ScholarGoogle Scholar
  56. [56] Kim Yunsu, Petrov Petre, Petrushkov Pavel, Khadivi Shahram, and Ney Hermann. 2019. Pivot-based transfer learning for neural machine translation between non-English languages. arXiv preprint arXiv:1909.09524 (2019).Google ScholarGoogle Scholar
  57. [57] Klementiev Alexandre, Irvine Ann, Callison-Burch Chris, and Yarowsky David. 2012. Toward statistical machine translation without parallel corpora. In Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics. 130140.Google ScholarGoogle ScholarDigital LibraryDigital Library
  58. [58] Kocmi Tom and Bojar Ondřej. 2018. Trivial transfer learning for low-resource neural machine translation. arXiv preprint arXiv:1809.00357 (2018).Google ScholarGoogle Scholar
  59. [59] Koehn Philipp, Och Franz J., and Marcu Daniel. 2003. Statistical Phrase-based Translation. Technical Report. University of Southern California Marina Del Rey Information Sciences Inst.Google ScholarGoogle ScholarCross RefCross Ref
  60. [60] Lakew Surafel M., Erofeeva Aliia, Negri Matteo, Federico Marcello, and Turchi Marco. 2018. Transfer learning in multilingual neural machine translation with dynamic vocabulary. arXiv preprint arXiv:1811.01137 (2018).Google ScholarGoogle Scholar
  61. [61] Lample Guillaume, Conneau Alexis, Denoyer Ludovic, and Ranzato Marc’Aurelio. 2017. Unsupervised machine translation using monolingual corpora only. arXiv preprint arXiv:1711.00043 (2017).Google ScholarGoogle Scholar
  62. [62] Lample Guillaume, Ott Myle, Conneau Alexis, Denoyer Ludovic, and Ranzato Marc’Aurelio. 2018. Phrase-based & neural unsupervised machine translation. arXiv preprint arXiv:1804.07755 (2018).Google ScholarGoogle Scholar
  63. [63] Lee Jason, Cho Kyunghyun, and Hofmann Thomas. 2017. Fully character-level neural machine translation without explicit segmentation. Trans. Assoc. Computat. Ling. 5 (2017), 365378.Google ScholarGoogle ScholarCross RefCross Ref
  64. [64] Leng Yichong, Tan Xu, Qin Tao, Li Xiang-Yang, and Liu Tie-Yan. 2019. Unsupervised pivot translation for distant languages. arXiv preprint arXiv:1906.02461 (2019).Google ScholarGoogle Scholar
  65. [65] Li Guanlin, Liu Lemao, Huang Guoping, Zhu Conghui, and Zhao Tiejun. 2019. Understanding data augmentation in neural machine translation: Two perspectives towards generalization. In Proceedings of the Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). 56935699.Google ScholarGoogle ScholarCross RefCross Ref
  66. [66] Li Rumeng, Wang Xun, and Yu Hong. 2020. MetaMT, a meta learning method leveraging multiple domain data for low resource machine translation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34. 82458252.Google ScholarGoogle ScholarCross RefCross Ref
  67. [67] Li Xiaoqing, Zhang Jiajun, and Zong Chengqing. 2016. One sentence one model for neural machine translation. arXiv preprint arXiv:1609.06490 (2016).Google ScholarGoogle Scholar
  68. [68] Liu Ding, Ma Ning, Yang Fangtao, and Yang Xuebin. 2019. A survey of low resource neural machine translation. In Proceedings of the 4th International Conference on Mechanical, Control and Computer Engineering (ICMCCE). IEEE, 39393.Google ScholarGoogle ScholarCross RefCross Ref
  69. [69] Liu Zihan, Winata Genta Indra, and Fung Pascale. 2021. Continual mixed-language pre-training for extremely low-resource neural machine translation. arXiv preprint arXiv:2105.03953 (2021).Google ScholarGoogle Scholar
  70. [70] Luo Gongxu, Yang Yating, Yuan Yang, Chen Zhanheng, and Ainiwaer Aizimaiti. 2019. Hierarchical transfer learning architecture for low-resource neural machine translation. IEEE Access 7 (2019), 154157154166.Google ScholarGoogle ScholarCross RefCross Ref
  71. [71] Luong Minh-Thang, Pham Hieu, and Manning Christopher D.. 2015. Effective approaches to attention-based neural machine translation. arXiv preprint arXiv:1508.04025 (2015).Google ScholarGoogle Scholar
  72. [72] Maimaiti Mieradilijiang, Liu Yang, Luan Huanbo, Pan Zegao, and Sun Maosong. 2021. Improving data augmentation for low-resource NMT guided by POS-Tagging and paraphrase embedding. Trans. Asian Low-resour. Lang. Inf. Process. 20, 6 (2021), 121.Google ScholarGoogle ScholarDigital LibraryDigital Library
  73. [73] Maimaiti Mieradilijiang, Liu Yang, Luan Huanbo, and Sun Maosong. 2019. Multi-round transfer learning for low-resource NMT using multiple high-resource languages. ACM Trans. Asian Low-resour. Lang. Inf. Process. 18, 4 (2019), 126.Google ScholarGoogle ScholarDigital LibraryDigital Library
  74. [74] Marie Benjamin, Rubino Raphael, and Fujita Atsushi. 2020. Tagged back-translation revisited: Why does it really work? In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 59905997.Google ScholarGoogle ScholarCross RefCross Ref
  75. [75] Nakayama Hideki and Nishida Noriki. 2017. Zero-resource machine translation by multimodal encoder–decoder network with multimedia pivot. Mach. Transl. 31, 1 (2017), 4964.Google ScholarGoogle ScholarDigital LibraryDigital Library
  76. [76] Nguyen Toan Q. and Chiang David. 2017. Transfer learning across low-resource, related languages for neural machine translation. arXiv preprint arXiv:1708.09803 (2017).Google ScholarGoogle Scholar
  77. [77] Nguyen Xuan-Phi, Joty Shafiq, Kui Wu, and Aw Ai Ti. 2019. Data diversification: A simple strategy for neural machine translation. arXiv preprint arXiv:1911.01986 (2019).Google ScholarGoogle Scholar
  78. [78] Pan Boyuan, Yang Yazheng, Li Hao, Zhao Zhou, Zhuang Yueting, Cai Deng, and He Xiaofei. 2019. MacNet: Transferring knowledge from machine comprehension to sequence-to-sequence models. arXiv preprint arXiv:1908.01816 (2019).Google ScholarGoogle Scholar
  79. [79] Papineni Kishore, Roukos Salim, Ward Todd, and Zhu Wei-Jing. 2002. Bleu: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics. 311318.Google ScholarGoogle ScholarDigital LibraryDigital Library
  80. [80] Paul Michael, Finch Andrew, and Sumita Eiichrio. 2013. How to choose the best pivot language for automatic translation of low-resource languages. ACM Trans. Asian Lang. Inf. Process. 12, 4 (2013), 117.Google ScholarGoogle ScholarDigital LibraryDigital Library
  81. [81] Peters Matthew E., Neumann Mark, Iyyer Mohit, Gardner Matt, Clark Christopher, Lee Kenton, and Zettlemoyer Luke. 2018. Deep contextualized word representations. arXiv preprint arXiv:1802.05365 (2018).Google ScholarGoogle Scholar
  82. [82] Pham Hieu, Wang Xinyi, Yang Yiming, and Neubig Graham. 2021. Meta back-translation. arXiv preprint arXiv:2102.07847 (2021).Google ScholarGoogle Scholar
  83. [83] Popović Maja. 2015. chrF: Character n-gram F-score for automatic MT evaluation. In Proceedings of the 10th Workshop on Statistical Machine Translation. 392395.Google ScholarGoogle ScholarCross RefCross Ref
  84. [84] Pourdamghani Nima, Aldarrab Nada, Ghazvininejad Marjan, Knight Kevin, and May Jonathan. 2019. Translating translationese: A two-step approach to unsupervised machine translation. arXiv preprint arXiv:1906.05683 (2019).Google ScholarGoogle Scholar
  85. [85] Pourdamghani Nima and Knight Kevin. 2017. Deciphering related languages. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. 25132518.Google ScholarGoogle ScholarCross RefCross Ref
  86. [86] Radford Alec, Wu Jeffrey, Child Rewon, Luan David, Amodei Dario, Sutskever Ilya, et al. 2019. Language models are unsupervised multitask learners. OpenAI Blog 1, 8 (2019), 9.Google ScholarGoogle Scholar
  87. [87] Ravi Sujith and Knight Kevin. 2011. Deciphering foreign language. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies. 1221.Google ScholarGoogle ScholarDigital LibraryDigital Library
  88. [88] Ren Shuo, Chen Wenhu, Liu Shujie, Li Mu, Zhou Ming, and Ma Shuai. 2018. Triangular architecture for rare language translation. arXiv preprint arXiv:1805.04813 (2018).Google ScholarGoogle Scholar
  89. [89] Ren Shuo, Wu Yu, Liu Shujie, Zhou Ming, and Ma Shuai. 2019. Explicit cross-lingual pre-training for unsupervised machine translation. arXiv preprint arXiv:1909.00180 (2019).Google ScholarGoogle Scholar
  90. [90] Ren Shuo, Wu Yu, Liu Shujie, Zhou Ming, and Ma Shuai. 2020. A retrieve-and-rewrite initialization method for unsupervised machine translation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 34983504.Google ScholarGoogle ScholarCross RefCross Ref
  91. [91] Sen Sukanta, Gupta Kamal Kumar, Ekbal Asif, and Bhattacharyya Pushpak. 2019. Multilingual unsupervised NMT using shared encoder and language-specific decoders. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 30833089.Google ScholarGoogle ScholarCross RefCross Ref
  92. [92] Sennrich Rico, Haddow Barry, and Birch Alexandra. 2015. Improving neural machine translation models with monolingual data. arXiv preprint arXiv:1511.06709 (2015).Google ScholarGoogle Scholar
  93. [93] Sennrich Rico, Haddow Barry, and Birch Alexandra. 2016. Edinburgh neural machine translation systems for WMT 16. arXiv preprint arXiv:1606.02891 (2016).Google ScholarGoogle Scholar
  94. [94] Sennrich Rico and Zhang Biao. 2019. Revisiting low-resource neural machine translation: A case study. arXiv preprint arXiv:1905.11901 (2019).Google ScholarGoogle Scholar
  95. [95] Shavarani Hassan S. and Sarkar Anoop. 2021. Better neural machine translation by extracting linguistic information from BERT. arXiv preprint arXiv:2104.02831 (2021).Google ScholarGoogle Scholar
  96. [96] Snover Matthew, Dorr Bonnie, Schwartz Richard, Micciulla Linnea, and Makhoul John. 2006. A study of translation edit rate with targeted human annotation. In Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers. 223231.Google ScholarGoogle Scholar
  97. [97] Sun Haipeng, Wang Rui, Chen Kehai, Utiyama Masao, Sumita Eiichiro, and Zhao Tiejun. 2019. Unsupervised bilingual word embedding agreement for unsupervised neural machine translation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 12351245.Google ScholarGoogle ScholarCross RefCross Ref
  98. [98] Sutskever Ilya, Vinyals Oriol, and Le Quoc V.. 2014. Sequence to sequence learning with neural networks. arXiv preprint arXiv:1409.3215 (2014).Google ScholarGoogle Scholar
  99. [99] Vaswani Ashish, Shazeer Noam, Parmar Niki, Uszkoreit Jakob, Jones Llion, Gomez Aidan N., Kaiser Lukasz, and Polosukhin Illia. 2017. Attention is all you need. arXiv preprint arXiv:1706.03762 (2017).Google ScholarGoogle Scholar
  100. [100] Vincent Pascal, Larochelle Hugo, Lajoie Isabelle, Bengio Yoshua, Manzagol Pierre-Antoine, and Bottou Léon. 2010. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion.J. Mach. Learn. Res. 11, 12 (2010).Google ScholarGoogle Scholar
  101. [101] Wang Rui, Tan Xu, Luo Renqian, Qin Tao, and Liu Tie-Yan. 2021. A survey on low-resource neural machine translation. arXiv preprint arXiv:2107.04239 (2021).Google ScholarGoogle Scholar
  102. [102] Wang Xinyi, Pham Hieu, Dai Zihang, and Neubig Graham. 2018. SwitchOut: An efficient data augmentation algorithm for neural machine translation. arXiv preprint arXiv:1808.07512 (2018).Google ScholarGoogle Scholar
  103. [103] Wei Hao-Ran, Zhang Zhirui, Chen Boxing, and Luo Weihua. 2020. Iterative domain-repaired back-translation. arXiv preprint arXiv:2010.02473 (2020).Google ScholarGoogle Scholar
  104. [104] Weng Rongxiang, Yu Heng, Huang Shujian, Cheng Shanbo, and Luo Weihua. 2020. Acquiring knowledge from pre-trained model to neural machine translation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34. 92669273.Google ScholarGoogle ScholarCross RefCross Ref
  105. [105] Weng Rongxiang, Yu Heng, Huang Shujian, Luo Weihua, and Chen Jiajun. 2019. Improving neural machine translation with pre-trained representation. arXiv preprint arXiv:1908.07688 (2019).Google ScholarGoogle Scholar
  106. [106] Wu Jiawei, Wang Xin, and Wang William Yang. 2019. Extract and edit: An alternative to back-translation for unsupervised neural machine translation. arXiv preprint arXiv:1904.02331 (2019).Google ScholarGoogle Scholar
  107. [107] Xia Yingce, He Di, Qin Tao, Wang Liwei, Yu Nenghai, Liu Tie-Yan, and Ma Wei-Ying. 2016. Dual learning for machine translation. arXiv preprint arXiv:1611.00179 (2016).Google ScholarGoogle Scholar
  108. [108] Xie Ziang, Wang Sida I., Li Jiwei, Lévy Daniel, Nie Aiming, Jurafsky Dan, and Ng Andrew Y.. 2017. Data noising as smoothing in neural network language models. arXiv preprint arXiv:1703.02573 (2017).Google ScholarGoogle Scholar
  109. [109] Yang Zhen, Chen Wei, Wang Feng, and Xu Bo. 2018. Unsupervised neural machine translation with weight sharing. arXiv preprint arXiv:1804.09057 (2018).Google ScholarGoogle Scholar
  110. [110] Zahabi Samira Tofighi, Bakhshaei Somayeh, and Khadivi Shahram. 2013. Using context vectors in improving a machine translation system with bridge language. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 318322.Google ScholarGoogle Scholar
  111. [111] Zhang Jiajun and Zong Chengqing. 2016. Exploiting source-side monolingual data in neural machine translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. 15351545.Google ScholarGoogle ScholarCross RefCross Ref
  112. [112] Zhang Meng, Liu Yang, Luan Huanbo, and Sun Maosong. 2017. Adversarial training for unsupervised bilingual lexicon induction. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 19591970.Google ScholarGoogle ScholarCross RefCross Ref
  113. [113] Zheng Hao, Cheng Yong, and Liu Yang. 2017. Maximum expected likelihood estimation for zero-resource neural machine translation. In Proceedings of the International Joint Conference on Artificial Intelligence. 42514257.Google ScholarGoogle ScholarCross RefCross Ref
  114. [114] Zheng Zaixiang, Zhou Hao, Huang Shujian, Li Lei, Dai Xin-Yu, and Chen Jiajun. 2019. Mirror-generative neural machine translation. In Proceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  115. [115] Zhou Chunting, Ma Xuezhe, Hu Junjie, and Neubig Graham. 2019. Handling syntactic divergence in low-resource machine translation. arXiv preprint arXiv:1909.00040 (2019).Google ScholarGoogle Scholar
  116. [116] Zoph Barret, Yuret Deniz, May Jonathan, and Knight Kevin. 2016. Transfer learning for low-resource neural machine translation. arXiv preprint arXiv:1604.02201 (2016).Google ScholarGoogle Scholar

Index Terms

  1. Low-resource Neural Machine Translation: Methods and Trends

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Asian and Low-Resource Language Information Processing
      ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 21, Issue 5
      September 2022
      486 pages
      ISSN:2375-4699
      EISSN:2375-4702
      DOI:10.1145/3533669
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 15 November 2022
      • Online AM: 15 March 2022
      • Accepted: 27 January 2022
      • Revised: 15 December 2021
      • Received: 10 June 2021
      Published in tallip Volume 21, Issue 5

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Refereed

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text

    HTML Format

    View this article in HTML Format .

    View HTML Format