Abstract
Language reconciles the ideas, beliefs and values of people from diverse cultural, social, economic, religious and professional backgrounds, and it is imperative for sustainable development. Undoubtedly, the emergence of Neural Machine Translation has gained significant advancement in language translation automation, and consequently, establishing seamless reconciliation between two communicating entities with diverse backgrounds. Comprising of sequence to sequence architecture, Neural Machine Translation models have outperformed statistical and rule-based models in terms of concordance and syntax, regardless of how complex the language structure is. However, structured language resources are scarce for low-resource languages like Akan. Research contributions toward the revitalization of these languages are limited, along with subtle adverse events such as cultural assimilation and language imperialism. In order to solve the problem of low-resource machine translation from Akan to English, we mine and use the first parallel corpus for the Akan-English translation. We establish a benchmark for the Akan-English translation using a deep hierarchical end-to-end attention-based neural machine translation model trained on the parallel corpus. Experimental results further confirm the effectiveness of our approach towards the task. Nevertheless, a confirmed grammatical parity between the two languages bequests improvement approaching a substantial revitalization of the Akan language.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
https://github.com/KingsleyNA/Parallel-Data-For-The-Revitalization-of-Tw-or-The-Akan-Language.
- 2.
- 3.
References
A Statistical approach to machine translation. In: Readings in Machine Translation (2018). https://doi.org/10.7551/mitpress/5779.003.0039
The Oxford Handbook of Endangered Languages (2018). https://doi.org/10.1093/oxfordhb/9780190610029.001.0001
Arkoh, R., Matthewson, L.: A familiar definite article in Akan. Lingua (2013). https://doi.org/10.1016/j.lingua.2012.09.012
Artetxe, M., Labaka, G., Agirre, E., Cho, K.: Unsupervised neural machine translation. In: 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings (2018)
Bahdanau, D., Cho, K.H., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings (2015)
Bhattacharyya, P.: Machine translation (2015). https://doi.org/10.1201/b18004
Brants, T., Popat, A.C., Xu, P., Och, F.J., Dean, J.: Large language models in machine translation. In: EMNLP-CoNLL 2007 - Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (2007)
Caterini, A.L., Chang, D.E.: Recurrent neural networks. In: SpringerBriefs in Computer Science (2018). https://doi.org/10.1007/978-3-319-75304-1_5
Charoenpornsawat, P., Sornlertlamvanich, V., Charoenporn, T.: Improving translation quality of rule-based machine translation (2002). https://doi.org/10.3115/1118794.1118799
Chung, J., Ahn, S., Bengio, Y.: Hierarchical multiscale recurrent neural networks. In: 5th International Conference on Learning Representations, ICLR 2017 - Conference Track Proceedings (2019)
Costa-Jussà, M.R., Fonollosa, J.A.: Latest trends in hybrid machine translation and its applications. Comput. Speech Lang. (2015). https://doi.org/10.1016/j.csl.2014.11.001
Dillinger, M.: An introduction to machine translation. In: AMTA 2010 - 9th Conference of the Association for Machine Translation in the Americas (2010). https://doi.org/10.2307/3721978
Ding, Y., Liu, Y., Luan, H., Sun, M.: Visualizing and understanding neural machine translation. In: ACL 2017 - 55th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (2017). https://doi.org/10.18653/v1/P17-1106
Duff, P.A.: Transnationalism, multilingualism, and identity. Annu. Rev. Appl. Linguist. (2015). https://doi.org/10.1017/S026719051400018X
Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: AISTATS (2010)
He, D., Xia, Y., Qin, T., Wang, L., Yu, N., Liu, T.Y., Ma, W.Y.: Dual learning for machine translation. In: Advances in Neural Information Processing Systems (2016)
Hermes, M., Bang, M., Marin, A.: Designing Indigenous language revitalization (2012). https://doi.org/10.17763/haer.82.3.q8117w861241871j
Hutchins, J.: Example-based machine translation: a review and commentary. Mach. Trans. (2005). https://doi.org/10.1007/s10590-006-9003-9
Johnson, M., Schuster, M., Le, Q.V., Krikun, M., Wu, Y., Chen, Z., Thorat, N., Viégas, F., Wattenberg, M., Corrado, G., Hughes, M., Dean, J.: Google’s multilingual neural machine translation system: enabling zero-shot translation. Trans. Assoc. Comput. Linguist. (2017). https://doi.org/10.1162/tacl_a_00065
Koehn, P., Knowles, R.: Six challenges for neural machine translation (2017). https://doi.org/10.18653/v1/w17-3204
Lample, G., Conneau, A., Denoyer, L., Ranzato, M.: Unsupervised machine translation using monolingual corpora only. In: 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings (2018)
Luong, M.T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. In: Conference Proceedings - EMNLP 2015: Conference on Empirical Methods in Natural Language Processing (2015). https://doi.org/10.18653/v1/d15-1166
Napoles, C., Sakaguchi, K., Post, M., Tetreault, J.: Ground truth for grammatical error correction metrics. In: ACL-IJCNLP 2015 - 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing, Proceedings of the Conference (2015). https://doi.org/10.3115/v1/p15-2097
Papineni, K., Roukos, S., Ward, T., Zhu, W.J.: BLEU: a method for automatic evaluation of machine translation. In: Computational Linguistics (2002)
Pascanu, R., Gulcehre, C., Cho, K., Bengio, Y.: How to construct deep recurrent neural networks. In: 2nd International Conference on Learning Representations, ICLR 2014 - Conference Track Proceedings (2014)
Sands, B., Sands, B.: Language Revitalization in Africa. In: The Oxford Handbook of Endangered Languages (2018). https://doi.org/10.1093/oxfordhb/9780190610029.013.29
Schmidhuber, J.: Deep Learning in neural networks: an overview (2015). https://doi.org/10.1016/j.neunet.2014.09.003
Sennrich, R., Haddow, B., Birch, A.: Neural machine translation of rare words with subword units. In: 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016 - Long Papers (2016). https://doi.org/10.18653/v1/p16-1162
Simard, M., Ueffing, N., Isabelle, P., Kuhn, R.: Rule-based translation with statistical phrase-based post-editing (2007). https://doi.org/10.3115/1626355.1626383
Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems (2014)
Tiede, M.K.: An MRI-based study of pharyngeal volume contrasts in Akan and English. J. Phonetics (1996). https://doi.org/10.1006/jpho.1996.0022
Tucker, G.R.: A global perspective on bilingualism and bilingual education. In: Georgetown University Round table on Languages and Linguistics 1999 (2001)
Vaswani, A., Brain, G., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., Polosukhin, I.: Attention is all you need. In: Advances in Neural Information Processing Systems (2017)
Wang, X., Lu, Z., Tu, Z., Li, H., Xiong, D., Zhang, M.: Neural machine translation advised by statistical machine translation. In: 31st AAAI Conference on Artificial Intelligence, AAAI 2017 (2017)
Wiseman, S., Rush, A.M.: Sequence-to-sequence learning as beam-search optimization. In: EMNLP 2016 - Conference on Empirical Methods in Natural Language Processing, Proceedings (2016). https://doi.org/10.18653/v1/d16-1137
Wu, J., Hou, H., Shen, Z., Du, J., Li, J.: Adapting attention-based neural network to low-resource Mongolian-Chinese machine translation. In: NLPCC/ICCPOL (2016)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Acheampong, K.N., Sackey, N.N.O. (2021). Language Revitalization: A Benchmark for Akan-to-English Machine Translation. In: Arai, K., Kapoor, S., Bhatia, R. (eds) Intelligent Systems and Applications. IntelliSys 2020. Advances in Intelligent Systems and Computing, vol 1251. Springer, Cham. https://doi.org/10.1007/978-3-030-55187-2_20
Download citation
DOI: https://doi.org/10.1007/978-3-030-55187-2_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-55186-5
Online ISBN: 978-3-030-55187-2
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)