Abstract
The translation and sharing of languages around the world has become a necessary precondition for the movement of people. Teaching Chinese as a foreign language (TCFL) undertakes international function of spreading national culture. How to translate Chinese as a foreign language into English has become an important task. Machine translation has moved beyond the realm of theory to practical use as a result of advancements in computing. Deep learning is a prominent and relatively young subfield of machine learning that has shown promising results in a variety of fields. This paper aims to develop a TCFL-oriented English-Chinese neural machine translation model. First, this paper proposes a hyperbolic tangent long short-term memory network (HTLSTM). This will integrate future information and historical information to extract more sufficient contextual semantic information. Secondly, this paper proposes a multi-subspace attention mechanism. This integrates multiple attention calculation functions in the multi-subspace attention mechanism (MSATT). Thirdly, this paper combines HTLSTM with MSATT to construct an English-Chinese bilingual neural translation model called ECBTNet. The multi-subspace attention maps hidden state of hyperbolic tangent long-term short-term memory network to multiple subspaces. This then uses multiple attention calculation functions in the multi-attention mechanism when calculating the attention score. By applying different attention calculation functions in different subspaces to extract omni-directional context information features, accurate attention calculation results can be obtained. Finally, a systematic experiment is carried out, and the experimental data verify the feasibility of applying ECBTNet to the field of English-Chinese translation in TCFL.








Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The datasets used during the current study are available from the corresponding author upon reasonable request.
References
Rivera-Trigueros I (2022) Machine translation systems and quality assessment: a systematic review[J]. Lang Resour Eval 56(2):593–619
Klimova B, Pikhart M, Benites AD et al (2023) Neural machine translation in foreign language teaching and learning: a systematic review[J]. Educ Inf Technol 28(1):663–682
Ranathunga S, Lee ESA, Prifti Skenduli M et al (2023) Neural machine translation for low-resource languages: a survey[J]. ACM Comput Surv 55(11):1–37
Lee SM (2023) The effectiveness of machine translation in foreign language education: a systematic review and meta-analysis[J]. Comput Assist Lang Learn 36(1–2):103–125
Guerberof-Arenas A, Toral A (2022) Creativity in translation: machine translation as a constraint for literary texts[J]. Transl Spaces 11(2):184–212
Stahlberg F (2020) Neural machine translation: a review[J]. J Artif Intell Res 69:343–418
Ryu J, Kim Y, Park S, et al. (2022) Exploring foreign language students’ perceptions of the guided use of machine translation (GUMT) model for Korean writing[J]. L2 J. 14(1)
Mondal SK, Zhang H, Kabir HMD et al (2023) Machine translation and its evaluation: a study[J]. Artif Intell Rev 1:1–90
Pei J, Zhong K, Yu Z, et al. (2022) Scene graph semantic inference for image and text matching[J]. Transactions on Asian and Low-Resource Language Information Processing, 1
Saunders D (2022) Domain adaptation and multi-domain adaptation for neural machine translation: a survey[J]. J Artif Intell Res 75:351–424
Samant RM, Bachute MR, Gite S et al (2022) Framework for deep learning-based language models using multi-task learning in natural language understanding: a systematic literature review and future directions[J]. IEEE Access 10:17078–17097
Dabre R, Chu C, Kunchukuttan A (2020) A survey of multilingual neural machine translation[J]. ACM Comput Surv (CSUR) 53(5):1–38
Andrabi SAB, Wahid A (2022) Machine translation system using deep learning for English to Urdu[J]. Comput Intell Neurosci
Al-Sayed MM (2022) Workload time series cumulative prediction mechanism for cloud resources using neural machine translation technique[J]. J Grid Comput 20(2):16
Nguyen PT, Di Rocco J, Rubei R et al (2022) DeepLib: Machine translation techniques to recommend upgrades for third-party libraries[J]. Expert Syst Appl 202:117267
Bensalah N, Ayad H, Adib A, et al. (2022) CRAN: an hybrid CNN-RNN attention-based model for Arabic machine translation[C]. Networking, Intelligent Systems and Security: Proceedings of NISS 2021. Springer Singapore, 87–102
Chiche A, Yitagesu B (2022) Part of speech tagging: a systematic review of deep learning and machine learning approaches[J]. J Big Data 9(1):1–25
Fan A, Bhosale S, Schwenk H et al (2021) Beyond english-centric multilingual machine translation[J]. J Mach Learn Res 22(1):4839–4886
Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks[J]. Adv Neural Inf Process Syst 27:3104–3112
Cho K, van Merriënboer B, Gulcehre C, et al. (2014) Learning Phrase Representations Using RNN Encoder–Decoder for Statistical Machine Translation[C]. Conference on Empirical Methods in Natural Language Processing, 1724–1734
Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate[J]. arXiv preprint arXiv:1409.0473
Luong M T, Pham H, Manning C D (2015) Effective approaches to attention-based neural machine translation[C]. Conference on Empirical Methods in Natural Language Processing 1412–1421
Jean S, Cho K, Memisevic R., Bengio, Y (2015) On using very large target vocabulary for neural machine translation[C]. Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing 1–10
Junczys-Dowmunt M, Dwojak T, Hoang H (2016) Is neural machine translation ready for deployment[J]. A case study on, 30
Gehring J, Auli M, Grangier D, et al. (2017) Convolutional sequence to sequence learning[C]. International conference on machine learning 1243–1252
Sennrich R, Haddow B, Birch A (2016) Neural machine translation of rare words with subword units[C]. Annual Meeting of the Association for Computational Linguistics 1715–1725
Vaswani A, Shazeer N, Parmar N, et al. (2017) Attention is All You Need[C]. International Conference on Neural Information Processing Systems, 6000–6010.
Hassan H, Aue A, Chen C, et al. (2018) Achieving human parity on automatic chinese to english news translation[J]. arXiv preprint arXiv:1803.05567
Dehghani M, Gouws S, Vinyals O, et al. (2018) Universal transformers[J]. arXiv preprint arXiv:1807.03819
Dai Z, Yang Z, Yang Y, et al. (2019) Transformer-XL: attentive language models beyond a fixed-length context[C]. Annual Meeting of the Association for Computational Linguistics 2978–2988
Wang Q, Li B, Xiao T, et al. (2019) Learning deep transformer models for machine translation[C]. Annual Meeting of the Association for Computational Linguistics., 1810–1822
Dedes K, Utama ABP, Wibawa AP et al. (2022) Neural machine translation of Spanish-English food recipes using LSTM[J]. JOIV: Int J Informat Visual 6(2):290–297
Xiao Q, Chang X, Zhang X et al (2020) Multi-information spatial–temporal LSTM fusion continuous sign language neural machine translation[J]. IEEE Access 8:216718–216728
Sartipi A, Dehghan M, Fatemi A (2023) An evaluation of persian-english machine translation datasets with transformers[J]. arXiv preprint arXiv:2302.00321
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no conflict of interest exists.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Yang, J. ECBTNet: English-Foreign Chinese intelligent translation via multi-subspace attention and hyperbolic tangent LSTM. Neural Comput & Applic 35, 25001–25011 (2023). https://doi.org/10.1007/s00521-023-08624-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-023-08624-8