Skip to main content
Log in

ECBTNet: English-Foreign Chinese intelligent translation via multi-subspace attention and hyperbolic tangent LSTM

  • S.I.: Evolutionary Computation based Methods and Applications for Data Processing
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

The translation and sharing of languages around the world has become a necessary precondition for the movement of people. Teaching Chinese as a foreign language (TCFL) undertakes international function of spreading national culture. How to translate Chinese as a foreign language into English has become an important task. Machine translation has moved beyond the realm of theory to practical use as a result of advancements in computing. Deep learning is a prominent and relatively young subfield of machine learning that has shown promising results in a variety of fields. This paper aims to develop a TCFL-oriented English-Chinese neural machine translation model. First, this paper proposes a hyperbolic tangent long short-term memory network (HTLSTM). This will integrate future information and historical information to extract more sufficient contextual semantic information. Secondly, this paper proposes a multi-subspace attention mechanism. This integrates multiple attention calculation functions in the multi-subspace attention mechanism (MSATT). Thirdly, this paper combines HTLSTM with MSATT to construct an English-Chinese bilingual neural translation model called ECBTNet. The multi-subspace attention maps hidden state of hyperbolic tangent long-term short-term memory network to multiple subspaces. This then uses multiple attention calculation functions in the multi-attention mechanism when calculating the attention score. By applying different attention calculation functions in different subspaces to extract omni-directional context information features, accurate attention calculation results can be obtained. Finally, a systematic experiment is carried out, and the experimental data verify the feasibility of applying ECBTNet to the field of English-Chinese translation in TCFL.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data availability

The datasets used during the current study are available from the corresponding author upon reasonable request.

References

  1. Rivera-Trigueros I (2022) Machine translation systems and quality assessment: a systematic review[J]. Lang Resour Eval 56(2):593–619

    Article  Google Scholar 

  2. Klimova B, Pikhart M, Benites AD et al (2023) Neural machine translation in foreign language teaching and learning: a systematic review[J]. Educ Inf Technol 28(1):663–682

    Article  Google Scholar 

  3. Ranathunga S, Lee ESA, Prifti Skenduli M et al (2023) Neural machine translation for low-resource languages: a survey[J]. ACM Comput Surv 55(11):1–37

    Article  Google Scholar 

  4. Lee SM (2023) The effectiveness of machine translation in foreign language education: a systematic review and meta-analysis[J]. Comput Assist Lang Learn 36(1–2):103–125

    Article  Google Scholar 

  5. Guerberof-Arenas A, Toral A (2022) Creativity in translation: machine translation as a constraint for literary texts[J]. Transl Spaces 11(2):184–212

    Article  Google Scholar 

  6. Stahlberg F (2020) Neural machine translation: a review[J]. J Artif Intell Res 69:343–418

    Article  MathSciNet  Google Scholar 

  7. Ryu J, Kim Y, Park S, et al. (2022) Exploring foreign language students’ perceptions of the guided use of machine translation (GUMT) model for Korean writing[J]. L2 J. 14(1)

  8. Mondal SK, Zhang H, Kabir HMD et al (2023) Machine translation and its evaluation: a study[J]. Artif Intell Rev 1:1–90

    Google Scholar 

  9. Pei J, Zhong K, Yu Z, et al. (2022) Scene graph semantic inference for image and text matching[J]. Transactions on Asian and Low-Resource Language Information Processing, 1

  10. Saunders D (2022) Domain adaptation and multi-domain adaptation for neural machine translation: a survey[J]. J Artif Intell Res 75:351–424

    Article  MathSciNet  MATH  Google Scholar 

  11. Samant RM, Bachute MR, Gite S et al (2022) Framework for deep learning-based language models using multi-task learning in natural language understanding: a systematic literature review and future directions[J]. IEEE Access 10:17078–17097

    Article  Google Scholar 

  12. Dabre R, Chu C, Kunchukuttan A (2020) A survey of multilingual neural machine translation[J]. ACM Comput Surv (CSUR) 53(5):1–38

    Article  Google Scholar 

  13. Andrabi SAB, Wahid A (2022) Machine translation system using deep learning for English to Urdu[J]. Comput Intell Neurosci

  14. Al-Sayed MM (2022) Workload time series cumulative prediction mechanism for cloud resources using neural machine translation technique[J]. J Grid Comput 20(2):16

    Article  Google Scholar 

  15. Nguyen PT, Di Rocco J, Rubei R et al (2022) DeepLib: Machine translation techniques to recommend upgrades for third-party libraries[J]. Expert Syst Appl 202:117267

    Article  Google Scholar 

  16. Bensalah N, Ayad H, Adib A, et al. (2022) CRAN: an hybrid CNN-RNN attention-based model for Arabic machine translation[C]. Networking, Intelligent Systems and Security: Proceedings of NISS 2021. Springer Singapore, 87–102

  17. Chiche A, Yitagesu B (2022) Part of speech tagging: a systematic review of deep learning and machine learning approaches[J]. J Big Data 9(1):1–25

    Article  Google Scholar 

  18. Fan A, Bhosale S, Schwenk H et al (2021) Beyond english-centric multilingual machine translation[J]. J Mach Learn Res 22(1):4839–4886

    MathSciNet  MATH  Google Scholar 

  19. Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks[J]. Adv Neural Inf Process Syst 27:3104–3112

    Google Scholar 

  20. Cho K, van Merriënboer B, Gulcehre C, et al. (2014) Learning Phrase Representations Using RNN Encoder–Decoder for Statistical Machine Translation[C]. Conference on Empirical Methods in Natural Language Processing, 1724–1734

  21. Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate[J]. arXiv preprint arXiv:1409.0473

  22. Luong M T, Pham H, Manning C D (2015) Effective approaches to attention-based neural machine translation[C]. Conference on Empirical Methods in Natural Language Processing 1412–1421

  23. Jean S, Cho K, Memisevic R., Bengio, Y (2015) On using very large target vocabulary for neural machine translation[C]. Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing 1–10

  24. Junczys-Dowmunt M, Dwojak T, Hoang H (2016) Is neural machine translation ready for deployment[J]. A case study on, 30

  25. Gehring J, Auli M, Grangier D, et al. (2017) Convolutional sequence to sequence learning[C]. International conference on machine learning 1243–1252

  26. Sennrich R, Haddow B, Birch A (2016) Neural machine translation of rare words with subword units[C]. Annual Meeting of the Association for Computational Linguistics 1715–1725

  27. Vaswani A, Shazeer N, Parmar N, et al. (2017) Attention is All You Need[C]. International Conference on Neural Information Processing Systems, 6000–6010.

  28. Hassan H, Aue A, Chen C, et al. (2018) Achieving human parity on automatic chinese to english news translation[J]. arXiv preprint arXiv:1803.05567

  29. Dehghani M, Gouws S, Vinyals O, et al. (2018) Universal transformers[J]. arXiv preprint arXiv:1807.03819

  30. Dai Z, Yang Z, Yang Y, et al. (2019) Transformer-XL: attentive language models beyond a fixed-length context[C]. Annual Meeting of the Association for Computational Linguistics 2978–2988

  31. Wang Q, Li B, Xiao T, et al. (2019) Learning deep transformer models for machine translation[C]. Annual Meeting of the Association for Computational Linguistics., 1810–1822

  32. Dedes K, Utama ABP, Wibawa AP et al. (2022) Neural machine translation of Spanish-English food recipes using LSTM[J]. JOIV: Int J Informat Visual 6(2):290–297

  33. Xiao Q, Chang X, Zhang X et al (2020) Multi-information spatial–temporal LSTM fusion continuous sign language neural machine translation[J]. IEEE Access 8:216718–216728

    Article  Google Scholar 

  34. Sartipi A, Dehghan M, Fatemi A (2023) An evaluation of persian-english machine translation datasets with transformers[J]. arXiv preprint arXiv:2302.00321

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jing Yang.

Ethics declarations

Conflict of interest

The authors declare no conflict of interest exists.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yang, J. ECBTNet: English-Foreign Chinese intelligent translation via multi-subspace attention and hyperbolic tangent LSTM. Neural Comput & Applic 35, 25001–25011 (2023). https://doi.org/10.1007/s00521-023-08624-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-023-08624-8

Keywords