Skip to main content

Towards Better Translations from Classical to Modern Chinese: A New Dataset and a New Method

  • Conference paper
  • First Online:
Natural Language Processing and Chinese Computing (NLPCC 2023)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14302))

  • 1137 Accesses

Abstract

Classical Chinese (Ancient Chinese) is the written language that was used in ancient China and has been an important carrier of Chinese culture for thousands of years. Numerous ideas of modern disciplines have been influenced or derived from it, including mathematics, medicine, engineering, etc., which demonstrated the necessity for us to understand, inherit and disseminate it. Consequently, there is an urgent need to develop neural machine translation to facilitate the comprehension of classical Chinese sentences. In this paper, we introduce a high-quality and comprehensive dataset called C2MChn, consisting of about 615K sentence pairs for the translation between classical and modern Chinese. To the best of our knowledge, this is the first dataset covering a wide range of domains including history books, Buddhist classics, Confucian classics, etc. Furthermore, through the analysis of classical and modern Chinese, we have proposed a simple yet effective method, named Syntax-Semantics Awareness Transformer (SSAT). It’s capable of leveraging both syntactic and semantic information which are indispensable for better translating classical Chinese. Experiments show that our model can achieve better BLEU scores than several state-of-the-art methods as well as two general translation engines including Microsoft and Baidu APIs. The dataset and related resources will be released at: https://github.com/Zongyuan-Jiang/C2MChn.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://iwslt.org/.

  2. 2.

    https://www.statmt.org/wmt14/translation-task.html.

  3. 3.

    https://github.com/moses-smt/mosesdecoder/blob/master/scripts/generic/multi-bleu.perl.

  4. 4.

    https://fanyi.baidu.com/.

  5. 5.

    https://www.bing.com/translator/.

References

  1. Ataman, D., Negri, M., Turchi, M., Federicob, M.: Linguistically motivated vocabulary reduction for neural machine translation from Turkish to English. Prague Bull. Math. Linguist. 108, 331–342 (2017)

    Article  Google Scholar 

  2. Ba, J.L., Kiros, J.R., Hinton, G.E.: Layer normalization. arXiv preprint arXiv:1607.06450 (2016)

  3. Bahdanau, D., Cho, K.H., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: 3rd International Conference on Learning Representations, ICLR 2015 (2015)

    Google Scholar 

  4. Bapna, A., Chen, M.X., Firat, O., Cao, Y., Wu, Y.: Training deeper neural machine translation models with transparent attention. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 3028–3033 (2018)

    Google Scholar 

  5. Chang, E., Shiue, Y.T., Yeh, H.S., Demberg, V.: Time-aware ancient Chinese text translation and inference. In: Proceedings of the 2nd International Workshop on Computational Approaches to Historical Language Change 2021, pp. 1–6 (2021)

    Google Scholar 

  6. Dai, Z., Yang, Z., Yang, Y., Carbonell, J.G., Le, Q., Salakhutdinov, R.: Transformer-XL: attentive language models beyond a fixed-length context. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 2978–2988 (2019)

    Google Scholar 

  7. Dehghani, M., Gouws, S., Vinyals, O., Uszkoreit, J., Kaiser, L.: Universal transformers. In: International Conference on Learning Representations (2019)

    Google Scholar 

  8. Dou, Z.Y., Tu, Z., Wang, X., Shi, S., Zhang, T.: Exploiting deep representations for neural machine translation. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 4253–4262 (2018)

    Google Scholar 

  9. Gehring, J., Auli, M., Grangier, D., Yarats, D., Dauphin, Y.N.: Convolutional sequence to sequence learning. In: International Conference on Machine Learning, pp. 1243–1252. PMLR (2017)

    Google Scholar 

  10. Gu, J., Lu, Z., Li, H., Li, V.O.: Incorporating copying mechanism in sequence-to-sequence learning. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (vol. 1: Long Papers), pp. 1631–1640 (2016)

    Google Scholar 

  11. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  12. Hurskainen, A., Tiedemann, J.: Rule-based machine translation from English to Finnish. In: Proceedings of the Second Conference on Machine Translation, pp. 323–329 (2017)

    Google Scholar 

  13. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: International Conference on Learning Representations (2015)

    Google Scholar 

  14. Kitaev, N., Kaiser, L., Levskaya, A.: Reformer: the efficient transformer. In: International Conference on Learning Representations (2019)

    Google Scholar 

  15. Kontogianni, A., Ganetsos, T., Kousoulis, P., Papakitsos, E.C.: Computer-assisted translation of Egyptian-Coptic into Greek. J. Integr. Inf. Manage. (2020)

    Google Scholar 

  16. Liu, D., Yang, K., Qu, Q., Lv, J.: Ancient-modern Chinese translation with a new large training dataset. ACM Trans. Asian Low-Resour. Lang. Inf. Process. (TALLIP) 19(1), 1–13 (2019)

    Google Scholar 

  17. Liu, L., Liu, X., Gao, J., Chen, W., Han, J.: Understanding the difficulty of training transformers. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pp. 5747–5763 (2020)

    Google Scholar 

  18. Raganato, A., Tiedemann, J., et al.: An analysis of encoder representations in transformer-based machine translation. In: Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP. The Association for Computational Linguistics (2018)

    Google Scholar 

  19. So, D., Le, Q., Liang, C.: The evolved transformer. In: International Conference on Machine Learning, pp. 5877–5886. PMLR (2019)

    Google Scholar 

  20. Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems, vol. 27 (2014)

    Google Scholar 

  21. Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)

    Google Scholar 

  22. Wang, Q., et al.: Learning deep transformer models for machine translation. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 1810–1822 (2019)

    Google Scholar 

  23. Wang, Q., Li, F., Xiao, T., Li, Y., Li, Y., Zhu, J.: Multi-layer representation fusion for neural machine translation. In: Proceedings of the 27th International Conference on Computational Linguistics, pp. 3015–3026 (2018)

    Google Scholar 

  24. Wu, L., et al.: R-Drop: regularized dropout for neural networks. Adv. Neural. Inf. Process. Syst. 34, 10890–10905 (2021)

    Google Scholar 

  25. Xiong, R., et al.: On layer normalization in the transformer architecture. In: International Conference on Machine Learning, pp. 10524–10533. PMLR (2020)

    Google Scholar 

  26. Zhang, H., Yang, M., Zhao, T.: Exploring hybrid character-words representational unit in classical-to-modern Chinese machine translation. In: 2015 International Conference on Asian Language Processing (IALP), pp. 33–36. IEEE (2015)

    Google Scholar 

  27. Zhang, Z., Li, W., Su, Q.: Automatic translating between ancient Chinese and contemporary Chinese with limited aligned corpora. In: International Conference on Natural Language Processing and Chinese Computing, pp. 157–167 (2019)

    Google Scholar 

  28. Zhao, G., Sun, X., Xu, J., Zhang, Z., Luo, L.: MUSE: parallel multi-scale attention for sequence to sequence learning. arXiv preprint arXiv:1911.09483 (2019)

Download references

Acknowledgement

This research is supported in part by NSFC (Grant No.: 61936003) and Zhuhai Industry Core and Key Technology Research Project (no. 2220004002350). We would like to thank Mr. Xiandu Shi and Ms. Jing Zhang for providing some original data collation and data annotation for this work.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lianwen Jin .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Jiang, Z., Wang, J., Cao, J., Gao, X., Jin, L. (2023). Towards Better Translations from Classical to Modern Chinese: A New Dataset and a New Method. In: Liu, F., Duan, N., Xu, Q., Hong, Y. (eds) Natural Language Processing and Chinese Computing. NLPCC 2023. Lecture Notes in Computer Science(), vol 14302. Springer, Cham. https://doi.org/10.1007/978-3-031-44693-1_31

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-44693-1_31

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-44692-4

  • Online ISBN: 978-3-031-44693-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics