Skip to main content

Translating Classical Chinese Poetry into Modern Chinese with Transformer

  • Conference paper
  • First Online:
Chinese Lexical Semantics (CLSW 2021)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 13249))

Included in the following conference series:

  • 485 Accesses

Abstract

Classical Chinese poetry, as the cultural heritage of human beings, is very popular in Chinese community all over the world. Nearly every person in these regions can recite several poems to artistically express his or her emotion. However, due to the huge difference between the archaic and modern Chinese language, now, it is hard for people to understand these poems. Considering the Neural machine translation (NMT) has made great progress since the self-attention technology called Transformer was adopted, we followed the paradigm of NMT to translate the classical Chinese poems into modern Chinese texts in this study. First, we collect many ancient poems and their explanations as a parallel corpus. Then, a NMT system is trained with this corpus. Finally, we compared this system with Generative Pre-Training (GPT) and analysis their results in detail.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://github.com/whl5902601/Poetry-translation.

  2. 2.

    https://www.gushiwen.org/.

  3. 3.

    https://github.com/huggingface/transformers.

  4. 4.

    https://github.com/hughqiu/GPT2-Chinese.

References

  1. Zhang, X., Lapata, M.: Chinese poetry generation with recurrent neural networks. In: EMNLP, pp. 670–680 (2014)

    Google Scholar 

  2. Yang, X., Lin X., Suo, S., Li, M.: Generating thematic Chinese poetry using conditional variational autoencoders with hybrid decoders. In: IJCAI, pp. 4539–4545 (2018)

    Google Scholar 

  3. Liao, Y., Wang Y., Liu, Q., Jiang, X.: GPT-based generation for classical Chinese poetry. In: arxiv (2019)

    Google Scholar 

  4. Sutskever, I., Vinyals, O., Le, Q.: Sequence to sequence learning with neural networks. In: NIPS, pp. 2238–2244 (2014)

    Google Scholar 

  5. Vaswani, A., et al.: Attention is all you need. In: NeurIPS (2017)

    Google Scholar 

  6. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language Models are Unsupervised Multitask Learners. Technical report (2019)

    Google Scholar 

  7. Papineni, K., Roukos S., Ward, T., Zhu, W.: BLEU: a method for automatic evaluation of machine translation. In: ACL (2002)

    Google Scholar 

  8. Yan, R.: I, Poet: automatic poetry composition through recurrent neural networks with iterative polishing schema. In: IJCAI, pp. 2238–2244 (2016)

    Google Scholar 

  9. Yi, X., Li R., Sun, M.: Chinese poetry generation with a salient-clue mechanism. In: Proceedings of the SIGNLL Conference on Computational Natural Language Learning, pp. 241–250 (2018)

    Google Scholar 

  10. Bahdanau, D., Cho K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: ICLR (2015)

    Google Scholar 

  11. Li, J., et al.: Generating classical Chinese poems via conditional variational autoencoder and adversarial training. In: EMNLP, pp. 3890–3900 (2018)

    Google Scholar 

  12. Xu, W., Ritter A., Dolan, W., Grishman, R., Cherry, C.: Paraphrasing for style. In: COLING (2012)

    Google Scholar 

  13. Jang, S., Min J., Kwon, M. : Writing Style Conversion using Neural Machine Translation. Technical report (2017)

    Google Scholar 

  14. Tian, H.: Learning to interpret and translate classical Chinese poetry. In: Theory and Practice in Language Studies. vol. 3, no. 9, pp. 1546–1551. Springer, Heidelberg (2013). https://doi.org/10.4304/tpls.3.9

  15. Devlin, J., Chang M., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: NAACL-HLT, pp. 4171–4186 (2019)

    Google Scholar 

  16. Yi, X., Sun M., Li, R., Li, W.: Automatic poetry generation with mutual reinforcement learning. In: ACL (2018)

    Google Scholar 

  17. He, D., et al.: Dual learning for machine translation. In: NIPS (2016)

    Google Scholar 

  18. Hazra, R., Dutta P., Gupta, S., Qaathir, M., Dukkipati, A.: Active2 learning: actively reducing redundancies in active learning methods for sequence tagging and machine translation. In: NAACL (2021)

    Google Scholar 

Download references

Acknowledgments

This study is supported by the fundamental research of Leshan Normal University (grant no. 801 205190117) and Natural Science Foundation of China (no. 61373056). We will thank the anonymous reviewers for their valuable comments.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shushan Zhu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Jin, P., Wang, H., Ma, L., Wang, B., Zhu, S. (2022). Translating Classical Chinese Poetry into Modern Chinese with Transformer. In: Dong, M., Gu, Y., Hong, JF. (eds) Chinese Lexical Semantics. CLSW 2021. Lecture Notes in Computer Science(), vol 13249. Springer, Cham. https://doi.org/10.1007/978-3-031-06703-7_37

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-06703-7_37

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-06702-0

  • Online ISBN: 978-3-031-06703-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics