Skip to main content

A Novel Distributed Reinforcement Learning Method for Classical Chinese Poetry Generation

  • Conference paper
  • First Online:
Parallel and Distributed Computing, Applications and Technologies (PDCAT 2020)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 12606))

  • 1101 Accesses

Abstract

Poetry generation has been a classic natural language generation task recently. But so far the methods for this topic mainly imitate and reproduce the poems on the training data set, which indicates that they either have not much connotation or overfit too much like plagiarism of the existing poems. To solve this problem, unlike previous work, instead of tuning the trade-off between connotation and innovation, we propose a distributed reinforcement learning framework, which consists of two stages of training, to generate creative and meaningful poetry. At the first stage we train a model in parallel on a large poetry corpus at word level to master how poets write poems. At the second stage we train the model with a distributed architecture to learn how connotation is developed in human literary art works at sentence level and force the model to imitate itself when it composes some ‘good poems’ to further improve performance. Experiments on generating classical Chinese poetry demonstrate that the proposed model is able to achieve better performance and the high efficiency of training compared to the state-of-the-art.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The data sets are publicly available from: https://github.com/chinese-poetry/chinese-poetry.

  2. 2.

    For GPT-based method, User may have to register a Wechat account and add or .

  3. 3.

    seqGAN code is available from: https://github.com/LantaoYu/SeqGAN.

  4. 4.

    Jiuge is available from: http://118.190.162.99:8080/.

  5. 5.

    The Natural Language Processing Group at the Department of Computer Science and Technology, Tsinghua University.

References

  1. Chen, H., Yi, X., Sun, M., Li, W., Yang, C., Guo, Z.: Sentiment-controllable Chinese poetry generation. In: IJCAI, pp. 4925–4931 (2019)

    Google Scholar 

  2. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  3. Horgan, D., et al.: Distributed prioritized experience replay (2018)

    Google Scholar 

  4. Liang, S.: Unsupervised semantic generative adversarial networks for expert retrieval. In: WWW (2019)

    Google Scholar 

  5. Liao, Y., Wang, Y., Liu, Q., Jiang, X.: Gpt-based generation for classical chinese poetry. arXiv preprint arXiv:1907.00151 (2019)

  6. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. Computer ence (2013)

    Google Scholar 

  7. Mikolov, T., Karafiát, M., Burget, L., Cernock, J., Khudanpur, S.: Recurrent neural network based language model. In: INTERSPEECH 2010 (2010)

    Google Scholar 

  8. Oh, J., Guo, Y., Singh, S., Lee, H.: Self-imitation learning. arXiv preprint arXiv:1806.05635 (2018)

  9. Papineni, K., Roukos, S., Ward, T., Zhu, W.J.: Bleu: a method for automatic evaluation of machine translation, October 2002. https://doi.org/10.3115/1073083.1073135

  10. Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018)

    Google Scholar 

  11. Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms (2017)

    Google Scholar 

  12. Song, Y., Shi, S., Li, J., Zhang, H.: Directional skip-gram: explicitly distinguishing left and right context for word embeddings. In: Proceedings of ACL 2018 (2018)

    Google Scholar 

  13. Sun, M., Yi, X., Li, W.: Stylistic chinese poetry generation via unsupervised style disentanglement, pp. 3960–3969, January 2018. https://doi.org/10.18653/v1/D18-1430

  14. Sutton, R., Barto, A.: Reinforcement Learning: An Introduction. MIT Press, United States (1998)

    MATH  Google Scholar 

  15. Wang, Z., et al.: Chinese poetry generation with planning based neural network. arXiv:1610.09889 (2016)

  16. Yan, R., Jiang, H., Lapata, M., Lin, S.D., Lv, X., Li, X.: I, poet: automatic chinese poetry composition through a generative summarization framework under constrained optimization. In: 23rd IJCAI (2013)

    Google Scholar 

  17. Yi, X., Li, R., Sun, M.: Chinese poetry generation with a salient-clue mechanism. CoNLL, 241–250 (2018)

    Google Scholar 

  18. Yi, X., Sun, M., Li, R., Li, W.: Automatic poetry generation with mutual reinforcement learning. Proc. EMNLP 2018, 3143–3153 (2018)

    Google Scholar 

  19. Yi, X., Sun, M., Li, R., Zonghan, Y.: Chinese poetry generation with a working memory model, September 2018

    Google Scholar 

  20. Yu, L., Zhang, W., Wang, J., Yu, Y.: Seqgan: sequence generative adversarial nets with policy gradient. In: AAAI-17 (2017)

    Google Scholar 

  21. Zhipeng, G., et al.: Jiuge: a human-machine collaborative chinese classical poetry generation system. In: Proceedings of ACL 2019: System Demonstrations, pp. 25–30 (2019)

    Google Scholar 

  22. Zinkevich, M., Weimer, M., Smola, A.J., Li, L.: Parallelized stochastic gradient descent. In: Proceedings of NIPS 2010 (2011)

    Google Scholar 

Download references

Acknowledgements

This work is supported by National Key R & D Program of China Project #2017YFB0203201, Key-Area Research and Development Plan of Guangdong Province 2020B010164003.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hong Shen .

Editor information

Editors and Affiliations

A More Examples of Our Method

A More Examples of Our Method

Table 3. Generated poems of our method

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ma, L., Shen, H., Liang, S. (2021). A Novel Distributed Reinforcement Learning Method for Classical Chinese Poetry Generation. In: Zhang, Y., Xu, Y., Tian, H. (eds) Parallel and Distributed Computing, Applications and Technologies. PDCAT 2020. Lecture Notes in Computer Science(), vol 12606. Springer, Cham. https://doi.org/10.1007/978-3-030-69244-5_3

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-69244-5_3

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-69243-8

  • Online ISBN: 978-3-030-69244-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics