Skip to main content

A Novel Deep Reinforcement Learning Framework for Stock Portfolio Optimization

  • Conference paper
  • First Online:
  • 2460 Accesses

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1516))

Abstract

Deep reinforcement learning (DRL) is a recently concerned research field for stock portfolio optimization. The existing solutions face various challenge. In this paper, we propose a DRL framework to the stock portfolio optimization problem, which mainly includes the following three contributions: 1) We propose an Over-fitting Prevention Objective Function (OPOF) to avoid over-fitting in the training process. 2) An algorithm called Batch-Forward Recurrent Reinforcement Learning (BFRRL) is proposed to improve the stability of the training process. 3) A neural network called Multi Times Scale Transformer (MTS-Trans) is proposed to enhance stock series local feature extraction ability in multiple time scales. Compared with the current SOTA algorithm, our approach improves returns by 63% in the Chinese stock market and 138% in the U.S. stock market, while the strategy’s risk is also reduced.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Bai, S., Kolter, J.Z., Koltun, V.: An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271 (2018)

  2. Bakshaev, A.: Market-making with reinforcement-learning (SAC). arXiv preprint arXiv:2008.12275 (2020)

  3. Jiang, Z., Xu, D., Liang, J.: A deep reinforcement learning framework for the financial portfolio management problem. arXiv preprint arXiv:1706.10059 (2017)

  4. Jin, O., El-Saawy, H.: Portfolio management using reinforcement learning. Stanford University (2016)

    Google Scholar 

  5. Kelly Jr., J.L.: A new interpretation of information rate. In: The Kelly Capital Growth Investment Criterion: Theory and Practice, pp. 25–34. World Scientific (2011)

    Google Scholar 

  6. Li, X., Li, Y., Zhan, Y., Liu, X.Y.: Optimistic bull or pessimistic bear: adaptive deep reinforcement learning for stock portfolio allocation. arXiv preprint arXiv:1907.01503 (2019)

  7. Liang, Z., Chen, H., Zhu, J., Jiang, K., Li, Y.: Adversarial deep reinforcement learning in portfolio management. arXiv preprint arXiv:1808.09940 (2018)

  8. Lim, B., Arık, S.Ö., Loeff, N., Pfister, T.: Temporal fusion transformers for interpretable multi-horizon time series forecasting. Int. J. Forecast. 37, 1748–1764 (2021)

    Article  Google Scholar 

  9. Markowitz, H.: Portfolio selection. J. Financ. 7(1), 77–91 (1952)

    Google Scholar 

  10. Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)

    Article  Google Scholar 

  11. Moody, J., Wu, L., Liao, Y., Saffell, M.: Performance functions and reinforcement learning for trading systems and portfolios. J. Forecast. 17(5–6), 441–470 (1998)

    Article  Google Scholar 

  12. Mozer, M.C.: A focused back-propagation algorithm for temporal pattern recognition. Complex Syst. 3(4), 349–381 (1989)

    MATH  Google Scholar 

  13. Vaswani, A., et al.: Attention is all you need. arXiv preprint arXiv:1706.03762 (2017)

  14. Wang, X., Wang, Y., Weng, B., Vinel, A.: Stock2Vec: a hybrid deep learning framework for stock market prediction with representation learning and temporal convolutional network. arXiv preprint arXiv:2010.01197 (2020)

  15. Williams, R.J., Zipser, D.: A learning algorithm for continually running fully recurrent neural networks. Neural Comput. 1(2), 270–280 (1989)

    Article  Google Scholar 

  16. Xiong, Z., Liu, X.Y., Zhong, S., Yang, H., Walid, A.: Practical deep reinforcement learning approach for stock trading. arXiv preprint arXiv:1811.07522 (2018)

  17. Xu, K., Zhang, Y., Ye, D., Zhao, P., Tan, M.: Relation-aware transformer for portfolio policy learning. In: IJCAI (2020)

    Google Scholar 

Download references

Acknowledgements

This work was supported in part by the National Nature Science Foundation of China under Grant U2013201 and in part by the Pearl River Talent Plan of Guangdong Province under Grant 2019ZT08X603.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jianyong Chen .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Hu, S., Zheng, H., Chen, J. (2021). A Novel Deep Reinforcement Learning Framework for Stock Portfolio Optimization. In: Mantoro, T., Lee, M., Ayu, M.A., Wong, K.W., Hidayanto, A.N. (eds) Neural Information Processing. ICONIP 2021. Communications in Computer and Information Science, vol 1516. Springer, Cham. https://doi.org/10.1007/978-3-030-92307-5_24

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-92307-5_24

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-92306-8

  • Online ISBN: 978-3-030-92307-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics