Skip to main content
Log in

Improving deep reinforcement learning by safety guarding model via hazardous experience planning

  • Letter
  • Published:
Frontiers of Computer Science Aims and scope Submit manuscript

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

References

  1. Kai A, Deisenroth M P, Brundage M, Bharath A A. Deep reinforcement learning: a brief survey. IEEE Signal Processing Magazine, 2017, 34(6): 26–38

    Article  Google Scholar 

  2. Cheng R, Orosz G, Murray R M, Burdick J W. End-to-end safe reinforcement learning through barrier functions for safety-critical continuous control tasks. In: Proceedings of the AAAI Conference on Artificial Intelligence. 2019, 3387–3395

  3. Saunders W, Sastry G, Stuhlmueller A, Evans O. Trial without error: towards safe reinforcement learning via human intervention. In: Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems. 2018, 2067–2069

  4. Achiam J, Held D, Tamar A, Abbeel P. Constrained policy optimization. In: Proceedings of the International Conference on Machine Learning. 2017, 22–31

  5. García J, Fernández F. A comprehensive survey on safe reinforcement learning. Journal of Machine Learning Research, 2015, 16: 1437–1480

    MathSciNet  MATH  Google Scholar 

  6. Chatzilygeroudis K, Vassiliades V, Mouret J B. Reset-free trial-and-error learning for robot damage recovery. Robotics and Autonomous Systems, 2018, 100: 236–250

    Article  Google Scholar 

  7. Zhu F, Wu W, Fu Y, Liu Q. A dual deep network based secure deep reinforcement learning method. Chinese Journal of Computers, 2019, 42(8): 1812–1826

    Google Scholar 

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China (Grant No. 61303108), Natural Science Foundation of Jiangsu Province (BK20211102), Suzhou Key Industries Technological Innovation-Prospective Applied Research Project (SYG 201804); A Project Funded by the Priority Academic Program Development of Jiangsu Higher Education Institutions.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Fei Zhu.

Additional information

Supporting information

The supporting information is available online at journal. hep. com. cn and link. springer. com

Electronic supplementary material

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Peng, P., Zhu, F., Ling, X. et al. Improving deep reinforcement learning by safety guarding model via hazardous experience planning. Front. Comput. Sci. 16, 164320 (2022). https://doi.org/10.1007/s11704-021-0250-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11704-021-0250-y

Navigation