Skip to main content
Log in

Multi-constrained intelligent gliding guidance via optimal control and DQN

  • Research Paper
  • Published:
Science China Information Sciences Aims and scope Submit manuscript

Abstract

In order to improve the adaptability and robustness of gliding guidance under complex environments and multiple constraints, this study proposes an intelligent gliding guidance strategy based on the optimal guidance, predictor-corrector technique, and deep reinforcement learning (DRL). Longitudinal optimal guidance was introduced to satisfy the altitude and velocity inclination constraints, and lateral maneuvering was used to control the terminal velocity magnitude and position. The maneuvering amplitude was calculated by the analytical prediction of the terminal velocity, and the direction was learned and determined by the deep Q-learning network (DQN). In the direction decision model construction, the state and action spaces were designed based on the flight status and maneuvering direction, and a reward function was proposed using the terminal predicted state and terminal constraints. For DQN training, initial data samples were generated based on the heading-error corridor, and the experience replay pool was managed according to the terminal guidance error. The simulation results show that the intelligent gliding guidance strategy can satisfy various terminal constraints with high precision and ensure adaptability and robustness under large deviations.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Yu J L, Dong X W, Li Q D, et al. Cooperative guidance strategy for multiple hypersonic gliding vehicles system. Chin J Aeronaut, 2020, 33: 990–1005

    Article  Google Scholar 

  2. Lahanier H, Serre L. Trajectory and guidance scheme design for free flight test of hypersonic vehicle. In: Proceedings of the AIAA Guidance, Navigation, and Control Conference, 2017

  3. Joshi A, Sivan K, Amma S S. Predictor-corrector reentry guidance algorithm with path constraints for atmospheric entry vehicles. J Guid Control Dyn, 2007, 30: 1307–1318

    Article  Google Scholar 

  4. Zhu J, Zhang S. Adaptive optimal gliding guidance independent of QEGC. Aerospace Sci Tech, 2017, 71: 373–381

    Article  Google Scholar 

  5. Mnih V, Kavukcuoglu K, Silver D, et al. Human-level control through deep reinforcement learning. Nature, 2015, 518: 529–533

    Article  Google Scholar 

  6. Kober J, Bagnell J A, Peters J. Reinforcement learning in robotics: a survey. Int J Robot Res, 2013, 32: 1238–1274

    Article  Google Scholar 

  7. Bhopale P, Kazi F, Singh N. Reinforcement learning based obstacle avoidance for autonomous underwater vehicle. J Mar Sci Appl, 2019, 18: 228–238

    Article  Google Scholar 

  8. Junell J L, van Kampen E J, de Visser C C, et al. Reinforcement learning applied to a quadrotor guidance law in autonomous flight. In: Proceedings of AIAA Guidance, Navigation, and Control Conference, 2015

  9. Gaudet B, Furfaro R. Missile homing-phase guidance law design using reinforcement learning. In: Proceedings of AIAA Guidance, Navigation, and Control Conference, 2012

  10. Luo Z, Li X S, Wang L X, et al. Multiconstrained gliding guidance based on optimal and reinforcement learning method. Math Problems Eng, 2021, 2021: 1–12

    Article  Google Scholar 

  11. Yang J, You X, Wu G, et al. Application of reinforcement learning in UAV cluster task scheduling. Future Generation Comput Syst, 2019, 95: 140–148

    Article  Google Scholar 

  12. Chai R, Tsourdos A, Savvaris A, et al. Six-DOF spacecraft optimal trajectory planning and real-time attitude control: a deep neural network-based approach. IEEE Trans Neural Netw Learn Syst, 2019, 31: 5005–5013

    Article  Google Scholar 

  13. Hovell K, Ulrich S. On deep reinforcement learning for spacecraft guidance. In: Proceedings of AIAA Scitech 2020 Forum, 2020

  14. Hovell K, Ulrich S. Deep reinforcement learning for spacecraft proximity operations guidance. J Spacecraft Rockets, 2021, 58: 254–264

    Article  Google Scholar 

  15. Woodbury T D, Dunn C, Valasek J. Autonomous soaring using reinforcement learning for trajectory generation. In: Proceedings of the 52nd Aerospace Sciences Meeting, 2014

  16. Julian K D, Kochenderfer M J. Distributed wildfire surveillance with autonomous aircraft using deep reinforcement learning. J Guid Control Dyn, 2019, 42: 1768–1778

    Article  Google Scholar 

  17. Gaudeta B, Furfaroa R, Linares R. Reinforcement meta-learning for angle-only intercept guidance of maneuvering targets. In: Proceedings of the AIAA SciTech 2020 Forum, 2020

  18. Gao W, Zhou X, Pan M, et al. Acceleration control strategy for aero-engines based on model-free deep reinforcement learning method. Aerospace Sci Tech, 2022, 120: 107248

    Article  Google Scholar 

  19. Zhu J, Liu L, Tang G, et al. Highly constrained optimal gliding guidance. Proc Inst Mech Eng Part G-J Aerospace Eng, 2015, 229: 2321–2335

    Article  Google Scholar 

  20. Phillips T H. A common aero vehicle (CAV) model, description, and employment guide. Schafer Corp AFRL and AFSPC, 2003, 27: 1–12

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jianwen Zhu.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhu, J., Zhang, H., Zhao, S. et al. Multi-constrained intelligent gliding guidance via optimal control and DQN. Sci. China Inf. Sci. 66, 132202 (2023). https://doi.org/10.1007/s11432-022-3543-4

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11432-022-3543-4

Keywords

Navigation