Skip to main content

Realizing Human-like Walking and Running with Feedforward Enhanced Reinforcement Learning

  • Conference paper
  • First Online:
Intelligent Robotics and Applications (ICIRA 2023)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14270))

Included in the following conference series:

Abstract

Locomotion control of legged robots is a challenging problem. Recently, reinforcement learning has been applied to legged locomotion and made a great success. However, the reward signal design remains a challenging problem to produce a humanlike motion such as walking and running. Although imitation learning provides a way to mimic the behavior of humans or animals, the obtained motion may be restricted due to the over-constrained property of this method. Here we propose a novel and simple way to generate humanlike behavior by using feedforward enhanced reinforcement learning (FERL). In FERL, the control action is composed of a feedforward part and a feedback part, where the feedforward part is a periodic time-dependent signal generated by a state machine and the feedback part is a state-dependent signal obtained by a neural network. By using FERL with a simple feedforward of two feet stepping up and down alternately, we achieve humanlike walking and running for a simulated biped robot, Ranger Max. Comparison results show that the feedforward is key to generating humanlike behavior, while the policy trained with no feedforward only results in some strange gaits. FERL may also be extended to other legged robots to generate various locomotion styles, which provides a competitive alternative for imitation learning.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Heess, N., et al.: Emergence of locomotion behaviours in rich environments. arXiv preprint arXiv:1707.02286 (2017)

  2. Hwangbo, J., et al.: Learning agile and dynamic motor skills for legged robots. Sci. Robot. 4(26), eaau5872 (2019)

    Google Scholar 

  3. Xie, Z., Berseth, G., Clary, P., Hurst, J., van de Panne, M.: Feedback control for cassie with deep reinforcement learning. In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1241–1246. IEEE (2018)

    Google Scholar 

  4. Li, Z., et al.: Reinforcement learning for robust parameterized locomotion control of bipedal robots. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 2811–2817. IEEE (2021)

    Google Scholar 

  5. Peng, X.B., Coumans, E., Zhang, T., Lee, T.W., Tan, J., Levine, S.: Learning agile robotic locomotion skills by imitating animals. arXiv preprint arXiv:2004.00784 (2020)

  6. Shao, Y., Jin, Y., Liu, X., He, W., Wang, H., Yang, W.: Learning free gait transition for quadruped robots via phase-guided controller. IEEE Robot. Autom. Let. 7(2), 1230–1237 (2021)

    Article  Google Scholar 

  7. Wu, Q., Zhang, C., Liu, Y.: Custom sine waves are enough for imitation learning of bipedal gaits with different styles. In: 2022 IEEE International Conference on Mechatronics and Automation (ICMA), pp. 499–505. IEEE (2022)

    Google Scholar 

  8. Siekmann, J., Godse, Y., Fern, A., Hurst, J.: Sim-to-real learning of all common bipedal gaits via periodic reward composition. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 7309–7315. IEEE (2021)

    Google Scholar 

  9. Siekmann, J., Green, K., Warila, J., Fern, A., Hurst, J.: Blind bipedal stair traversal via sim-to-real reinforcement learning. arXiv preprint arXiv:2105.08328 (2021)

  10. Bellegarda, G., Chen, Y., Liu, Z., Nguyen, Q.: Robust high-speed running for quadruped robots via deep reinforcement learning. In: 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 10364–10370. IEEE (2022)

    Google Scholar 

  11. Vollenweider, E., Bjelonic, M., Klemm, V., Rudin, N., Lee, J., Hutter, M.: Advanced skills through multiple adversarial motion priors in reinforcement learning. arXiv preprint arXiv:2203.14912 (2022)

  12. Escontrela, A., et al.: Adversarial motion priors make good substitutes for complex reward functions. In: 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 25–32. IEEE (2022)

    Google Scholar 

  13. Li, T., Geyer, H., Atkeson, C.G., Rai, A.: Using deep reinforcement learning to learn high-level policies on the atrias biped. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 263–269. IEEE (2019)

    Google Scholar 

  14. Tan, W., et al.: A hierarchical framework for quadruped locomotion based on reinforcement learning. In: 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 8462–8468. IEEE (2021)

    Google Scholar 

  15. Castillo, G. A., Weng, B., Zhang, W., Hereid, A.: Robust feedback motion policy design using reinforcement learning on a 3d digit bipedal robot. In: 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5136–5143. IEEE (2021)

    Google Scholar 

  16. Wang, Z., Wei, W., Xie, A., Zhang, Y., Wu, J., Zhu, Q.: Hybrid bipedal locomotion based on reinforcement learning and heuristics. Micromachines 13(10), 1688 (2022)

    Article  Google Scholar 

  17. Bhounsule, P.A., et al.: Low-bandwidth reflex-based control for lower power walking: 65 km on a single battery charge. Int. J. Robot. Res. 33(10), 1305–1321 (2014)

    Article  Google Scholar 

Download references

Acknowledgment

This work was supported by the National Natural Science Foundation of China No. 62003188 and No.92248304.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Linqi Ye .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ye, L., Wang, X., Liang, B. (2023). Realizing Human-like Walking and Running with Feedforward Enhanced Reinforcement Learning. In: Yang, H., et al. Intelligent Robotics and Applications. ICIRA 2023. Lecture Notes in Computer Science(), vol 14270. Springer, Singapore. https://doi.org/10.1007/978-981-99-6492-5_38

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-6492-5_38

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-6491-8

  • Online ISBN: 978-981-99-6492-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics