Abstract
This paper proposes an improved Q-learning algorithm to solve the problem that the traditional Q-learning algorithm is applied to the path planning of mobile robot in complex environment, the convergence speed is slow due to large number of iterations, and even the actual reward signal is sparse so that the agent cannot get the optimal path. The improved algorithm reduces the number of iterative runs of the agent in the path planning process to improve the convergence speed by further updating the iterative formula of the Q-value. At the same time, adding sparse reward algorithm leads to finding the optimal path successfully. In order to verify the effectiveness of the algorithm, simulation experiments are carried out in two groups of environments: the simple and the complex. The final simulation results show that the improved algorithm can avoid obstacles effectively and find the optimal path to the target position after less iterations, which proves that the performance of the improved algorithm is better than the traditional Q-learning in the path planning.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Marin-Plaza, P., Hussein, A., Martin, D., et al.: Global and local path planning study in a ROS-based research platform for autonomous vehicles. J. Adv. Transp. 2018(PT.1), 1–10 (2018)
Li, G., Yamashita, A., Asama, H., et al.: An efficient improved artificial potential field based regression search method for robot path planning. In: International Conference on Mechatronics & Automation. IEEE (2012)
Chang, L., Shan, L., Jiang, C., et al.: Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment. Auton. Robot. 45(2), 51–76 (2020)
Liu, G., Lao, S.Y., Tan, D.F., et al.: Fast graphic converse method for path planning of anti-ship missile. J. Ballist. (2011)
Tang, X.R., Zhu, Y., Jiang, X.X.: Improved A-star algorithm for robot path planning in static environment. J. Phys. Conf. Ser. 1792(1), 012067 (2021). 8p.
Shen, Z., Hao, Y., Li, K.: Application research of an Adaptive Genetic Algorithms based on information entropy in path planning. Int. J. Inf. 13(6) (2010)
Wang, G., Zhou, J.: Dynamic robot path planning system using neural network. J. Intell. Fuzzy Syst. 40(2), 3055–3063 (2021)
Zhang, C., You, X.: Improved quantum ant colony algorithm of path planning for mobile robot based on grid model. Electron. Sci. Technol. (2016)
Gaskett, C., Fletcher, L., Zelinsky, A.: Reinforcement learning for a vision based mobile robot. IEEE (2018)
Kamalapurkar, R., Walters, P., Rosenfeld, J., et al.: Model-Based Reinforcement Learning for Approximate Optimal Control (2018)
Polydoros, A.S., Nalpantidis, L.: Survey of model-based reinforcement learning: applications on robotics. J. Intell. Rob. Syst. 86(2), 1–21 (2017)
Tong, L., Wang, J.: Application of reinforcement learning in robot path planning. Comput. Simul. 30(012), 351–355 (2013)
Huang, C., Sheng, Z., Jie, X., et al.: Markov Decision Process (2018)
Surya, S., Rakesh, N.: Traffic Congestion Prediction and Intelligent Signaling Based on Markov Decision Process and Reinforcement Learning (2018)
Chu, P., Vu, H., Yeo, D., Lee, B., Um, K., Cho, K.: Robot reinforcement learning for automatically avoiding a dynamic obstacle in a virtual environment. In: Park, J.J.H., Chao, H.-C., Arabnia, H., Yen, N.Y. (eds.) Advanced Multimedia and Ubiquitous Engineering. LNEE, vol. 352, pp. 157–164. Springer, Heidelberg (2015). https://doi.org/10.1007/978-3-662-47487-7_24
Chen, L.P., Mohri, M., Rostamizadeh, A., Talwalkar, A.: Foundations of machine learning, second edition. Statistical Papers 60 (2019)
Watkins, C., Dayan, P.: Q-learning. Machine Learning (1992)
Yang, R., Yan, J.P., Li, X.: Research on sparse reward algorithm for reinforcement learning: theory and experiment . J. Intell. Syst. 15(05), 888–899 (2020)
Riedmiller, M., Hafner, R., Lampe, T., et al.: Learning by Playing-Solving Sparse Reward Tasks from Scratch (2018)
Gullapalli, V., Barto, A.G.: Shaping as a Method for Accelerating Reinforcement Learning. IEEE (2002)
Wiewiora, E.: Reward Shaping. Springer, Heidelberg (2011)
Ng, A.Y., Harada, D., Russell, S.: Policy invariance under reward transformations: theory and application to reward shaping. In: ICML, vol. 99, pp. 278–287 (1999)
Acknowledgments
This works is partly supported by the Natural Science Foundation of Liaoning, China under Grant 2019MS008, Education Committee Project of Liaoning, China under Grant LJ2019003.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Liu, J., Zhang, A., Zhang, Y. (2021). Research on Path Planning Algorithm for Mobile Robot Based on Improved Reinforcement Learning. In: Huang, DS., Jo, KH., Li, J., Gribova, V., Hussain, A. (eds) Intelligent Computing Theories and Application. ICIC 2021. Lecture Notes in Computer Science(), vol 12837. Springer, Cham. https://doi.org/10.1007/978-3-030-84529-2_50
Download citation
DOI: https://doi.org/10.1007/978-3-030-84529-2_50
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-84528-5
Online ISBN: 978-3-030-84529-2
eBook Packages: Computer ScienceComputer Science (R0)