Abstract
This paper focuses on developing a Deep Reinforcement Learning (DRL)—based agent for real-time trajectory planning and tracking in a simulated parking environment, specifically low-speed maneuvers in a parking area with comb-shaped spaces and a random distribution of non-player vehicles. We rely on CARLA as a virtual driving simulator due to its realistic graphics and physics simulation capabilities, and on the Gymnasium and Stable-Baselines3 toolkits for training the agent. We show that the agent is able to achieve a success rate of 97% in reaching the target parking lot without collisions. However, integrating CARLA with DRL frameworks poses challenges, such as determining suitable environment and neural network update frequencies. Despite these issues, the results demonstrate the potential of DRL agents in developing automated driving functions.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Lazzaroni L, Bellotti F, Capello A, Cossu M, De Gloria A, Berta R (2023) Deep reinforcement learning for automated car parking. In: Berta R, De Gloria A (eds) Applications in electronics pervading industry, environment and society. Springer Nature Switzerland, Cham, pp 125–130. https://doi.org/10.1007/978-3-031-30333-3_16
Capello A, Forneris L, Pighetti A, Bellotti F, Lazzaroni L, Cossu M, De Gloria A, Berta R (2023) Investigating high-level decision making for automated driving. In: Berta R, De Gloria A (eds) Applications in electronics pervading industry, environment and society. Springer Nature Switzerland, Cham, pp 307–311. https://doi.org/10.1007/978-3-031-30333-3_41
Leurent E (2018) An environment for autonomous driving decision-making. https://github.com/eleurent/highway-env
Dosovitskiy A, Ros G, Codevilla F, Lopez A, Koltun V (2017) CARLA: an open urban driving simulator. http://arxiv.org/abs/1711.03938
Pérez-Gil Ó et al (2022) Deep reinforcement learning based control for autonomous vehicles in CARLA. Multimed Tools Appl. 81:3553–3576. https://doi.org/10.1007/s11042-021-11437-3
GutiĂ©rrez-Moreno R, Barea R, LĂ³pez-GuillĂ©n E, Araluce J, Bergasa LM (2022) Reinforcement learning-based autonomous driving at intersections in CARLA simulator. Sensors 22:8373. https://doi.org/10.3390/s22218373
Brockman G, Cheung V, Pettersson L, Schneider J, Schulman J, Tang J, Zaremba W (2016) OpenAI Gym. http://arxiv.org/abs/1606.01540
Raffin A, Hill A, Gleave A, Kanervisto A, Ernestus M, Dormann N (2021) Stable-Baselines3: reliable reinforcement learning implementations. J Mach Learn Res 22:1–8
NVIDIA PhysX (2023). https://github.com/NVIDIA-Omniverse/PhysX
Schulman J, Wolski F, Dhariwal P, Radford A, Klimov O (2017) Proximal policy optimization algorithms, http://arxiv.org/abs/1707.06347
Bengio Y, Louradour J, Collobert R, Weston J (2009) Curriculum learning. In: proceedings of the 26th Annual International Conference on Machine Learning pp 41–48. Association for Computing Machinery, New York, USA. https://doi.org/10.1145/1553374.1553380
Acknowledgements
The authors would like to thank all partners within the Hi-Drive project for their cooperation and valuable contribution. This project has received funding from the European Union's Horizon 2020 research and innovation programme under grant agreement No 101006664. The sole responsibility of this publication lies with the authors. Neither the European Commission nor CINEA—in its capacity of Granting Authority—can be made responsible for any use that may be made of the information this document contains.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Lazzaroni, L., Pighetti, A., Bellotti, F., Capello, A., Cossu, M., Berta, R. (2024). Automated Parking in CARLA: A Deep Reinforcement Learning-Based Approach. In: Bellotti, F., et al. Applications in Electronics Pervading Industry, Environment and Society. ApplePies 2023. Lecture Notes in Electrical Engineering, vol 1110. Springer, Cham. https://doi.org/10.1007/978-3-031-48121-5_50
Download citation
DOI: https://doi.org/10.1007/978-3-031-48121-5_50
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-48120-8
Online ISBN: 978-3-031-48121-5
eBook Packages: EngineeringEngineering (R0)