Abstract
Around view monitor (AVM) based parking slot detection (PSD) is an important task for autonomous parking. Among various algorithms proposed for this task, such as parking line detection or segmentation-based algorithms, corner point detection-based algorithms achieve higher precision since they detect the most fundamental feature of a parking slot. However, such algorithms cannot detect the parking slot when it is partially occluded by the vehicle: even a slight occlusion of one or two corner points significantly disrupts the detection. Moreover, they do not distinguish apart different parking slots when multiple slots exist. To address these problems, this paper proposes a parking slot detection and tracking (PSDT) algorithm using deep neural network, namely PSDT-Net. In order to accurately infer the parking slot position partially occluded in the camera’s blind spot, PSDT-Net detects not only the visible corner points but additional marking points near the blind spot. In addition, using DeepSORT algorithm, PSDT-Net assigns unique labels to different parking slots and tracks their location over time: initially targeted parking slot can be consistently tracked throughout the parking process. Real vehicle experiments show that PSDT-Net can detect and track the parking slot precisely during parking, compared to existing corner point detection-based PSD algorithms.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
References
Lee, S., Hyeon, D., Park, G., Baek, I.J., Kim, S.W., Seo, S.W.: Directional-dbscan: parking-slot detection using a clustering method in around-view monitoring system. In: 2016 IEEE Intelligent Vehicles Symposium (IV), pp. 349–354. IEEE (2016)
Li, Q., Lin, C., Zhao, Y.: Geometric features-based parking slot detection. Sensors 18(9), 2821 (2018)
Hamada, K., Hu, Z., Fan, M., Chen, H.: Surround view based parking lot detection and tracking. In: 2015 IEEE Intelligent Vehicles Symposium (IV), pp. 1106–1111. IEEE (2015)
Wu, Y., Yang, T., Zhao, J., Guan, L., Jiang, W.: Vh-hfcn based parking slot and lane markings segmentation on panoramic surround view. In: 2018 IEEE Intelligent Vehicles Symposium (IV), pp. 1767–1772. IEEE (2018)
Jiang, W., Wu, Y., Guan, L., Zhao, J.: Dfnet: semantic segmentation on panoramic images with dynamic loss weights and residual fusion block. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 5887–5892. IEEE (2019)
Zhang, L., Huang, J., Li, X., Xiong, L.: Vision-based parking-slot detection: a dcnn-based approach and a large-scale benchmark dataset. IEEE Trans. Image Process. 27(11), 5350–5364 (2018)
Zhang, L., Li, X., Huang, J., Shen, Y., Wang, D.: Vision-based parking-slot detection: a benchmark and a learning-based approach. Symmetry 10(3), 64 (2018)
Li, W., Cao, L., Yan, L., Li, C., Feng, X., Zhao, P.: Vacant parking slot detection in the around view image based on deep learning. Sensors 20(7), 2138 (2020)
Suhr, J.K., Jung, H.G.: Fully-automatic recognition of various parking slot markings in around view monitor (avm) image sequences. In: 2012 15th International IEEE Conference on Intelligent Transportation Systems, pp. 1294–1299. IEEE (2012)
Suhr, J.K., Jung, H.G.: Full-automatic recognition of various parking slot markings using a hierarchical tree structure. Opt. Eng. 52(3), 037203 (2013)
Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481–2495 (2017)
Harris, C., Stephens, M., et al.: A combined corner and edge detector. In: Alvey Vision Conference, vol. 15, pp. 10–5244. Citeseer (1988)
Redmon, J., Farhadi, A.: Yolo9000: better, faster, stronger. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7263–7271 (2017)
Girshick, R.: Fast r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1440–1448 (2015)
Wojke, N., Bewley, A., Paulus, D.: Simple online and realtime tracking with a deep association metric. In: 2017 IEEE International Conference on Image Processing (ICIP), pp. 3645–3649. IEEE (2017)
Jocher, G., et al.: ultralytics/yolov5: v5.0–YOLOv5-P6 1280 models, AWS, Supervise.ly and YouTube integrations, April 2021
He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729–9738 (2020)
Bewley, A., Ge, Z., Ott, L., Ramos, F., Upcroft, B.: Simple online and realtime tracking. In: 2016 IEEE International Conference on Image Processing (ICIP), pp. 3464–3468. IEEE (2016)
Kim, M., Alletto, S., Rigazio, L.: Similarity mapping with enhanced siamese network for multi-object tracking. arXiv preprint arXiv:1609.09156 (2016)
Zhang, S., et al.: Tracking persons-of-interest via adaptive discriminative features. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 415–433. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46454-1_26
Leal-Taixé, L., Canton-Ferrer, C., Schindler, K.: Learning by tracking: siamese cnn for robust target association. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 33–40 (2016)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Park, Y., Ahn, J., Park, J. (2022). Deep Learning Based Parking Slot Detection and Tracking: PSDT-Net. In: Kim, J., et al. Robot Intelligence Technology and Applications 6. RiTA 2021. Lecture Notes in Networks and Systems, vol 429. Springer, Cham. https://doi.org/10.1007/978-3-030-97672-9_26
Download citation
DOI: https://doi.org/10.1007/978-3-030-97672-9_26
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-97671-2
Online ISBN: 978-3-030-97672-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)