Abstract
This work presents a vision system based on the YOLO algorithm to identify static objects that could be obstacles in the path of a mobile robot. In order to identify the objects and its distances a Microsoft Kinect sensor was used. In addition, a Nvidia Jetson TX2 GPU was used to increase the image processing algorithm performance. Our experimental results indicate that the YOLO network has detected all the predefined obstacles for which it has been trained with good reliability and the calculus of the distance using the depth information returned by Microsoft Kinect had an error below of 3,64%.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Conference on Computer Vision and Pattern Recognition - CVPR (2016)
Silva, R.M., Cuadros, M.A.S.L., Gamarra, D.F.T.: Comparison of a backstepping and a fuzzy controller for tracking a trajectory with a mobile robot. In: The International Conference on Intelligent Systems Design and Applications (ISDA), Vellore (2018)
Moura, R.G.: Utilizando o Microsoft Kinect na obtenção de atributos antropométricos. 2014. Work of conclusion of course (Bachelor in Information Systems) - Lutheran University Center of Palmas, Palmas, TO (2014)
Pegas, G.L.: Rastreamento visual para robôs usando microsoft kinect. Completion of course work (Electrical Engineer) - Universidade Federal of São Carlos, São Carlos, SP (2014)
Wang, Y., Song, G., Qiao, G., Zhang, Y., Zhang, J., Wang, W.: Wheeled robot control based on gesture recognition using the Kinect sensor. In: IEEE International Conference on Robotics and Biomimetics (ROBIO), Shenzhen (2013)
Fadli, H., Machbub, C., Hidayat, E.: Human gesture imitation on NAO humanoid robot using kinect based on inverse kinematics method. In: International Conference on Advanced Mechatronics, Intelligent Manufacture, and Industrial Automation (ICAMIMIA), Surabaya (2017)
Tenguria, R., Parkhedar, S., Modak, N., Madan, R., Tondwalkar, A.: Design framework for general purpose object recognition on a robotic plataform. In: International Conference on Communication and Signal Processing, India (2017)
Lucian, A., Sandu, A., Radu, O., Moldovan, D.: Human leg detection from depth sensing. R&D Innovation Center, Romênia (2018)
Bersan, D., Martins, R., Campos, M., Nascimento, E.R.: Semantic map augmentation for robot navigation: a learning approach based on visual and depth data. In: Brazilian Symposium on Robotics e 2018 Workshop on Robotics in Education (2018)
Zhou, J., Feng, L., Chellali, R., Zhu, H.: Detecting and tracking objects in HRI: YOLO networks for the NAO “I see you” Function. In: 27th IEEE International Symposium on Robot and Human Interactive Communication (2018)
Gu, S., Chen, X., Zeng, N., Wang, X.: A deep learning tennis ball collection robot and the implementation on Nvidia Jetson TX1 board. In: Conference on Advanced Intelligent Mechatronics (AIM) (2018)
Zapf, M.P., Gupta, A., Saiki, L.Y.M., Kawanabe, M.: Data-driven, 3-D classification of person-object relationship and semantic context clustering for robotics and AI application. In: 27th IEEE International Symposium on Robot and Human Interactive Communication (2018)
Maolanon, P., Sukvichai, K., Chayopitak, N., Takashi, A.: Indoor room identify and mapping with virtual based SLAM using furnitures and household objects relationship based on CNNs. In: 10th International Conference of Information and Communication Technology for Embedded Systems (IC-ICTES) (2019)
Zhao, X., Jia, H., Ni, Y.: A novel three-dimensional object detection with the modified you only look once method. Int. J. Adv. Robot. Syst. 15(2), 1–13 (2018)
Pfitscher, M., Welfer, D., Nascimento, E.J., Cuadros, M.A.S.L., Gamarra, D.F.T.: Users activity gesture recognition on kinect sensor using convolutional neural networks and FastDTW for controlling movements of a mobile robot. Inteligência Artificial 22, 121–134 (2019)
Pfitscher, M., Welfer, D., Cuadros, M.A.S.L., Gamarra, D.F.T.: Activity gesture recognition on Kinect sensor using convolutional neural networks and FastDTW for the MSRC-12 dataset. In: International Conference on Intelligent Systems Design and Applications (ISDA), Vellore (2018)
Redmon, J., Farhadi, A.: YOLO9000: better, faster, stronger. In: Conference on Computer Vision and Pattern Recognition - CVPR (2017)
Openkinect: https://openkinect.org/wiki/Main_Page. Accessed 6 May 2018
Dos Reis, D.H., Welfer, D., Cuadros, M.A.S.L., Gamarra, D.F.T.: Mobile robot navigation using object recognition software with RGBD images and the YOLO algorithm. Appl. Artif. Intell. 33, 1290–1305 (2019)
Acknowledgment
We would like to thank the National Institute of Space Research (INPE) located at the Federal University of Santa Maria (UFSM), especially Adriano Petry, for their cooperation.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Henke dos Reis, D., Welfer, D., de Souza Leite Cuadros, M.A., Tello Gamarra, D.F. (2021). Object Recognition Software Using RGBD Kinect Images and the YOLO Algorithm for Mobile Robot Navigation. In: Abraham, A., Siarry, P., Ma, K., Kaklauskas, A. (eds) Intelligent Systems Design and Applications. ISDA 2019. Advances in Intelligent Systems and Computing, vol 1181. Springer, Cham. https://doi.org/10.1007/978-3-030-49342-4_25
Download citation
DOI: https://doi.org/10.1007/978-3-030-49342-4_25
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-49341-7
Online ISBN: 978-3-030-49342-4
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)