Abstract
Detection of fiducial markers in challenging lighting conditions can be useful in fields such as industry, medicine, or any other setting in which lighting cannot be controlled (e.g., outdoor environments or indoors with poor lighting). However, explicitly dealing with such conditions has not been done before. Hence, we propose DeepArUco, a deep learning-based framework that aims to detect ArUco markers in lighting conditions where the classical ArUco implementation fails. The system is built around Convolutional Neural Networks, performing the job of detecting and decoding ArUco markers. A method to generate synthetic data to train the networks is also proposed. Furthermore, a real-life dataset of ArUco markers in challenging lighting conditions is introduced and used to evaluate our system, which will be made publicly available alongside the implementation.
Code available in GitHub: https://github.com/AVAuco/deeparuco/.
Supported by the MCIN Project TED2021-129151B-I00/AEI/10.13039/ 501100011033/ European Union NextGenerationEU/PRTR, project PID2019-103871GB-I00 of the Spanish Ministry of Economy, Industry and Competitiveness, and project PAIDI P20_00430 of the Junta de Andalucía, FEDER. We thank H. Sarmadi his contribution to the data preparation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Fiala, M.: Designing highly reliable fiducial markers. IEEE PAMI 32(7), 1317–1324 (2010). https://doi.org/10.1109/TPAMI.2009.146
Garrido-Jurado, S., Muñoz-Salinas, R., Madrid-Cuevas, F.J., Marín-Jiménez, M.J.: Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recogn. 47(6), 2280–2292 (2014). https://doi.org/10.1016/j.patcog.2014.01.005
Hu, D., DeTone, D., Malisiewicz, T.: Deep ChArUco: dark ChArUco marker pose estimation. In: CVPR (2019)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: NeurIPS, vol. 25 (2012)
Li, B., Wu, J., Tan, X., Wang, B.: ArUco marker detection under occlusion using convolutional neural network. In: International Conference on Automation, Control and Robotics Engineering (CACRE), pp. 706–711 (2020)
Lin, T.Y., et al.: Microsoft COCO: common objects in context. arXiv:1405.0312 [cs] (2015)
Mondéjar-Guerra, V.M., Garrido-Jurado, S., Muñoz-Salinas, R., Marín-Jiménez, M.J., Carnicer, R.M.: Robust identification of fiducial markers in challenging conditions. Expert Syst. Appl. 93, 336–345 (2018). https://doi.org/10.1016/j.eswa.2017.10.032
Olson, E.: AprilTag: a robust and flexible visual fiducial system. In: 2011 IEEE International Conference on Robotics and Automation, pp. 3400–3407 (2011). https://doi.org/10.1109/ICRA.2011.5979561. iSSN: 1050-4729
Ramírez, F.J.R., Muñoz-Salinas, R., Carnicer, R.M.: Tracking fiducial markers with discriminative correlation filters. Image Vis. Comput. 107, 104094 (2021). https://doi.org/10.1016/j.imavis.2020.104094
Romero-Ramirez, F.J., Muñoz-Salinas, R., Medina-Carnicer, R.: Speeded up detection of squared fiducial markers. Image Vis. Comput. 76, 38–47 (2018). https://doi.org/10.1016/j.imavis.2018.05.004
Sani, M.F., Karimian, G.: Automatic navigation and landing of an indoor AR. Drone quadrotor using ArUco marker and inertial sensors. In: 2017 International Conference on Computer and Drone Applications (IConDA), pp. 102–107 (2017). https://doi.org/10.1109/ICONDA.2017.8270408
Sarmadi, H., Muñoz-Salinas, R., Álvaro Berbís, M., Luna, A., Medina-Carnicer, R.: 3D Reconstruction and alignment by consumer RGB-D sensors and fiducial planar markers for patient positioning in radiation therapy. Comput. Methods Programs Biomed. 180, 105004 (2019). https://doi.org/10.1016/j.cmpb.2019.105004
Sarmadi, H., Muñoz-Salinas, R., Álvaro Berbís, M., Luna, A., Medina-Carnicer, R.: Joint scene and object tracking for cost-Effective augmented reality guided patient positioning in radiation therapy. Comput. Methods Programs Biomed. 209, 106296 (2021). https://doi.org/10.1016/j.cmpb.2021.106296
Strisciuglio, N., Vallina, M.L., Petkov, N., Muñoz Salinas, R.: Camera localization in outdoor garden environments using artificial landmarks. In: IWOBI, pp. 1–6 (2018)
Wang, J., Olson, E.: AprilTag 2: efficient and robust fiducial detection. In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4193–4198 (2016). https://doi.org/10.1109/IROS.2016.7759617
Zhang, Z., Hu, Y., Yu, G., Dai, J.: DeepTag: a general framework for fiducial marker design and detection. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 2931–2944 (2023). https://doi.org/10.48550/arXiv.2105.13731
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Berral-Soler, R., Muñoz-Salinas, R., Medina-Carnicer, R., Marín-Jiménez, M.J. (2023). DeepArUco: Marker Detection and Classification in Challenging Lighting Conditions. In: Pertusa, A., Gallego, A.J., Sánchez, J.A., Domingues, I. (eds) Pattern Recognition and Image Analysis. IbPRIA 2023. Lecture Notes in Computer Science, vol 14062. Springer, Cham. https://doi.org/10.1007/978-3-031-36616-1_16
Download citation
DOI: https://doi.org/10.1007/978-3-031-36616-1_16
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-36615-4
Online ISBN: 978-3-031-36616-1
eBook Packages: Computer ScienceComputer Science (R0)