Abstract
Localization is an important task in various fields. Especially, a robust and accurate localization system is required for a mobile device to be used in AR (augmented reality) or MR (mixed reality) applications. In this paper, we present an accurate visual-inertial odometry algorithm based on both point and line features. This algorithm utilizes a sliding window non-linear optimization-based method to reduce the computational cost. We implemented our proposed algorithm on a smartphone. Also, we show that our proposed algorithm works well in indoor and outdoor environments in real-time.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Kerl, M., Sturm, J., Cremers, D.: Dense visual SLAM for RGB-D cameras. In: Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2100–2106 (2013)
Bloesch, M., Omari, S., Hutter, M., Siegwart, R.: Robust visual inertial odometry using a direct EKF-based approach. In: Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 298–304 (2015)
Engel, J., Koltun, V., Cremers, D.: Direct sparse odometry. IEEE Trans. Pattern Anal. Mach. Intell. 40(3), 611–625 (2017)
Schneider, T., et al.: MapLab: an open framework for research in visual-inertial mapping and localization. IEEE Robot. Autom. Lett. 3(3), 1418–1425 (2017)
Davison, A.J., Reid, I.D., Molton, N.D., Stasse, O.: MonoSLAM: real-time single camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 29(6), 1052–1067 (2007)
Leutenegger, S., Lynen, S., Bosse, M., Siegwart, R., Furgale, P.: Keyframe-based visual-inertial odometry using nonlinear optimization. Int. J. Robot. Res. 34(3), 314–334 (2015)
Qin, T., Li, P., Shen, S.: VINS-Mono: a robust and versatile monocular visual-inertial state estimator. IEEE Trans. Rob. 34(4), 1004–1020 (2018)
Campos, C., Elvira, R., Rodríguez, J.J.G., Montiel, J.M.M., Tardós, J.D.: ORB-SLAM3: an accurate open-source library for visual, visual-inertial, and multimap SLAM. IEEE Trans. Robot. 37, 1874–1890 (2021)
He, Y., Zhao, J., Guo, Y., He, W., Yuan, K.: PL-VIO: tightly-coupled monocular visual-inertial odometry using point and line features. Sensors 18(4), 1159 (2018)
Pumarola, A., Vakhitov, A., Agudo, A., Sanfeliu, A., Moreno-Noguer, F.: PL-SLAM: real-time monocular visual SLAM with points and lines. In: Proceedings of IEEE International Conference on Robotics and Automation (ICRA), pp. 4503–4508 (2017)
Jung, K., Kim, Y., Lim, H., Myung, H.: ALVIO: adaptive line and point feature-based visual inertial odometry for robust localization in indoor environments. In: Proceedings of RiTA, pp. 171–184 (2020)
Lim, H., Kim, Y., Jung, K., Hu, S., Myung, H.: Avoiding degeneracy for monocular visual SLAM with point and line features. arXiv preprint arXiv:2103.01501 (2021)
Rosinol, A., Abate, M., Chang, Y., Carlone, L.: Kimera: an open-source library for real-time metric-semantic localization and mapping. In: Proceedings of IEEE International Conference on Robotics and Automation (ICRA), pp. 1689–1696 (2020)
Li, X., He, Y., Lin, J., Liu, X.: Leveraging planar regularities for point line visual-inertial odometry. In: Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5120–5127 (2020)
Li, P., Qin, T., Hu, B., Zhu, F., Shen, S.: Monocular visual-inertial state estimation for mobile augmented reality. In: Proceedings of IEEE International Symposium on Mixed and Augmented Reality (ISMAR), pp. 11–21 (2017)
Von, G., Rafael, G., Jakubowicz, J., Morel, J., Randall, G.: LSD: a line segment detector. Image Process. On Line 2, 35–55 (2012)
Zhang, L., Koch, R.: An efficient and robust line segment matching approach based on LBD descriptor and pairwise geometric consistency. J. Vis. Commun. Image Represent. 24(7), 794–805 (2013)
Bradski, G.: The openCV library. Dr. Dobb’s J. Softw. Tools Professional Program. 25(11), 120–123 (2000)
Acknowledgement
This work was supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No. 2021-0-00230, development of real\(\cdot \)virtual environmental analysis based adaptive interaction technology) and the Defense Challengeable Future Technology Program of Agency for Defense Development, Republic of Korea. The students are supported by Korea Ministry of Land, Infrastructure and Transport (MOLIT) as “Innovative Talent Education Program for Smart City” and BK21 FOUR.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Kim, Y., Lim, H., Myung, H. (2022). Point and Line Feature-Based VIO for Mobile Devices. In: Kim, J., et al. Robot Intelligence Technology and Applications 6. RiTA 2021. Lecture Notes in Networks and Systems, vol 429. Springer, Cham. https://doi.org/10.1007/978-3-030-97672-9_25
Download citation
DOI: https://doi.org/10.1007/978-3-030-97672-9_25
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-97671-2
Online ISBN: 978-3-030-97672-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)