Abstract
Visual and inertial sensors are used collaboratively in many applications because of their complementary properties. The problem associated with sensor fusion is relative coordinate transformations. This paper presents a quaternion-based method to estimate the relative rotation between visual and inertial sensors. Rotation between a camera and an inertial measurement unit (IMU) is represented by quaternions, which are separately measured to allow the sensor to be optimized individually. Relative quaternions are used so that the global reference is not required to be known. The accuracy of the coordinate transformation was evaluated by comparing with a ground-truth tracking system. The experiment analysis proves the effectiveness of the proposed method in terms of accuracy and robustness.
Hongsheng He’s—work was supported in part by NSFC grant 61305114.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Tao, Y., Hu, H., Zhou, H.: Integration of vision and inertial sensors for 3d arm motion tracking in home-based rehabilitation. Int. J. Robot. Res. 26(6), 607–624 (2007)
He, H., Li, Y., Guan, Y., Tan, J.: Wearable ego-motion tracking for blind navigation in indoor environments. IEEE Trans. Autom. Sci. Eng. 12(4), 1181–1190 (2015)
Hwangbo, M., Kim, J.S., Kanade, T.: Inertial-aided klt feature tracking for a moving camera. In: IEEE/RSJ International Conference on Intelligent Robots and Systems: IROS 2009, pp. 1909–1916 (2009)
Achtelik, M.W., Lynen, S., Weiss, S., Kneip, L., Chli, M., Siegwart, R.: Visual-inertial slam for a small helicopter in large outdoor environments. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2651–2652. IEEE (2012)
Lobo, J., Dias, J.: Vision and inertial sensor cooperation using gravity as a vertical reference. Pattern Anal. Mach. Intell. IEEE Trans. 25(12), 1597–1608 (2003)
Alves, J., Lobo, J., Dias, J.: Camera-inertial sensor modelling and alignment for visual navigation. Mach. Intell. Robot. Control 5(3), 103–112 (2003)
Daniilidis, K.: Hand-eye calibration using dual quaternions. Int. J. Robot. Res. 18(3), 286–298 (1999)
Ovrén, H., Forssen, P., Tornqvist, D.: Why would i want a gyroscope on my rgb-d sensor? In: IEEE Workshop on Robot Vision (WORV), pp. 68–75. IEEE (2013)
Mirzaei, F.M., Roumeliotis, S.I.: A kalman filter-based algorithm for imu-camera calibration: Observability analysis and performance evaluation. Robot. IEEE Trans. 24(5), 1143–1156 (2008)
Lobo, J., Dias, J.: Relative pose calibration between visual and inertial sensors. Int. J. Robot. Res. 26(6), 561–575 (2007)
Gemeiner, P., Einramhof, P., Vincze, M.: Simultaneous motion and structure estimation by fusion of inertial and vision data. Int. J. Robot. Res. 26(6), 591–605 (2007)
Kelly, J., Sukhatme, G.S.: Visual-inertial sensor fusion: Localization, mapping and sensor-to-sensor self-calibration. Int. J. Robot. Res. 30(1), 56–79 (2011)
Hol, J.D., Schön, T.B., Gustafsson, F.: A new algorithm for calibrating a combined camera and imu sensor unit. In: 10th International Conference on Control, Automation, Robotics and Vision, ICARCV 2008, pp. 1857–1862. IEEE (2008)
Fleps, M., Mair, E., Ruepp, O., Suppa, M., Burschka, D.: Optimization based imu camera calibration. In: 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3297–3304. IEEE (2011)
Lang, P., Pinz, A.: Calibration of hybrid vision/inertial tracking systems. In: InerVis, Barcelona, Spain (2005)
Shoemake, K.: Quaternion calculus for animation. SIGGRAPH Course 23, 15–23 (1989)
Dam, E.B., Koch, M., Lillholm, M.: Quaternions, interpolation and animation. Københavns Universitet, Datalogisk Institut (1998)
Keat, J.: Analysis of least-squares attitude determination routine doaop. Technical Report CSC/TM-77/6034, Computer Sciences Corp, Technical Report (1977)
Markley, F.L., Mortari, D.: How to estimate attitude from vector observations. Technical Report, NASA (1999)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
He, H., Li, Y., Tan, J. (2016). Rotational Coordinate Transformation for Visual-Inertial Sensor Fusion. In: Agah, A., Cabibihan, JJ., Howard, A., Salichs, M., He, H. (eds) Social Robotics. ICSR 2016. Lecture Notes in Computer Science(), vol 9979. Springer, Cham. https://doi.org/10.1007/978-3-319-47437-3_42
Download citation
DOI: https://doi.org/10.1007/978-3-319-47437-3_42
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-47436-6
Online ISBN: 978-3-319-47437-3
eBook Packages: Computer ScienceComputer Science (R0)