Abstract
The position and orientation of moving platform mainly depends on global positioning system and inertial navigation system in the field of low-altitude surveying, mapping and remote sensing and land-based mobile mapping system. However, GPS signal is unavailable in the application of deep space exploration and indoor robot control. In such circumstances, image-based methods are very important for self-position and orientation of moving platform. Therefore, this paper firstly introduces state of the art development of the image-based self-position and orientation method (ISPOM) for moving platform from the following aspects: 1) A comparison among major image-based methods (i.e., visual odometry, structure from motion, simultaneous localization and mapping) for position and orientation; 2) types of moving platform; 3) integration schemes of image sensor with other sensors; 4) calculation methodology and quantity of image sensors. Then, the paper proposes a new scheme of ISPOM for mobile robot — depending merely on image sensors. It takes the advantages of both monocular vision and stereo vision, and estimates the relative position and orientation of moving platform with high precision and high frequency. In a word, ISPOM will gradually speed from research to application, as well as play a vital role in deep space exploration and indoor robot control.
Similar content being viewed by others
References
Nakju D, Howie C, Wan K C. Accurate relative localization using odometry. In: IEEE International Conference on Robotics and Automation. Taipei, 2003. 1606–1612
Randeniya D I B, Gunaratne M, Sarkar S, et al. Calibration of inertial and vision systems as a prelude to multi-sensor fusion. Transport Res C: Emer, 2008, 16: 255–274
Iván F M, Pascual C, Carol M, et al. Omnidirectional vision applied to unmanned aerial vehicles (UAVs) attitude and heading estimation. Robot Auton Syst, 2010, 58: 809–819
Farid K, Kenzo N. A visual navigation system for autonomous flight of micro air vehicles. In: IEEE International Conference on Intelligent Robots and Systems. Saint Louis, 2009. 3888–3893
Sergio A R F, Vincent F, Philippe B, et al. An embedded multi-modal system for object localization and tracking. In: IEEE Symposium on Intelligent Vehicles. San Diego, 2010. 211–216
Friedrich F, Davide S, Marc P. A constricted bundle adjustment parameterization for relative scale estimation in visual odometry. In: IEEE International Conference on Robotics and Automation. Anchorage, 2010. 1899–1904
Brad G, Andreas K, Mongi A A. A comparison of pose estimation techniques: hardware vs. video. In: Proceedings Of SPIE Unmanned Vehicle Technology VII. Orlando, 2005. 166–173
Munir Z. High resolution relative localisation using two cameras. Robot Auton Syst, 2007, 55: 685–692
Kai O A, Nicola T, Björn T J, et al. Multisensor on-the-fly localization: precision and reliability for applications. Robot Auton Syst, 2001, 34: 131–143
Kim J H, Chung M J, Choi B T. Recursive estimation of motion and a scene model with a two-camera system of divergent view. Pattern Recogn, 2010, 43: 2265–2280
Yang C, Mark M, Larry M. Visual odometry on the mars exploration rovers. In: IEEE International Conference on Systems, Man and Cybernetics. Hawaii, 2005. 903–910
Di K C, Xu F L, Wang J, et al. Photogrammetric processing of rover imagery of the 2003 mars exploration rover mission. ISPRS J Photogramm, 2008, 63: 181–201
Li R X, Di K C, Matthies L H, et al. Rover localization and landing site mapping technology for the 2003 mars exploration rover mission. Photogramm Eng Rem S, 2004, 70: 77–90
Li R X, Squyres S W, Arvidson R E, et al. Initial results of rover localization and topographic mapping for the 2003 mars exploration rover mission. Photogramm Eng Rem S, 2005, 71: 1129–1144
Jochen K, Jan A, Frank K. A robust vision-based hover control for ROV. In: IEEE Kobe Techno-Ocean. Kobe, 2008. 1–7
Dennis K, Homayoun N. Development of visual simultaneous localization and mapping (VSLAM) for a pipe inspection robot. In: International Symposium on Computational Intelligence in Robotics and Automation. Jacksonville, 2007. 344–349
Castle R O, Klein G, Murray D W. Combining monoSLAM with object recognition for scene augmentation using a wearable camera. Image Vision Comput, 2010, 28: 1548–1556
Liu J J, Cody P, Kostas D. Video-based localization without 3D mapping for the visually impaired. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops. San Francisco, 2010. 23–30
Jari H, Pekka S, Janne H. Vision-based motion estimation for interaction with mobile devices. Comput Vis Image Und, 2007, 108: 188–195
Arvind R, Anning C, Jay A F. Observability analysis of INS and lever-arm error states with CDGPS — camera aiding. In: IEEE Symposium on Position Location and Navigation. Indian Wells, 2010. 1197–1230
Mattia D A, Andrea L, Francesco N, et al. GIMPhI: A novel vision-based navigation approach for low cost MMS. In: IEEE Position Location and Navigation Symposium. Indian Wells, 2010. 1238–1244
Li D R, Zhao S M, Lu Y H, et al. Combined block adjustment for airborne three-line CCD scanner images. Acta Geod Cartogr Sin, 2007, 36: 245–250
Yuan X X. A novel method of systematic error compensation for a position and orientation system. Prog Nat Sci, 2008, 18: 953–963
Li D R, Yuan X X. GPS-supported bundle block adjustment-an empirical result from test field taiyuan. Acta Geod Cartogr Sin, 1995, 24: 1–7
Li D R, Yuan X X. Some investigation for GPS Supported acrotriangulation. Acta Geod Cartogr Sin, 1997, 26: 14–19
Li D R, Yuan X X, Wu Z C. GPS supported automatic aerotriangulation. J Rem Sens, 1997, 1: 306–310
Li D R, Yuan X X. Airborne mapping system with GPS-supported aerotriangulation. In: Proceedings International Workshop on Mobile Mapping Technology. Bangkok, 1999. 351–358
Yuan X X, Li D R. GPS-supported determination method for interior orientation of aerial camera. Acta Geod Cartogr Sin, 1995, 24: 103–196
Yuan X X, Li D R. GPS-Supported Aerial Triangulation in China, Spatial Information Science, Technology and its Applications-RS, GPS, GIS their Integration and Applications. Wuhan: Technical University of Surveying and Mapping Press, 1998. 671–681
Yuan X X, Li D R. Application of GPS-supported aerotriangulation in the establishment for hainan land resource foundation information system. J WTUSM, 1999, 24: 38–42
Wang W, Wang D. Land vehicle navigation using odometry/INS/vision integrated system. In: IEEE International Conference on Cybernetics and Intelligent Systems. Chengdu, 2008. 754–759
Zhang T G, Kang Y, Markus A, et al. Autonomous hovering of a vision/IMU guided quadrotor. In: International Conference on Mechatronics and Automation. Changchun, 2009. 2870–2875
Sagüés C, Guerrero J J. Visual correction for mobile robot homing. Robot Auton Syst, 2005, 50: 41–49
Wang P L, Shi S D, Hong X W. A SLAM algorithm based on monocular vision and odometer. Comput Simul, 2008, 25: 172–175
García G R, Sotelo M A, Parra I, et al. 2D visual odometry method for global positioning measurement. In: IEEE International Symposium on Intelligent Signal Processing. Xiamen, 2007. 1–6
Mouragnon E, Lhuillier M, Dhome M, et al. Generic and real time structure from motion using local bundle adjustment. Image Vision Comput, 2009, 27: 1178–1193
Kourogi M, Kurata T, Sakaue K. A panorama-based method of personal positioning and orientation and its real-time applications for wearable computers. In: Proceedings of 5th IEEE International Symposium on Wearable Computing. Zurich, 2001. 107–114
Andrew E J, Yang C, Larry H M. Machine vision for autonomous small body navigation. In: IEEE Aerospace Conference. Big Sky, 2000. 661–671
Wang X H, Fu W P, Su L. Research on binocular vision SLAM with odometer in indoor environment. J Xi’an Univ Technol, 2009, 25: 466–471
Comport A I, Malis E, Rives P. Accurate quadrifocal tracking for robust 3D visual odometry. In: IEEE International Conference on Robotics and Automation. Roma, 2007. 40–45
Bernd K, Andreas G, Henning L. Visual odometry based on stereo image sequences with RANSAC-based outlier rejection scheme. In: IEEE Symposium on Intelligent Vehicles. San Diego, 2010. 487–492
Cornelia F, Yiannis A. Observability of 3D motion. Int J Comput Vision, 2000, 37: 43–63
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Li, D., Liu, Y. & Yuan, X. Image-based self-position and orientation method for moving platform. Sci. China Inf. Sci. 56, 1–14 (2013). https://doi.org/10.1007/s11432-012-4649-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11432-012-4649-9