ABSTRACT
In this paper we present a novel visual-inertial 6-DOF localization approach that can be directly integrated in a wearable immersive system for simulation and training. In this context, while CAVE environments typically require complex and expensive set-up, our approach relies on visual and inertial information provided by commodity hardware, i.e. a consumer monocular camera and an Inertial Measurement Unit (IMU).
We propose a novel robust pipeline based on state-of-the-art image-based localization and sensor fusion approaches. A loosely-coupled sensor fusion approach, which makes use of robust orientation information from the IMU, is employed to cope with failures in visual tracking (e.g. due to camera fast motion) in order to limit motion jitters. Fast and smooth re-localization is also provided to track position following visual tracking outage and guarantee continued operation. The 6-DOF information is then used to render consistently VR contents on a stereoscopic HMD. The proposed system, demonstrated in the context of Construction, runs at 30 fps on a standard PC and requires a very limited set-up for its intended application.
Supplemental Material
Available for Download
Supplemental material.
- Aron, M., Simon, G., and Berger, M.-O. 2007. Use of inertial sensors to support video tracking. Comput. Animat. Virtual Worlds 18, 1 (Feb.), 57--68. Google ScholarDigital Library
- Bleser, G., and Stricker, D. 2008. Advanced tracking through efficient image processing and visual-inertial sensor fusion. In IEEE VR '08, 137--144.Google Scholar
- Chen, W., Plancoulaine, A., Férey, N., Touraine, D., Nelson, J., and Bourdot, P. 2013. 6DoF Navigation in Virtual Worlds: Comparison of joystick-based and head-controlled paradigms. In ACM VRST '13, 111--114. Google ScholarDigital Library
- Cruz-Neira, C., Sandin, D. J., DeFanti, T. A., Kenyon, R. V., and C. Hart, J. 1992. The CAVE: Audio visual experience automatic virtual environment. Commun. ACM 35, 6 (June), 64--72. Google ScholarDigital Library
- Gauglitz, S., Höllerer, T., and Turk, M. 2011. Evaluation of interest point detectors and feature descriptors for visual tracking. Int. J. Comput. Vis. 94, 3, 335--360. Google ScholarDigital Library
- Haralick, B., Lee, C.-N., Ottenberg, K., and Nolle, M. 1994. Review and analysis of solutions of the three point perspective pose estimation problem. Int. J. Comput. Vis. 13, 3, 331--356. Google ScholarDigital Library
- Hay, S., Newman, J., and Harle, R. 2008. Optical tracking using commodity hardware. In ISMAR 2008, 159--160. Google ScholarDigital Library
- Heinly, J., Dunn, E., and Frahm, J.-M. 2012. Comparative evaluation of binary features. In ECCV 2012. 759--773. Google ScholarDigital Library
- Lim, H., Sinha, S. N., Cohen, M. F., and Uyttendaele, M. 2012. Real-time image-based 6-DOF localization in large-scale environments. In IEEE CVPR '12, 1043--1050. Google ScholarDigital Library
- Oskiper, T., Chiu, H.-P., Zhu, Z., Samaresekera, S., and Kumar, R. 2011. Stable vision-aided navigation for large-area augmented reality. In IEEE VR 2011, 63--70. Google ScholarDigital Library
- Perea, L., How, J., Breger, L., and Elosegui, P. 2007. Nonlinearity in sensor fusion: Divergence issues in EKF, modified truncated SOF, and UKF. In Proc. AIAA Guidance, Navigation, and Control Conf. 2007.Google ScholarCross Ref
- Shi, J., and Tomasi, C. 1994. Good features to track. In Proceedings CVPR '94, 593--600.Google Scholar
- Snavely, N., Seitz, S. M., and Szeliski, R. 2008. Modeling the world from internet photo collections. Int. J. Comput. Vision 80, 2 (Nov.), 189--210. Google ScholarDigital Library
- Welch, G., and Foxlin, E. 2002. Motion tracking: No silver bullet, but a respectable arsenal. IEEE Comput. Graph. Appl. 22, 6 (Nov.), 24--38. Google ScholarDigital Library
- Williams, B., Narasimham, G., Rump, B., McNamara, T. P., Carr, T. H., Rieser, J., and Bodenheimer, B. 2007. Exploring large virtual environments with an HMD when physical space is limited. In Proc. 4th APGV Symposium, 41--48. Google ScholarDigital Library
- Zhu, Z., Oskiper, T., Samarasekera, S., Kumar, R., and Sawhney, H. 2008. Real-time global localization with a pre-built visual landmark database. In IEEE CVPR 2008, 1--8.Google Scholar
Index Terms
- Robust 6-DOF immersive navigation using commodity hardware
Recommendations
A Multi-Camera 6-DOF Pose Tracker
ISMAR '04: Proceedings of the 3rd IEEE/ACM International Symposium on Mixed and Augmented RealityMost of the work in head-pose tracking has concentrated on single-camera systems with a relatively small field of view which have limited accuracy because features are only observed in a single viewing direction. We present a multi-camera pose tracker ...
A 6-DOF ARTag-based tracking system
This paper presents a system to track position and orientation of a generic mobile device equipped with a camera using a set of variable size fiducial markers. The system provides six degrees-of-freedom (6-DOF) by tracking fiducial markers through the ...
Stereo by Multiperspective Imaging under 6 DOF Camera Motion
3DIM '05: Proceedings of the Fifth International Conference on 3-D Digital Imaging and ModelingMultiperspective imaging has been used to recover the structure of a scene. Although several algorithms for structure recovery have been developed as typified by stereo panoramas, there exists no common framework which subsumes various camera motions to ...
Comments