skip to main content
10.1145/3352593.3352599acmotherconferencesArticle/Chapter ViewAbstractPublication PagesairConference Proceedingsconference-collections
research-article

Real-time depth estimation using camera and IMU on the unstabilized platform of a spherical robot

Published:27 January 2020Publication History

ABSTRACT

Vision based robot navigation relies on the image sequences that are captured by the camera attached to its platform. In many robotic applications such as in case of spherical robots used for surveillance, the platform on which the camera resides is often unsteady and unwanted relative motion exists between camera and scene. This unwanted relative motion in case of a spherical robot is due to the pitching motion associated with its platform called yoke. A depth estimation algorithm that handles the effect of pitching using non-linear observer approach is proposed. The object in the scene whose depth is to be estimated is detected as features in the acquired images and are tracked by using concepts of optical flow. A discrete time state space model that fuses the information from camera and IMU data attached to the unsteady platform of spherical robot is derived. Extended Kalman filter (EKF) is used as the non-linear estimation technique for extraction of depth information from the proposed state space model. The convergence aspects of the extended Kalman filter when used as a deterministic observer for the proposed non-linear discrete-time model is analyzed with local observability and it is shown that there is boundedness of error covariance between the observed and actual depth. It is shown that the estimation error converges to zero irrespective of the initialization provided to the observer.

References

  1. Jakob Engel, Thomas Schöps, and Daniel Cremers. LSD-SLAM: Large-scale direct monocular SLAM. In European Conference on Computer Vision, pages 834--849. Springer, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  2. Zhenfei Yang and Shaojie Shen. Monocular visual--inertial state estimation with online initialization and camera--IMU extrinsic calibration. IEEE Transactions on Automation Science and Engineering, 14(1):39--51, 2017.Google ScholarGoogle ScholarCross RefCross Ref
  3. Jan Smisek, Michal Jancosek, and Tomas Pajdla. 3d with kinect. In Consumer depth cameras for computer vision, pages 3--25. Springer, 2013.Google ScholarGoogle ScholarCross RefCross Ref
  4. Vikranth Reddy D. Dhiraj Gandhi Leena Vachhani, Shantanu Thakar and Akshay Khatri. Design and development of a gearless spherical robot, Dec 2015. IPA No. 4717/MUM/2015.Google ScholarGoogle Scholar
  5. Ruo Zhang, Ping-Sing Tsai, James Edwin Cryer, and Mubarak Shah. Shape-from-shading: a survey. IEEE transactions on pattern analysis and machine intelligence, 21(8):690--706, 1999.Google ScholarGoogle Scholar
  6. B. J. Super and A. C. Bovik. Shape from texture using local spectral moments. IEEE Transactions on Pattern Analysis and Machine Intelligence, 17(4):333--343, April 1995.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Ilan Shimshoni, Yael Moses, and Michael Lindenbaum. Shape reconstruction of 3d bilaterally symmetric surfaces. International Journal of Computer Vision, 39(2):97--110, 2000.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Kevin Karsch, Ce Liu, and Sing Bing Kang. Depth transfer: Depth extraction from video using non-parametric sampling. IEEE transactions on pattern analysis and machine intelligence, 36(11):2144--2158, 2014.Google ScholarGoogle Scholar
  9. Lubor Ladicky, Jianbo Shi, and Marc Pollefeys. Pulling things out of perspective. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 89--96, 2014.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Fayao Liu, Chunhua Shen, Guosheng Lin, and Ian D Reid. Learning depth from single monocular images using deep convolutional neural fields. IEEE Trans. Pattern Anal. Mach. Intell., 38(10):2024--2039, 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Dan Xu, Elisa Ricci, Wanli Ouyang, Xiaogang Wang, and Nicu Sebe. Monocular depth estimation using multi-scale continuous CRFs as sequential deep networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018.Google ScholarGoogle Scholar
  12. Clément Godard, Oisin Mac Aodha, and Gabriel J Brostow. Unsupervised monocular depth estimation with left-right consistency. In CVPR, volume 2, page 7, 2017.Google ScholarGoogle ScholarCross RefCross Ref
  13. Tong Qin, Peiliang Li, and Shaojie Shen. VINS-mono: A robust and versatile monocular visual-inertial state estimator. IEEE Transactions on Robotics, 34(4):1004--1020, 2018.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. David A Forsyth and Jean Ponce. A modern approach. Computer vision: a modern approach, 2003.Google ScholarGoogle Scholar
  15. Yongkyu Song and Jessy W Grizzle. The extended Kalman filter as a local asymptotic observer for non-linear discrete-time systems. In American Control Conference, 1992, pages 3365--3369. IEEE, 1992.Google ScholarGoogle ScholarCross RefCross Ref
  16. Chris Harris and Mike Stephens. A combined corner and edge detector. In Alvey vision conference, volume 15, pages 10--5244. Citeseer, 1988.Google ScholarGoogle ScholarCross RefCross Ref
  17. Bruce D. Lucas and Takeo Kanade. An iterative image registration technique with an application to stereo vision. In Proceedings of the 7th International Joint Conference on Artificial Intelligence - Volume 2, IJCAI'81, pages 674--679, San Francisco, CA, USA, 1981. Morgan Kaufmann Publishers Inc.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. J Deyst and C Price. Conditions for asymptotic stability of the discrete minimum-variance linear estimator. IEEE Transactions on Automatic Control, 13(6):702--705, 1968.Google ScholarGoogle ScholarCross RefCross Ref
  19. Lei Chen, Junhong Li, and Ruifeng Ding. Identification for the second-order systems based on the step response. Mathematical and computer modelling, 53(5-6):1074--1083, 2011.Google ScholarGoogle Scholar
  20. Jean-Yves Bouguet. Pyramidal implementation of the affine lucas kanade feature tracker description of the algorithm. Intel Corporation, 5(1-10):4, 2001.Google ScholarGoogle Scholar

Index Terms

  1. Real-time depth estimation using camera and IMU on the unstabilized platform of a spherical robot

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Other conferences
        AIR '19: Proceedings of the 2019 4th International Conference on Advances in Robotics
        July 2019
        423 pages
        ISBN:9781450366502
        DOI:10.1145/3352593

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 27 January 2020

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed limited

        Acceptance Rates

        AIR '19 Paper Acceptance Rate69of140submissions,49%Overall Acceptance Rate69of140submissions,49%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader