Skip to main content
Log in

Full-parameter vision navigation based on scene matching for aircrafts

  • Research Paper
  • Published:
Science China Information Sciences Aims and scope Submit manuscript

Abstract

Vision navigation based on scene matching between real-time images and a reference image has many advantages over the commonly used inertial navigation system (INS), such as no cumulative measurement errors for long-endurance flight. Recent developments in vision navigation are mainly used for partial navigation parameters measurements, such as the position and the relative velocity, which cannot meet the requirements of completely autonomous navigation. We present the concept, principle and method of full-parameter vision navigation (FPVN) based on scene matching. The proposed method can obtain the three-dimensional (3D) position and attitude angles of an aircraft by the scene matching for multiple feature points instead of a single point in existing vision navigations. Thus, FPVN can achieve the geodetic position coordinates and attitude angles of the aircraft and then the velocity vector, attitude angular velocity and acceleration can be derived by the time differentials, which provide a full set of navigation parameters for aircrafts and unmanned aerial vehicles (UAVs). The method can be combined with the INS and the cumulative errors of the INS can be corrected using the measurements of FPVN rather than satellite navigation systems. The approach provides a completely autonomous and accurate navigation method for long-endurance flight without the help of satellites.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Bonin-Font F, Ortiz A, Oliver G. Visual navigation for mobile robots: a survey. J Intell Robot Syst, 2008, 53: 263–296

    Article  Google Scholar 

  2. Lin F, Lum K-Y, Chen B M, et al. Development of a vision-based ground target detection and tracking system for a small unmanned helicopter. Sci China Ser F-Inf Sci, 2009, 52: 2201–2215

    Article  MATH  Google Scholar 

  3. Li L C, Yu Q F, Shang Y, et al. A new navigation approach of terrain contour matching based on 3-D terrain reconstruction from onboard image sequence. Sci China Technol Sci, 2010, 53: 1176–1183

    Article  Google Scholar 

  4. Chaumette F, Hutchinson S. Visual servo control. Part II: advanced approaches. IEEE Robot Autom Mag, 2007, 14: 109–118

    Article  Google Scholar 

  5. Chesi G, Vicino A. Visual servoing for large camera displacements. IEEE Trans Robot, 2004, 20: 724–735

    Article  Google Scholar 

  6. Cesetti A, Frontoni E, Mancini A, et al. A visual global positioning system for unmanned aerial vehicles used in photogrammetric applications. J Intell Robot Syst, 2011, 61: 157–168

    Article  Google Scholar 

  7. Mourikis A I, Trawny N, Roumeliotis S I, et al. Vision-aided inertial navigation for spacecraft entry, descent, and landing. IEEE Trans Robot, 2009, 25: 264–280

    Article  Google Scholar 

  8. Cheng Y, Ansar A. Landmark based position estimation for pinpoint landing on mars. In: IEEE International Conference on Robotics and Automation Barcelona, Spain, 2005. 4470–4475

    Google Scholar 

  9. Bras S, Cunha R, Vasconcelos J F, et al. A nonlinear attitude observer based on active vision and inertial measurements. IEEE Trans Robot, 2011, 27: 664–677

    Article  Google Scholar 

  10. Li S, Cui P Y, Cui H T. Vision-aided inertial navigation for pinpoint planetary landing. Aerospace Sci Technol, 2007, 11: 499–506

    Article  MathSciNet  Google Scholar 

  11. Mourikis A I, Trawny N. Vision-aided inertial navigation for spacecraft entry, descent, and landing. IEEE Trans Robot, 2009, 25: 264–280

    Article  Google Scholar 

  12. Lupton O, Sukkarieh S. Visual-inertial-aided navigation for high-dynamic motion in built environments without initial conditions. IEEE Trans Robot, 2012, 28: 61–76

    Article  Google Scholar 

  13. Sim D G, Park R H, Kim R C, et al. Integrated position estimation using aerial image sequences. IEEE Trans Patt Anal Mach Intell, 2002, 24: 1–18

    Article  Google Scholar 

  14. Conte G. Vision-based localization and guidance for unmanned aerial vehicles. Dissertation for the Doctoral Degree, Linkoping University, 2009

    Google Scholar 

  15. Caballero F, Merino L, Ferruz J, et al. A visual odometer without 3D reconstruction for aerial vehicles, applications to building inspection. In: Proceedings of the 2005 IEEE International Conference on Robotics and Automation, Barcelona, 2005. 4673–4678

    Chapter  Google Scholar 

  16. Kaiser K, Gans N, Dixon W. Localization and control of an aerial vehicle through chained, vision-based pose reconstruction. In: Proceedings of the 2007 American Control Conference, New York, 2007. 5934–5939

    Chapter  Google Scholar 

  17. Amidi O. An autonomous vision-guided helicopter. Dissertation for the Doctoral Degree, Carnegie Mellon University, 1996

    Google Scholar 

  18. Kendoul F, Nonami K, Fantonik I, et al. An adaptive vision-based autopilot for mini flying machines guidance, navigation and control. Auton Robot, 2009, 27: 165–188

    Article  Google Scholar 

  19. Yu Q, Shang Y. Videometrics: Principles and Researches. Beijing: Science Press, 2009

    Google Scholar 

  20. Xu D, Li Y, Tan M. A general recursive linear method and unique solution pattern design for the perspective-n-point problem. Image Vis Comput, 2008, 26: 740–750

    Article  Google Scholar 

  21. Fishler M, Bolles R. Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun ACM, 1981, 24: 381–395

    Article  Google Scholar 

  22. Samadzadegan F, Hahn M, Saeedi S. Position estimation of aerial vehicle based on a vision aided navigation system. http://www.geovisualisierung.net/isprs2007/docs/Samadzadegan.pdf, 2007

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to QiFeng Yu.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yu, Q., Shang, Y., Liu, X. et al. Full-parameter vision navigation based on scene matching for aircrafts. Sci. China Inf. Sci. 57, 1–10 (2014). https://doi.org/10.1007/s11432-014-5094-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11432-014-5094-8

Keywords

Navigation