Skip to main content
Log in

Global image signature for visual loop-closure detection

  • Published:
Autonomous Robots Aims and scope Submit manuscript

Abstract

This work details a new method for loop-closure detection based on using multiple orthogonal projections to generate a global signature for each image of a video sequence. The new multi-projection function permits the detection of images corresponding to the same scene, but taken from different points of view. The signature generation process preserves enough information for robust loop-closure detection, although it transforms each image to a simple and compact representation. Thanks to these characteristics, a real-time operation is possible, even for long sequences with thousands of images. In addition, it has proved to work on very different scenarios without the need to change the parameters or to perform an onffline training stage, which makes it very independent on the environment and camera configuration. Results of an extensive set of experiments of the algorithm on several datasets, both indoors and outdoors and including underwater scenarios, are presented. Furthermore, an implementation, named HALOC, is available at a public repository as a C++ library for its use under the BSD license.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

References

  • Angeli, A., Filliat, D., Doncieux, S., & Meyer, J. A. (2008). Fast and incremental method for loop-closure detection using bags of visual words. IEEE Transactions on Robotics, 24(5), 1027–1037.

    Article  Google Scholar 

  • Arandjelovic, R., & Zisserman, A. (2013). All about vlad. In 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 1578–1585). IEEE, Piscataway.

  • Bay, H., Tuytelaars, T., & Van Gool, L. (2006). Surf: Speeded up robust features. In Proceedings of the European Conference on Computer Vision (ECCV) (pp. 404–417). Springer, Berlin.

  • Calonder, M., Lepetit, V., Strecha, C., & Fua, P. (2010). Brief: Binary robust independent elementary features. In Proceedings of IEEE European Conference on Computer Vision (ECCV).

  • Clemente, L. A., Davison, A. J., Reid, I. D., Neira, J., & Tardós, J. D. (2007). Mapping large loops with a single hand-held camera. In Proceedings of Robotics: Science and Systems.

  • Cummins, M., & Newman, P. (2008). Accelerated appearance-only slam. In IEEE international conference on Robotics and automation, 2008 (ICRA 2008) (pp. 1828–1833). IEEE, Piscataway.

  • Cummins, M., & Newman, P. (2008). Fab-map: Probabilistic localization and mapping in the space of appearance. The International Journal of Robotics Research, 27(6), 647–665.

    Article  Google Scholar 

  • Cummins, M. & Newman, P. (2008) The oxford mobile robotics group datasets, 2008. http://www.robots.ox.ac.uk/~mobile/IJRR_2008_Dataset/data.html.

  • Cummins, M., & Newman, P. (2010) Fab-map: Appearance-based place recognition and mapping using a learned visual vocabulary model. In Proceedings of the International Conference on Machine Learning

  • Diephuis, M., Voloshynovskiy, S., Koval, O., & Beekhof, F. (2011) Statistical analysis of binarized sift descriptors. In 7th International Symposium on Image and Signal Processing and Analysis (ISPA) (pp. 460–465). IEEE, Piscataway.

  • Durrant-Whyte, H., & Bailey, T. (2006). Simultaneous localization and mapping: Part i. IEEE Robotics Automation Magazine, 13(2), 99–110.

    Article  Google Scholar 

  • Filliat, D. (2007) A visual bag of words method for interactive qualitative localization and mapping. In IEEE International Conference on Robotics and Automation (pp. 3921–3926). IEEE, Piscataway.

  • Garcia-Fidalgo, E., & Ortiz, A. (2014). Vision-based topological mapping and localization by means of local invariant features and map refinement. Robotica, 33, 1446–1470.

    Article  Google Scholar 

  • Garcia-Fidalgo, E., & Ortiz, A. (2015). Vision-based topological mapping and localization methods: A survey. Robotics and Autonomous Systems, 64, 1–20.

    Article  Google Scholar 

  • Geiger, A., Ziegler, J., & Stiller, C. (2011). Stereoscan: Dense 3d reconstruction in real-time. In IEEE Intelligent Vehicles Symposium, Baden-Baden, June 2011.

  • Gionis, A., Indyk, P., Motwani, R., et al. (1999). Similarity search in high dimensions via hashing. VLDB, 99, 518–529.

    Google Scholar 

  • Glover, A., Maddern, W., Warren, M., Reid, S., Milford, M., & Wyeth, G.: (2011). Openfabmap: An open source toolbox for appearance-based loop closure detection. In The International Conference on Robotics and Automation, St Paul, Minnesota, 2011. IEEE, Piscataway.

  • Hartley, R., & Zisserman, A. (2003) Multiple View Geometry in Computer Vision. Cambridge University Press, Cambridge. ISBN: 0521623049.

  • Jégou, H., Douze, M., Schmid, C., & Pérez, P. (2010) Aggregating local descriptors into a compact image representation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2010 (pp. 3304–3311). IEEE, Piscataway.

  • Konečnỳ, J., & Hagara, M. (2014). One-shot-learning gesture recognition using hog-hof features. The Journal of Machine Learning Research, 15(1), 2513–2532.

    MathSciNet  Google Scholar 

  • Košecká, J., Li, F., & Yang, X. (2005). Global localization and relative positioning based on scale-invariant keypoints. Robotics and Autonomous Systems, 52(1), 27–38.

    Article  Google Scholar 

  • Lin, S., Ozsu, M. T., Oria, V., & Ng, R. (2001). An extendible hash for multi-precision similarity querying of image databases. In Proceedings of International Conference of Very Large Data Bases (VLDB), 1, 221–230.

    Google Scholar 

  • Liu, M., & Siegwart, R. (2012). Dp-fact: Towards topological mapping and scene recognition with color for omnidirectional cameras. In Proceedings of IEEE International Conference on Robotics and Automation (ICRA) (pp. 3503–3508), May 2012.

  • Liu, M., & Siegwart, R. (2014). Topological mapping and scene recognition with lightweight color descriptors for an omnidirectional camera. IEEE Transaction on Robotics, 30(2), 310–324.

    Article  Google Scholar 

  • Liu, Y., & Zhang, H. (2012). Visual loop closure detection with a compact image descriptor. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 1051–1056).

  • Lowe, D. G. (2004). Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60(2), 91–110.

    Article  Google Scholar 

  • Mıhçak, M. K., & Venkatesan, R. (2002) New iterative geometric methods for robust perceptual image hashing. In Security and privacy in digital rights management (pp. 13–21). Springer, Berlin.

  • Milford, M., & Wyeth, G. (2012). Visual route-based navigation for sunny summer days and stormy winter nights. In The International Conference on Robotics and Automation

  • Monga, V., & Evans, B. L. (2006). Perceptual image hashing via feature points: Performance evaluation and tradeoffs. IEEE Transactions on Image Processing, 15(11), 3453–3466.

    Article  Google Scholar 

  • Negre, P. L., & Bonin-Font, F. (2014) libhaloc. Retrieved May 2014 from, https://github.com/srv/libhaloc.

  • Oliva, A., & Torralba, A. (2001). Modelling the shape of the scene: A holistic representation of the spatial envelope. International Journal on Computer Vision, 42(3), 145–175.

    Article  MATH  Google Scholar 

  • Roy, S., Zhu, X., Yuan, J., & Chang, E. C. (2007). On preserving robustness-false alarm tradeoff in media hashing. In SPIE Proceedings on Visual Communications and Image Processing, vol. 6508.

  • Schaefer, G., & Stich, M. (2004). Ucid—an uncompressed colour image database. In Proceedings of SPIE, Storage and Retrieval Methods and Applications for Multimedia (pp. 472–480), San Jose.

  • Shahbazi, H., & Zhang, H. (2011). Application of locality sensitive hashing to realtime loop closure detection. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems(IROS) (pp. 1228–1233), San Francisco.

  • Smith, J. R., & Chang, S. (1997). Visually searching the web for content. IEEE Multimedia, 4(3), 12–20.

    Article  Google Scholar 

  • Sünderhauf, N., & Protzel, P. (2011). Brief-gist - closing the loop by simple means. In Proceedings of IEEE International Conference on Intelligent Robots and Systems (IROS).

  • Sunderhauf, N., & Protzel, P. (2011). Brief-gist-closing the loop by simple means. In IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 1234–1241). IEEE, Piscataway.

  • Swaminathan, A., Mao, Y., & Wu, M. (2006). Robust and secure image hashing. IEEE Transactions on Information Forensics and Security, 1(2), 215–230.

    Article  Google Scholar 

  • Ulrich, I., & Nourbakhsh, I. (2000). Appearance-based place recognition for topological localization. In Proceedings of IEEE International Conference on Robotics and Automation, 2000 (ICRA’00) (vol. 2, pp. 1023–1029). IEEE, Piscataway.

  • Vedaldi, A., & Fulkerson, B. (2008). VLFeat: An open and portable library of computer vision algorithms. http://www.vlfeat.org/.

  • Venkatesan, R., Koon, S. M., Jakubowski, M. H., & Moulin, P. (2000) Robust image hashing. In Proceedings of International Conference on Image Processing, 2000 (vol. 3, pp. 664–666). IEEE, Piscataway.

  • Wan, J., Ruan, Q., Li, W., An, G., & Zhao, R. (2014). 3d smosift: Three-dimensional sparse motion scale invariant feature transform for activity recognition from rgb-d videos. Journal of Electronic Imaging, 23(2), 023017–023017.

    Article  Google Scholar 

  • Wan, J., Ruan, Q., Li, W., & Deng, S. (2013). One-shot learning gesture recognition from RGB-D data using bag of features. The Journal of Machine Learning Research, 14(1), 2549–2582.

    Google Scholar 

  • Williams, B., Cummins, M., Neira, J., Newman, P., Reid, I., & Tardós, J. (2008). An image-to-map loop closing method for monocular slam. In IEEE/RSJ International Conference on Intelligent Robots and Systems, 2008 (IROS 2008) (pp. 2053–2059). IEEE, Piscataway.

  • Williams, B., Cummins, M., Neira, J., Newman, P., Reid, I., & Tardós, J. (2009). A comparison of loop closing techniques in monocular slam. Robotics and Autonomous Systems, 57(12), 1188–1197.

    Article  Google Scholar 

  • Wirth, S., Negre Carrasco, P. L., & Oliver, G. (2013). Visual odometry for autonomous underwater vehicles. In Proceedings of the IEEE Oceans, Bergen.

Download references

Acknowledgments

This work is partially supported by Ministry of Economy and Competitiveness under contracts TIN2014-58662-R, DPI2014-57746-C3-2-R, Govern de les Illes Balears (AAEE60/2014) and FEDER funds.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Pep Lluis Negre Carrasco.

Additional information

See https://github.com/srv/libhaloc.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Negre Carrasco, P.L., Bonin-Font, F. & Oliver-Codina, G. Global image signature for visual loop-closure detection. Auton Robot 40, 1403–1417 (2016). https://doi.org/10.1007/s10514-015-9522-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10514-015-9522-4

Keywords

Navigation