Skip to main content

Limit Characterization for Visual Place Recognition in Underwater Scenes

  • Conference paper
  • First Online:
ROBOT2022: Fifth Iberian Robotics Conference (ROBOT 2022)

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 589))

Included in the following conference series:

Abstract

The underwater environment has some structures that still need regular inspection. However, the nature of this environment presents a number of challenges in achieving accurate vehicle position and consequently successful image similarity detection. Although there are some factors - water turbidity or light attenuation - that degrade the quality of the captured images, visual sensors have shown a strong impact on mission scenarios - close range operations. Therefore, the purpose of this paper is to study whether these data are capable of addressing the aforementioned underwater challenges on their own. Considering the lack of available data in this context, a typical underwater scenario was recreated using the Stonefish simulator. Experiments were conducted on two predefined trajectories containing appearance scene changes. The loop closure situations provided by the bag-of-words (BoW) approach are correctly detected, but it is sensitive to some severe conditions.

FCT - Fundação para a Ciência e a Tecnologia.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://github.com/dorian3d/DBoW2.

  2. 2.

    https://github.com/dorian3d/DLoopDetector.

References

  1. Lu, H., Li, Y., Zhang, Y., Chen, M., Serikawa, S., Kim, H.: Underwater optical image processing: a comprehensive review. Mob. Netw. Appl. 22(6), 1204–1211 (2017). https://doi.org/10.1007/s11036-017-0863-4

    Article  Google Scholar 

  2. Melo, J., Matos, A.: Survey on advances on terrain based navigation for autonomous underwater vehicles. Ocean Eng. 139, 250–264 (2017). http://dx.doi.org/10.1016/j.oceaneng.2017.04.047

  3. Tareen, S.A.K., Saleem, Z.: A comparative analysis of SIFT, SURF, KAZE, AKAZE, ORB, and BRISK. In: International Conference on Computing, Mathematics and Engineering Technologies, pp. 1–10 (2014)

    Google Scholar 

  4. Fuentes-Pacheco, J., Ruiz-Ascencio, J., Rendón-Mancha, J.M.: Visual simultaneous localization and mapping: a survey. Artif. Intell. Rev. 43, 55–81 (2015)

    Article  Google Scholar 

  5. Kejriwal, N., Kumar, S., Shibata, T.: High performance loop closure detection using bag of word pairs. Robot. Auton. Syst. 77, 55–65 (2016)

    Article  Google Scholar 

  6. Gaspar, A.R., Nunes, A., Matos, A.: Evaluation of bags of binary words for place recognition in challenging scenarios. In: 2021 IEEE International Conference on Autonomous Robot Systems and Competitions, ICARSC 2021, pp. 19–24 (2021)

    Google Scholar 

  7. Hidalgo, F., Bräunl, T.: Evaluation of several feature detectors/extractors on underwater images towards vSLAM. Sensors. 08, 20 (2020)

    Google Scholar 

  8. Gálvez-López, D., Tardós, J.D.: Real-time loop detection with bags of binary words. In: International Conference on Intelligent Robots and Systems, pp. 51–58 (2011)

    Google Scholar 

  9. Mur-Artal, R., Tardós, J.D.: Fast relocalisation and Loop closing in Keyframe-based SLAM. In: IEEE International Conference on Robotics and Automation, pp. 846–853, June 2014

    Google Scholar 

  10. Nicosevici, T., Garcia, R.: Automatic visual bag-of-words for online robot navigation and mapping. IEEE Trans. Rob. 28(4), 886–898 (2012)

    Article  Google Scholar 

  11. Negre, P.L., Bonin-Font, F., Oliver, G.: Cluster-based loop closing detection for underwater SLAM in feature-poor regions. In: IEEE International Conference on Robotics and Automation. IEEE, pp. 2589–2595 (2016)

    Google Scholar 

  12. Sharma, S., Gupta, V., Juneja, M.: A survey of image data indexing techniques. Artif. Intell. Rev. 08, 52 (2019)

    Google Scholar 

  13. Wang, H., Wang, C., Xie, L.: Online visual place recognition via saliency re-identification. In: IEEE RSJ International Conference on Intelligent Robots and Systems, pp. 5030–5036 (2020)

    Google Scholar 

  14. Cieślak, P.: Stonefish: an advanced open-source simulation tool designed for marine robotics, with a ROS interface. In: OCEANS 2019 - Marseille, pp. 1–6 (2019)

    Google Scholar 

Download references

Acknowledgements

This work is financed by FCT - Fundação para a Ciência e a Tecnologia - and by FSE - Fundo Social Europeu through of the Norte 2020 – Programa Operacional Regional do Norte - through of the doctoral scholarship SFRH/BD/146460/2019. This work is also financed by K2D Project - Knowledge and Data from the Deep to the Space (POCI-01-0247-FEDER-045941) funded within the scope of MIT Portugal.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ana Rita Gaspar .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Gaspar, A.R., Nunes, A., Matos, A. (2023). Limit Characterization for Visual Place Recognition in Underwater Scenes. In: Tardioli, D., Matellán, V., Heredia, G., Silva, M.F., Marques, L. (eds) ROBOT2022: Fifth Iberian Robotics Conference. ROBOT 2022. Lecture Notes in Networks and Systems, vol 589. Springer, Cham. https://doi.org/10.1007/978-3-031-21065-5_6

Download citation

Publish with us

Policies and ethics