Skip to main content

Automated Marking of Underwater Animals Using a Cascade of Neural Networks

  • Conference paper
  • First Online:
Computational Science and Its Applications – ICCSA 2021 (ICCSA 2021)

Abstract

In this work, a multifactorial problem of analyzing the seabed state of plants and animals using photo and video materials is considered. Marine research to monitor benthic communities and automatic mapping of underwater landscapes make it possible to qualitatively assess the state of biomes. The task includes several components: preparation of a methodology for data analysis, their aggregation, analysis, presentation of results. In this work, we focused on methods for automating detection and data presentation.

For deep-sea research, which involves the detection, counting and segmentation of plants and animals, it is difficult to use traditional computer vision techniques. Thanks to modern automated monitoring technologies, the speed and quality of research can be increased several times while reducing the required human resources using machine learning and interactive visualization methods.

The proposed approach significantly improves the quality of the segmentation of objects underwater. The algorithm includes three main stages: correction of image distortions underwater, image segmentation, selection of individual objects. Combining neural networks that successfully solve each of the tasks separately into a cascade of neural networks is the optimal method for solving the problem of segmentation of aquaculture and animals.

Using the results obtained, it is possible to facilitate the control of the ecological state in the world, to automate the task of monitoring underwater populations.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Fish Recognition Ground-Truth data. http://groups.inf.ed.ac.uk/f4k/groundtruth/recog. Accessed 20 Mar 2021

  2. Fish Species Recognition. http://www.perceivelab.com/datasets. Accessed 20 Mar 2021

  3. Ozfish. https://aims.github.io/ozfish. Accessed 21 Mar 2021

  4. Fish Dataset. https://wiki.qut.edu.au/display/raq/Fish+Dataset. Accessed 22 Mar 2021

  5. Kepler.gl. https://github.com/keplergl/kepler.gl. Accessed 22 Mar 2021

  6. LP DAAC - SRTMGL1. https://lpdaac.usgs.gov/products/srtmgl1v003. Accessed 22 May 2021

  7. Azad, R., Fayjie, A.R., Kauffmann, C., Ben Ayed, I., Pedersoli, M., Dolz, J.: On the texture bias for few-shot CNN segmentation. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 2674–2683 (2021)

    Google Scholar 

  8. Bakiev, M., Khasanov, K.: Comparison of digital elevation models for determining the area and volume of the water reservoir. Int. J. Geoinform. 17(1), 37–45 (2021)

    Article  Google Scholar 

  9. Benjamin, J., et al.: Aboriginal artefacts on the continental shelf reveal ancient drowned cultural landscapes in northwest Australia. PLoS ONE 15(7), e0233912 (2020)

    Article  Google Scholar 

  10. Cao, S., Zhao, D., Sun, Y., Liu, X., Ruan, C.: Automatic coarse-to-fine joint detection and segmentation of underwater non-structural live crabs for precise feeding. Comput. Electron. Agric. 180, 105905 (2021)

    Article  Google Scholar 

  11. Dong, N., Xing, E.P.: Few-shot semantic segmentation with prototype learning. In: BMVC, vol. 3 (2018)

    Google Scholar 

  12. Ghorbani, M.A., Deo, R.C., Kim, S., Hasanpour Kashani, M., Karimi, V., Izadkhah, M.: Development and evaluation of the cascade correlation neural network and the random forest models for river stage and river flow prediction in Australia. Soft Comput. 24(16), 12079–12090 (2020). https://doi.org/10.1007/s00500-019-04648-2

    Article  Google Scholar 

  13. Islam, M.J., et al.: Semantic segmentation of underwater imagery: dataset and benchmark. arXiv preprint arXiv:2004.01241 (2020)

  14. Islam, M.J., Enan, S.S., Luo, P., Sattar, J.: Underwater image super-resolution using deep residual multipliers. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 900–906. IEEE (2020)

    Google Scholar 

  15. Jian, M., Liu, X., Luo, H., Lu, X., Yu, H., Dong, J.: Underwater image processing and analysis: a review. Sig. Process. Image Commun., 116088 (2020)

    Google Scholar 

  16. Jung, A.B., et al.: imgaug (2020). https://github.com/aleju/imgaug. Accessed 1 Feb 2020

  17. Li, C., Anwar, S., Porikli, F.: Underwater scene prior inspired deep underwater image and video enhancement. Pattern Recogn. 98, 107038 (2020)

    Article  Google Scholar 

  18. Li, X., Wei, T., Chen, Y.P., Tai, Y.W., Tang, C.K.: FSS-1000: a 1000-class dataset for few-shot segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2869–2878 (2020)

    Google Scholar 

  19. Liu, S., Yu, J., Ke, Z., Dai, F., Chen, Y.: Aerial-ground collaborative 3D reconstruction for fast pile volume estimation with unexplored surroundings. Int. J. Adv. Robot. Syst. 17(2), 1729881420919948 (2020)

    Google Scholar 

  20. Miao, J., Wei, Y., Yang, Y.: Memory aggregation networks for efficient interactive video object segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10366–10375 (2020)

    Google Scholar 

  21. Nocerino, E., Menna, F., Chemisky, B., Drap, P.: 3D sequential image mosaicing for underwater navigation and mapping. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 43, 991–998 (2020)

    Article  Google Scholar 

  22. Roach, T.N., et al.: A field primer for monitoring benthic ecosystems using structure-from-motion photogrammetry. JoVE (J. Vis. Exp.) 170, e61815 (2021)

    Google Scholar 

  23. Shu, C., Yu, K., Duan, Z., Yang, K.: Feature-metric loss for self-supervised learning of depth and egomotion. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12364, pp. 572–588. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58529-7_34

    Chapter  Google Scholar 

  24. Urbina-Barreto, I., et al.: Quantifying the shelter capacity of coral reefs using photogrammetric 3D modeling: from colonies to reefscapes. Ecol. Ind. 121, 107151 (2021)

    Article  Google Scholar 

  25. Yang, Z., Wei, Y., Yang, Y.: Collaborative video object segmentation by foreground-background integration. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12350, pp. 332–348. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58558-7_20

    Chapter  Google Scholar 

Download references

Acknowledgments

The authors would like to acknowledge the Reviewers for the valuable recommendations that helped in the improvement of this paper.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Oleg Iakushkin .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Iakushkin, O. et al. (2021). Automated Marking of Underwater Animals Using a Cascade of Neural Networks. In: Gervasi, O., et al. Computational Science and Its Applications – ICCSA 2021. ICCSA 2021. Lecture Notes in Computer Science(), vol 12956. Springer, Cham. https://doi.org/10.1007/978-3-030-87010-2_34

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-87010-2_34

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-87009-6

  • Online ISBN: 978-3-030-87010-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics