Skip to main content

Advertisement

Log in

Dolphin movement direction recognition using contour-skeleton information

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Detecting and tracking social marine mammals, including dolphins, can help to explain their social dynamics, predict their behavior, and measure the impact of human interference. The underwater environment is very special and different from the land environment: acoustic recorders are the main equipment for researchers to track dolphins at a long distance. Close-range detection of visual data is often seriously affected by the underwater environment, because the underwater environment is highly dynamic and the light source has attenuation in the deep water. Nonetheless, compared with acoustic information, visual data can provide more detailed information at low cost. The videos and images of dolphins provide researchers with great convenience in studying the body structure and social behavior of dolphins. At the same time, dolphin movement direction recognition helps researchers to learn more about dolphins through a series of accurate movement data. In this paper, we proposed an approach to detect the movement direction of dolphins effectively. First, the contours and skeletons are detected, which could reduce the impact of wrong detections significantly. And then another CNN-based model obtains the movement directions with feature images extracted from the previous steps. The experiment result shows the correctness and efficiency of the proposed method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Arbelaez P, Pont-Tuset J, Barron J, Marques F, Malik J (2014) Multiscale combinatorial grouping. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp 328–335

  2. Arnab A, Torr PH (2017) Pixelwise instance segmentation with a dynamically instantiated network. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 441–450

  3. Cao Z, Simon T, Wei S.E, Sheikh Y (2017) Realtime multi-person 2D pose estimation using part affinity fields. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7291–7299

  4. Dai J, He K, Sun J (2016) Instance-aware semantic segmentation via multi-task network cascades. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 3150–3158

  5. Eichner M, Ferrari V (2010) We are family: joint pose estimation of multiple persons. In: European conference on computer vision, pp 228–242. Springer, Berlin

  6. Fang HS, Xie S, Tai YW, Lu C (2017) RMPE: Regional multi-person pose estimation. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp 2334–2343

  7. Gkioxari G, Hariharan B, Girshick R, Malik J (2014) Using k-poselets for detecting people and localizing their keypoints. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 3582–3589

  8. He K, Gkioxari G, Dollár P, Girshick R (2017) Mask R-CNN. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp 2961–2969

  9. Hosang J, Benenson R, Dollár P, Schiele B (2015) What makes for effective detection proposals? IEEE Trans Pattern Anal Mach Intell 38(4):814–830

    Article  Google Scholar 

  10. Insafutdinov E, Pishchulin L, Andres B, Andriluka M, Schiele B (2016) DeeperCut: A deeper, stronger, and faster multi-person pose estimation model. In: European Conference on Computer Vision, pp 34–50. Springer, Berlin

  11. Jiang Y, Gou, Y, Zhang T, Wang K, Hu C (2017) A machine learning approach to argo data analysis in a thermocline. In: Sensors 17(10):2225

  12. Karnowski J, Hutchins E, Johnson C (2015) Dolphin detection and tracking. In: 2015 IEEE Winter Applications and Computer Vision Workshops, pp 51–56. IEEE

  13. Krizhevsky A, Sutskever I, Hinton GE (2017) ImageNet classification with deep convolutional neural networks. Commun ACM 60(6):84–90

    Article  Google Scholar 

  14. Li H, Ge X (2019) Design and application of an image classification algorithm based on semantic discrimination. Trait Signal 36(5):439–444

    Article  MathSciNet  Google Scholar 

  15. Papanikolopoulos NP, Khosla PK (1993) Adaptive robotic visual tracking: theory and experiments. IEEE Trans Autom Control 38(3):429–445

    Article  MathSciNet  MATH  Google Scholar 

  16. Pinheiro PO, Collobert R, Dollár P (2015) Learning to segment object candidates. In: Advances in Neural Information Processing Systems, pp 1990–1998

  17. Pishchulin L, Insafutdinov E, Tang S, Andres B, Andriluka M, Gehler PV, Schiele B (2016) DeepCut: joint subset partition and labeling for multi person pose estimation. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 4929–4937

  18. Prasad B, Agrawal A, Viswanathan V, Chowdhury AR, Kumar R, Panda SK (2015) A visually guided spherical underwater robot. In: 2015 IEEE Underwater Technology (UT), pp. 1–6. IEEE

  19. Qin H, Wang C, Jiang Y, Deng Z, Zhang, W (2018) Trend prediction of the 3D thermocline’s lateral boundary based on the SVR method. In: EURASIP J Wirel Commun Netw 252

  20. Ren S, He K, Girshick R, Sun J (2017) Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):1137–1149

    Article  Google Scholar 

  21. Ronchi MR, Perona P (2017) Benchmarking and error diagnosis in multi-instance pose estimation. In: Proceedings of the IEEE International Conference on Computer Vision, pp 369–378

  22. Russell BC, Torralba A, Murphy KP, Freeman WT (2008) LabelMe: a database and web-based tool for image annotation. Int J Comput Vis 77:157–173

    Article  Google Scholar 

  23. Sattar J, Dudek G (2009) Robust servo-control for underwater robots using banks of visual filters. In: 2009 IEEE International Conference on Robotics and Automation, pp 3583–3588. IEEE

  24. Wajeed MA, Sreenivasulu V (2019) Image based tumor cells identification using convolutional neural network and auto encoders. Trait Signal 36(5):445–453

    Article  Google Scholar 

  25. Wiggins SM, McDonald MA, Hildebrand JA (2012) Beaked whale and dolphin tracking using a multichannel autonomous acoustic recorder. J Acoust Soc Am 131(1):156–163

    Article  Google Scholar 

  26. Wiggins SM, Frasier KE, Elizabeth Henderson E, Hildebrand JA (2013) Tracking dolphin whistles using an autonomous acoustic recorder array. J Acoust Soc Am 133(6):3813–3818

    Article  Google Scholar 

  27. Yu SC, Ura T, Fujii T, Kondo H (2001) Navigation of autonomous underwater vehicles based on artificial underwater landmarks. In: MTS/IEEE Oceans 2001. An Ocean Odyssey. Conference Proceedings (IEEE Cat. No. 01CH37295), vol 1, pp 409–416. IEEE

  28. Zhao M, Hu C, Wei F, Wang K, Wang C, Jiang Y (2019) Real-time underwater image recognition with FPGA embedded system for convolutional neural network. In: Sensors 19(2):350

Download references

Funding

This work was supported by the National Natural Science Foundation of China under Grant 51679105, Grant 51809112, and Grant 51939003.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yu Jiang.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Qi, H., Xue, M., Peng, X. et al. Dolphin movement direction recognition using contour-skeleton information. Multimed Tools Appl 82, 21907–21923 (2023). https://doi.org/10.1007/s11042-020-09659-y

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-020-09659-y

Keywords

Navigation