Abstract
Detecting and tracking social marine mammals, including dolphins, can help to explain their social dynamics, predict their behavior, and measure the impact of human interference. The underwater environment is very special and different from the land environment: acoustic recorders are the main equipment for researchers to track dolphins at a long distance. Close-range detection of visual data is often seriously affected by the underwater environment, because the underwater environment is highly dynamic and the light source has attenuation in the deep water. Nonetheless, compared with acoustic information, visual data can provide more detailed information at low cost. The videos and images of dolphins provide researchers with great convenience in studying the body structure and social behavior of dolphins. At the same time, dolphin movement direction recognition helps researchers to learn more about dolphins through a series of accurate movement data. In this paper, we proposed an approach to detect the movement direction of dolphins effectively. First, the contours and skeletons are detected, which could reduce the impact of wrong detections significantly. And then another CNN-based model obtains the movement directions with feature images extracted from the previous steps. The experiment result shows the correctness and efficiency of the proposed method.
Similar content being viewed by others
References
Arbelaez P, Pont-Tuset J, Barron J, Marques F, Malik J (2014) Multiscale combinatorial grouping. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp 328–335
Arnab A, Torr PH (2017) Pixelwise instance segmentation with a dynamically instantiated network. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 441–450
Cao Z, Simon T, Wei S.E, Sheikh Y (2017) Realtime multi-person 2D pose estimation using part affinity fields. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7291–7299
Dai J, He K, Sun J (2016) Instance-aware semantic segmentation via multi-task network cascades. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 3150–3158
Eichner M, Ferrari V (2010) We are family: joint pose estimation of multiple persons. In: European conference on computer vision, pp 228–242. Springer, Berlin
Fang HS, Xie S, Tai YW, Lu C (2017) RMPE: Regional multi-person pose estimation. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp 2334–2343
Gkioxari G, Hariharan B, Girshick R, Malik J (2014) Using k-poselets for detecting people and localizing their keypoints. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 3582–3589
He K, Gkioxari G, Dollár P, Girshick R (2017) Mask R-CNN. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp 2961–2969
Hosang J, Benenson R, Dollár P, Schiele B (2015) What makes for effective detection proposals? IEEE Trans Pattern Anal Mach Intell 38(4):814–830
Insafutdinov E, Pishchulin L, Andres B, Andriluka M, Schiele B (2016) DeeperCut: A deeper, stronger, and faster multi-person pose estimation model. In: European Conference on Computer Vision, pp 34–50. Springer, Berlin
Jiang Y, Gou, Y, Zhang T, Wang K, Hu C (2017) A machine learning approach to argo data analysis in a thermocline. In: Sensors 17(10):2225
Karnowski J, Hutchins E, Johnson C (2015) Dolphin detection and tracking. In: 2015 IEEE Winter Applications and Computer Vision Workshops, pp 51–56. IEEE
Krizhevsky A, Sutskever I, Hinton GE (2017) ImageNet classification with deep convolutional neural networks. Commun ACM 60(6):84–90
Li H, Ge X (2019) Design and application of an image classification algorithm based on semantic discrimination. Trait Signal 36(5):439–444
Papanikolopoulos NP, Khosla PK (1993) Adaptive robotic visual tracking: theory and experiments. IEEE Trans Autom Control 38(3):429–445
Pinheiro PO, Collobert R, Dollár P (2015) Learning to segment object candidates. In: Advances in Neural Information Processing Systems, pp 1990–1998
Pishchulin L, Insafutdinov E, Tang S, Andres B, Andriluka M, Gehler PV, Schiele B (2016) DeepCut: joint subset partition and labeling for multi person pose estimation. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 4929–4937
Prasad B, Agrawal A, Viswanathan V, Chowdhury AR, Kumar R, Panda SK (2015) A visually guided spherical underwater robot. In: 2015 IEEE Underwater Technology (UT), pp. 1–6. IEEE
Qin H, Wang C, Jiang Y, Deng Z, Zhang, W (2018) Trend prediction of the 3D thermocline’s lateral boundary based on the SVR method. In: EURASIP J Wirel Commun Netw 252
Ren S, He K, Girshick R, Sun J (2017) Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):1137–1149
Ronchi MR, Perona P (2017) Benchmarking and error diagnosis in multi-instance pose estimation. In: Proceedings of the IEEE International Conference on Computer Vision, pp 369–378
Russell BC, Torralba A, Murphy KP, Freeman WT (2008) LabelMe: a database and web-based tool for image annotation. Int J Comput Vis 77:157–173
Sattar J, Dudek G (2009) Robust servo-control for underwater robots using banks of visual filters. In: 2009 IEEE International Conference on Robotics and Automation, pp 3583–3588. IEEE
Wajeed MA, Sreenivasulu V (2019) Image based tumor cells identification using convolutional neural network and auto encoders. Trait Signal 36(5):445–453
Wiggins SM, McDonald MA, Hildebrand JA (2012) Beaked whale and dolphin tracking using a multichannel autonomous acoustic recorder. J Acoust Soc Am 131(1):156–163
Wiggins SM, Frasier KE, Elizabeth Henderson E, Hildebrand JA (2013) Tracking dolphin whistles using an autonomous acoustic recorder array. J Acoust Soc Am 133(6):3813–3818
Yu SC, Ura T, Fujii T, Kondo H (2001) Navigation of autonomous underwater vehicles based on artificial underwater landmarks. In: MTS/IEEE Oceans 2001. An Ocean Odyssey. Conference Proceedings (IEEE Cat. No. 01CH37295), vol 1, pp 409–416. IEEE
Zhao M, Hu C, Wei F, Wang K, Wang C, Jiang Y (2019) Real-time underwater image recognition with FPGA embedded system for convolutional neural network. In: Sensors 19(2):350
Funding
This work was supported by the National Natural Science Foundation of China under Grant 51679105, Grant 51809112, and Grant 51939003.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Qi, H., Xue, M., Peng, X. et al. Dolphin movement direction recognition using contour-skeleton information. Multimed Tools Appl 82, 21907–21923 (2023). https://doi.org/10.1007/s11042-020-09659-y
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-020-09659-y