Skip to main content
Log in

Slim-YOLO-PR_KD: an efficient pose-varied object detection method for underground coal mine

  • Research
  • Published:
Journal of Real-Time Image Processing Aims and scope Submit manuscript

Abstract

Real-time object detection in underground coal mine is a crucial task in the development of AI-assisted supervision systems. Due to the complex environment of the underground coal mine, limited computing resources, and the variability of object poses, the general object detection algorithms cannot provide good performance. Hence, an improved underground pose-varied object detection method named Slim-YOLO-PR_KD has been proposed. By designing an efficient pose-varied attention module (EPA) for the backbone network, providing a receive field block (RFB) module for the neck network, and optimizing the loss function, the underground pose-varied detection model YOLO-PR is obtained, which achieved good accuracy but reduced speed. For YOLO-PR, the study improved the original module by designing RFB_SK, a lightweight C2f_GSG module, a shared parameter detection head and selectively replaced modules to slim down the whole network, resulting in a lightweight detection model Slim-YOLO-PR. By using an attention guided knowledge distillation of underground object detection method and using YOLO-PR as the teacher model, the efficient pose-varied detection model Slim-YOLO-PR_KD for coal mine underground is proposed. The experimental results show that compared with the baseline model, the proposed Slim-YOLO-PR_KD has a faster detection speed, achieving higher detection accuracy while reducing model parameters and computational complexity by 42% and 46% respectively, making it capable of performing real-time underground detection tasks. Compared with other general detection models, Slim-YOLO-PR_KD exhibits excellent performance in real-time pose-varied object detection tasks in complex environments of underground coal mines.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18
Fig. 19

Similar content being viewed by others

Data availability

The supporting datasets in this article can be obtained through the following link: https://pan.baidu.com/s/1LpdL3p6kXfkqqE2Umyzh7A?pwd=2024.

References

  1. Talaei Khoei, T., Ould Slimane, H., Kaabouch, N.: Deep learning: systematic review, models, challenges, and research directions. Neural Comput. Appl. 35, 23103–23124 (2023)

    Article  Google Scholar 

  2. Tian, Y., Yang, X., Yang, J., Mao, K., Yao, Y.: Evolution dynamic of intelligent construction strategy of coal mine enterprises in China. Heliyon 8, e10933 (2022)

    Article  Google Scholar 

  3. Du, Y., Zhang, H., Liang, L., Zhang, J., Song, B.: Applications of machine vision in coal mine fully mechanized tunneling faces: a review. IEEE Access 11, 102871–102898 (2023)

    Article  Google Scholar 

  4. Miao, D.J., Wang, W.H., Lv, Y.Y., Liu, L., Yao, K.X., Sui, X.H.: Research on the classification and control of human factor characteristics of coal mine accidents based on K-Means clustering analysis. Int. J. Ind. Ergon. 97, 103481 (2023)

    Article  Google Scholar 

  5. You, S., Zhu, H., Li, M., Li, Y., Tang, C.: Tracking system for a coal mine drilling robot for low-illumination environments. Appl. Sci.-Basel 13, 568 (2023)

    Article  Google Scholar 

  6. Guo, J., Zou, X., Wang, Z., Pan, J.: IEEE: Joint detection and tracking with movable camera and its application to a drilling robot in underground coal mine. In: 19th IEEE International Conference on Mechatronics and Automation (IEEE ICMA), pp. 946–951 (2022)

  7. Kaur, R., Singh, S.: A comprehensive review of object detection with deep learning. Digit Signal Process 132, 103812 (2023)

    Article  Google Scholar 

  8. Bing, Z., Wang, X., Dong, Z., Dong, L., He, T.: A novel edge computing architecture for intelligent coal mining system. Wirel Netw 29, 1545–1554 (2023)

    Article  Google Scholar 

  9. Liu, J., Mu, X., Liu, Z., Li, H.: Human skeleton behavior recognition model based on multi-object pose estimation with spatiotemporal semantics. Mach. Vis. Appl. 34(3), 44 (2023)

    Article  Google Scholar 

  10. Wu, Q.T., Zhou, Y.M.: IEEE: Multi-Pose Moving Human Detection based on Unmanned Aerial Vehicle in Real-Time. In: 14th IEEE Conference on Industrial Electronics and Applications (ICIEA), pp. 608–613. (2019)

  11. Gao, C.Q., Ye, S., Tian, H., Yan, Y.: Multi-scale single-stage pose detection with adaptive sample training in the classroom scene. Knowl.-Based Syst. 222, 107008 (2021)

    Article  Google Scholar 

  12. Bao, Q., Liu, W., Cheng, Y.H., Zhou, B.Y., Mei, T.: Pose-guided tracking-by-detection: robust multi-person pose tracking. IEEE Trans. Multimedia 23, 161–175 (2021)

    Article  Google Scholar 

  13. Wu, L., Huang, C., Fei, L.K., Zhao, S.P., Zhao, J.C., Cui, Z.W., Xu, Y.: Video-based fall detection using human pose and constrained generative adversarial network. IEEE Trans. Circuits Syst. Video Technol. 34, 2179–2194 (2024)

    Article  Google Scholar 

  14. Wu, L.G., Zhang, L., Chen, L., Shi, J.H., Wan, J.F.: A lightweight and multisource information fusion method for real-time monitoring of lump coal on mining conveyor belts. Int. J. Intell. Syst. 2023(1), 5327122 (2023)

    Google Scholar 

  15. Chen, K., Du, B., Wang, Y., Wang, G., He, J.: The real-time detection method for coal gangue based on YOLOv8s-GSC. J. Real-Time Image Process. 21(2), 37 (2024)

    Article  Google Scholar 

  16. Zhang, Y., Zhou, Y.M.: IEEE: YOLOv5 Based Pedestrian Safety Detection in Underground Coal Mines. In: IEEE International Conference on Robotics and Biomimetics (IEEE ROBIO), pp. 1700–1705. (2021)

  17. Zeng, Q.L., Xu, W.Q., Gao, K.D.: Measurement method and experiment of hydraulic support group attitude and straightness based on binocular vision. IEEE Trans. Instrum. Meas. 72, 1–14 (2023)

    Article  Google Scholar 

  18. Pan, L.H., Duan, Y.X., Zhang, Y.J., Xie, B.H., Zhang, R.: A lightweight algorithm based on YOLOv5 for relative position detection of hydraulic support at coal mining faces. J. Real-Time Image Process. 20(2), 40 (2023)

    Article  Google Scholar 

  19. Fan, Y.B., Mao, S.J., Li, M., Wu, Z., Kang, J.T.: CM-YOLOv8: lightweight YOLO for coal mine fully mechanized mining face. Sensors 24(6), 1866 (2024)

    Article  Google Scholar 

  20. Shao, X., Liu, S., Li, X., Lyu, Z., Li, H.: Rep-YOLO: an efficient detection method for mine personnel. J. Real-Time Image Process. 21(2), 28 (2024)

    Article  Google Scholar 

  21. Ouyang, D., He, S., Zhan, J., Guo, H., Huang, Z., Luo, M.L., Zhang, G.-L.: Efficient Multi-Scale Attention Module with Cross-Spatial Learning. In: ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1–5 (2023)

  22. Xiong, Y., Li, Z., Chen, Y., Wang, F., Zhu, X., Luo, J., Wang, W., Lu, T., Li, H., Qiao, Y., Lu, L., Zhou, J., Dai, J.: Efficient Deformable ConvNets: Rethinking Dynamic and Sparse Operator for Vision Applications. ArXiv abs/2401.06197 (2024)

  23. Liu, S., Huang, D., Wang, Y.: Receptive Field Block Net for Accurate and Fast Object Detection. In: 15th European Conference on Computer Vision (ECCV), pp. 404–419. (2018)

  24. Liu, C., Wang, K., Li, Q., Zhao, F., Zhao, K., Ma, H.: Powerful-IoU: More straightforward and faster bounding box regression loss with a nonmonotonic focusing mechanism. Neural Netw. 170, 276–284 (2024)

    Article  Google Scholar 

  25. Lau, K.W., Po, L.-M., Rehman, Y.A.U.: Large separable kernel attention: rethinking the large kernel attention design in CNN. Expert Syst. Appl. 236, 121352 (2024)

    Article  Google Scholar 

  26. Chollet, F.: IEEE: Xception—Deep Learning with Depthwise Separable Convolutions. In: 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1800–1807 (2017)

  27. Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: IEEE: GhostNet: More Features from Cheap Operations. In: IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1577–1586 (2020)

  28. Li, J., Wen, Y., He, L.: IEEE: SCConv: Spatial and Channel Reconstruction Convolution for Feature Redundancy. In: IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6153–6162 (2023)

  29. Zhang, X., Guo, W., Xing, Y., Wang, W., Yin, H., Zhang, Y.: AugFCOS: augmented fully convolutional one-stage object detection network. Pattern Recognit. 134, 109098 (2023)

    Article  Google Scholar 

  30. Zhang, L., Ma, K.: Structured knowledge distillation for accurate and efficient object detection. IEEE Trans. Pattern Anal. Mach. Intell. 45, 15706–15724 (2023)

    Article  Google Scholar 

  31. Yang, W., Zhang, X., Ma, B., Wang, Y., Wu, Y., Yan, J., Liu, Y., Zhang, C., Wan, J., Wang, Y.: An open dataset for intelligent recognition and classification of abnormal condition in longwall mining. Sci. Data 10, 416 (2023)

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported in part by the National Natural Science Foundation of China under Grants 52274160 and 51874300, the National Natural Science Foundation of China-Shanxi Provincial People's Government Coal-Based Low Carbon Joint Fund under Grant U1510115, "Jiangsu Distinguished Professor" project in Jiangsu Province (140923070), and the Fundamental Research Funds for the Central Universities Grants 2023QN1079, and 2024QN11050.

Funding

National Natural Science Foundation of China, 52274160.

Author information

Authors and Affiliations

Authors

Contributions

HM made contributions to theory, methodology, partial experiments, and initial draft writing. JL participated in theoretical argumentation, article review, and final draft writing. YG made contributions to some experiments. WC participated in the revision and polishing of the paper. TX and ZW provided experimental equipment and made certain contributions to chart drawing.

Corresponding author

Correspondence to Wei Chen.

Ethics declarations

Conflict of interest

The authors have no competing interests to declare that are relevant to the content of this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Mu, H., Liu, J., Guan, Y. et al. Slim-YOLO-PR_KD: an efficient pose-varied object detection method for underground coal mine. J Real-Time Image Proc 21, 160 (2024). https://doi.org/10.1007/s11554-024-01539-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11554-024-01539-0

Keywords

Navigation