Skip to main content
Log in

Real-time detection model of electrical work safety belt based on lightweight improved YOLOv5

  • Research
  • Published:
Journal of Real-Time Image Processing Aims and scope Submit manuscript

Abstract

Aiming at the issue that the existing aerial work safety belt wearing detection model cannot meet the real-time operation on edge devices, this paper proposes a lightweight aerial work safety belt detection model with higher accuracy. First, the model is made lightweight by introducing Ghost convolution and model pruning. Second, for complex scenarios involving occlusion, color confusion, etc., the model’s performance is optimized by introducing the new up-sampling operator, the attention mechanism, and the feature fusion network. Lastly, the model is trained using knowledge distillation to compensate for accuracy loss resulting from the lightweight design, thereby maintain a higher accuracy. Experimental results based on the Guangdong Power Grid Intelligence Challenge safety belt wearable dataset show that, in the comparison experiments, the improved model, compared with the mainstream object detection algorithm YOU ONLY LOOK ONCE v5s (YOLOv5s), has only 8.7% of the parameters of the former with only 3.7% difference in the mean Average Precision (mAP.50) metrics and the speed is improved by 100.4%. Meanwhile, the ablation experiments show that the improved model’s parameter count is reduced by 66.9% compared with the original model, while mAP.50 decreases by only 1.9%. The overhead safety belt detection model proposed in this paper combines the model’s lightweight design, SimAM attention mechanism, Bidirectional Feature Pyramid Network feature fusion network, Carafe operator, and knowledge distillation training strategy, enabling the model to maintain lightweight and real-time performance while achieving high detection accuracy.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

References

  1. Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multiperson 2d pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7291–7299 (2017). https://doi.org/10.1109/CVPR.2017.143

  2. Cao, J., Guo, Z., Pan, L., Ding, X.: Seat belt wearing detection in aerial work scenarios. J. Hunan Univ. Sci. Technol. 37, 92–99 (2022). https://doi.org/10.13582/j.cnki.1672-9102.2022.01.013

    Article  Google Scholar 

  3. Fang, C., Xiang, H., Leng, C., Chen, J., Yu, Q.: Research on real-time detection of safety harness wearing of workshop personnel based on yolov5 and openpose. Sustainability (2022). https://doi.org/10.3390/su14105872

    Article  Google Scholar 

  4. Feng, Z., Zhang, W., Zheng, Z.: Seat belt detection for working at height based on mask r-cnn. Comput Syst Appl 30, 202–207 (2021). https://doi.org/10.15888/j.cnki.csa.007812

    Article  Google Scholar 

  5. Ge, Z., Liu, S., Wang, F., Li, Z., Sun, J.: Yolox: Exceeding yolo series in 2021 (2021). arXiv preprint arXiv:2107.08430. https://doi.org/10.48550/arXiv.2107.08430

  6. Guo, H., Lin, H., Zhang, S., Li, S.: Image-based seat belt detection. In: Proceedings of 2011 IEEE International Conference on Vehicular Electronics and Safety, pp. 161–164. IEEE (2011). https://doi.org/10.1109/ICVES.2011.5983807

  7. Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: Ghostnet: more features from cheap operations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1580–1589 (2020). https://doi.org/10.1109/CVPR42600.2020.00165

  8. He, K., Gkioxari, G., Doll’ar, P., Girshick, R.: Mask rcnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961–2969 (2017). https://doi.org/10.1109/ICCV.2017.322

  9. Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network (2015). arXiv preprint arXiv:1503.02531. https://doi.org/10.48550/arXiv.1503.02531

  10. Howard, A., Sandler, M., Chu, G., Chen, L.C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., Vasudevan, V., et al.: Searching for mobilenetv3 (2019). In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1314–1324. https://doi.org/10.1109/ICCV.2019.00140

  11. Li, H., Kadav, A., Durdanovic, I., Samet, H., Graf, H.P.: Pruning filters for efficient convnets (2016). CoRR arXiv:abs/1608.08710. https://doi.org/10.48550/arXiv.1608.08710

  12. Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: Ssd: single shot multibox detector. In: 14th European Conference on Computer Vision (ECCV), vol. 9905, pp. 21–37. Springer International Publishing Ag, Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_2

  13. Mingxing, T., Ruoming, P., Le Quoc, V.E.: Scalable and efficient object detection. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2020). https://doi.org/10.1109/CVPR42600.2020.01079

  14. Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788 (2016). https://doi.org/10.1109/cvpr.2016.91

  15. Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137–1149 (2016). https://doi.org/10.1109/TPAMI.2016.2577031

    Article  Google Scholar 

  16. Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 618–626 (2017). https://doi.org/10.1109/ICCV.2017.74

  17. Tian, Z., Chu, X., Wang, X., Wei, X., Shen, C.: Fully convolutional one-stage 3d object detection on lidar range images. Adv. Neural Inf. Process. Syst. 35, 34899–34911 (2022). https://doi.org/10.48550/arXiv.2205.13764

    Article  Google Scholar 

  18. Wang, J., Chen, K., Xu, R., Liu, Z., Loy, C.C., Lin, D.: Carafe: content-aware reassembly of features. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3007–3016 (2019). https://doi.org/10.1109/ICCV.2019.00310

  19. Wang, W., Xie, E., Song, X., Zang, Y., Wang, W., Lu, T., Yu, G., Shen, C.: Efficient and accurate arbitrary shaped text detection with pixel aggregation network. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8440–8449 (2019). https://doi.org/10.48550/arXiv.1908.05900

  20. Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: Yolov7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464–7475 (2023). https://doi.org/10.48550/arXiv.2207.02696

  21. Yang, L., Zhang, R.Y., Li, L., Xie, X.: Simam: a simple, parameter-free attention module for convolutional neural networks. In: International Conference on Machine Learning, pp. 11863–11874. PMLR (2021)

  22. Zhao, J., Li, Y.: Falling risk and prevention analysis of high altitude workers in power system. J. New Ind. 9, 34–39 (2019). https://doi.org/10.19335/j.cnki.2095-6649.2019.11.008

    Article  Google Scholar 

Download references

Acknowledgements

This work is supported by the Hebei Province Graduate Student Innovation Ability Training Funding Project (Grant:CXZZSS2024163) and the Key Research and Development Projects in Hebei Province (Grant:20310103D). The authors would like to thank Mr.Tang from Electric Power Research Institute of Yunnan Electric Power Grid for data support. Special thanks to the Mr.Zhang from The University of Queensland for language support. Special thanks to Ms. Zhang from the Department of Computer of North China Electric Power University for her guidance in the drawing of some of the figure.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kaiye Huang.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liu, L., Huang, K., Bai, Y. et al. Real-time detection model of electrical work safety belt based on lightweight improved YOLOv5. J Real-Time Image Proc 21, 151 (2024). https://doi.org/10.1007/s11554-024-01533-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11554-024-01533-6

Keywords