Abstract
Aiming at the issue that the existing aerial work safety belt wearing detection model cannot meet the real-time operation on edge devices, this paper proposes a lightweight aerial work safety belt detection model with higher accuracy. First, the model is made lightweight by introducing Ghost convolution and model pruning. Second, for complex scenarios involving occlusion, color confusion, etc., the model’s performance is optimized by introducing the new up-sampling operator, the attention mechanism, and the feature fusion network. Lastly, the model is trained using knowledge distillation to compensate for accuracy loss resulting from the lightweight design, thereby maintain a higher accuracy. Experimental results based on the Guangdong Power Grid Intelligence Challenge safety belt wearable dataset show that, in the comparison experiments, the improved model, compared with the mainstream object detection algorithm YOU ONLY LOOK ONCE v5s (YOLOv5s), has only 8.7% of the parameters of the former with only 3.7% difference in the mean Average Precision (mAP.50) metrics and the speed is improved by 100.4%. Meanwhile, the ablation experiments show that the improved model’s parameter count is reduced by 66.9% compared with the original model, while mAP.50 decreases by only 1.9%. The overhead safety belt detection model proposed in this paper combines the model’s lightweight design, SimAM attention mechanism, Bidirectional Feature Pyramid Network feature fusion network, Carafe operator, and knowledge distillation training strategy, enabling the model to maintain lightweight and real-time performance while achieving high detection accuracy.









Similar content being viewed by others
References
Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multiperson 2d pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7291–7299 (2017). https://doi.org/10.1109/CVPR.2017.143
Cao, J., Guo, Z., Pan, L., Ding, X.: Seat belt wearing detection in aerial work scenarios. J. Hunan Univ. Sci. Technol. 37, 92–99 (2022). https://doi.org/10.13582/j.cnki.1672-9102.2022.01.013
Fang, C., Xiang, H., Leng, C., Chen, J., Yu, Q.: Research on real-time detection of safety harness wearing of workshop personnel based on yolov5 and openpose. Sustainability (2022). https://doi.org/10.3390/su14105872
Feng, Z., Zhang, W., Zheng, Z.: Seat belt detection for working at height based on mask r-cnn. Comput Syst Appl 30, 202–207 (2021). https://doi.org/10.15888/j.cnki.csa.007812
Ge, Z., Liu, S., Wang, F., Li, Z., Sun, J.: Yolox: Exceeding yolo series in 2021 (2021). arXiv preprint arXiv:2107.08430. https://doi.org/10.48550/arXiv.2107.08430
Guo, H., Lin, H., Zhang, S., Li, S.: Image-based seat belt detection. In: Proceedings of 2011 IEEE International Conference on Vehicular Electronics and Safety, pp. 161–164. IEEE (2011). https://doi.org/10.1109/ICVES.2011.5983807
Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: Ghostnet: more features from cheap operations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1580–1589 (2020). https://doi.org/10.1109/CVPR42600.2020.00165
He, K., Gkioxari, G., Doll’ar, P., Girshick, R.: Mask rcnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961–2969 (2017). https://doi.org/10.1109/ICCV.2017.322
Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network (2015). arXiv preprint arXiv:1503.02531. https://doi.org/10.48550/arXiv.1503.02531
Howard, A., Sandler, M., Chu, G., Chen, L.C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., Vasudevan, V., et al.: Searching for mobilenetv3 (2019). In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1314–1324. https://doi.org/10.1109/ICCV.2019.00140
Li, H., Kadav, A., Durdanovic, I., Samet, H., Graf, H.P.: Pruning filters for efficient convnets (2016). CoRR arXiv:abs/1608.08710. https://doi.org/10.48550/arXiv.1608.08710
Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: Ssd: single shot multibox detector. In: 14th European Conference on Computer Vision (ECCV), vol. 9905, pp. 21–37. Springer International Publishing Ag, Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_2
Mingxing, T., Ruoming, P., Le Quoc, V.E.: Scalable and efficient object detection. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2020). https://doi.org/10.1109/CVPR42600.2020.01079
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788 (2016). https://doi.org/10.1109/cvpr.2016.91
Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137–1149 (2016). https://doi.org/10.1109/TPAMI.2016.2577031
Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 618–626 (2017). https://doi.org/10.1109/ICCV.2017.74
Tian, Z., Chu, X., Wang, X., Wei, X., Shen, C.: Fully convolutional one-stage 3d object detection on lidar range images. Adv. Neural Inf. Process. Syst. 35, 34899–34911 (2022). https://doi.org/10.48550/arXiv.2205.13764
Wang, J., Chen, K., Xu, R., Liu, Z., Loy, C.C., Lin, D.: Carafe: content-aware reassembly of features. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3007–3016 (2019). https://doi.org/10.1109/ICCV.2019.00310
Wang, W., Xie, E., Song, X., Zang, Y., Wang, W., Lu, T., Yu, G., Shen, C.: Efficient and accurate arbitrary shaped text detection with pixel aggregation network. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8440–8449 (2019). https://doi.org/10.48550/arXiv.1908.05900
Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: Yolov7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464–7475 (2023). https://doi.org/10.48550/arXiv.2207.02696
Yang, L., Zhang, R.Y., Li, L., Xie, X.: Simam: a simple, parameter-free attention module for convolutional neural networks. In: International Conference on Machine Learning, pp. 11863–11874. PMLR (2021)
Zhao, J., Li, Y.: Falling risk and prevention analysis of high altitude workers in power system. J. New Ind. 9, 34–39 (2019). https://doi.org/10.19335/j.cnki.2095-6649.2019.11.008
Acknowledgements
This work is supported by the Hebei Province Graduate Student Innovation Ability Training Funding Project (Grant:CXZZSS2024163) and the Key Research and Development Projects in Hebei Province (Grant:20310103D). The authors would like to thank Mr.Tang from Electric Power Research Institute of Yunnan Electric Power Grid for data support. Special thanks to the Mr.Zhang from The University of Queensland for language support. Special thanks to Ms. Zhang from the Department of Computer of North China Electric Power University for her guidance in the drawing of some of the figure.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Liu, L., Huang, K., Bai, Y. et al. Real-time detection model of electrical work safety belt based on lightweight improved YOLOv5. J Real-Time Image Proc 21, 151 (2024). https://doi.org/10.1007/s11554-024-01533-6
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11554-024-01533-6