Abstract
The target detection algorithm has a high accuracy rate for the detection of general objects. However, for specific kinds of objects, especially those with complex application scenarios and high requirements on recognition rates, the accuracy and real-time performance are still far behind the actual application requirements. Meanwhile, the high real-time performance of most current target detection algorithm depends on high-performance computers or Graphics Processing Units (GPUs), which limits their application in low-performance embedded system. In this paper, an embedded vehicle detection algorithm was proposed. Compared to the existing general detection algorithms, the network structure of the proposed algorithm has been improved in many aspects. Firstly, the overall network structure of YOLOv3-tiny has been improved, and the way of feature extraction has been modified to increase the network in terms of detection speed and recognition accuracy. Secondly, a spatial pyramid pooling structure is introduced into the network to enhance the learning capability, and the spatial pyramid pooling structure is also improved by increasing the number of pooling layers in the structure and cascading multiple groups of pooling layers of different sizes. Thirdly, to speed up target detection process, the original clustering algorithm is replaced by Kmean++ clustering algorithm, and the original IoU (The Intersection over Union) loss function is replaced by GIoU (Generalized IoU). Fourthly, according to the scale factor of batch normalization layer, the model is compressed and trimmed to remove redundant channels, and a small model more suitable for embedded system application is obtained. Finally, Raspberry Pi 4B and Inter NCS2 neural computing stick are used as the embedded system with the input image size of 416 × 416. Experimental tests on UA-DETRAC dataset obtain higher mAP than other advanced existing methods, indicating that the network model proposed in this paper could effectively improve the real time and accuracy of multi-vehicle target detection.
Similar content being viewed by others
References
Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision And Pattern Recognition pp 580–587
Min R, Quan H, Cui Z, Cao Z, Pi Y, Xu, Z (2019) SAR Target detection using AdaBoost via GPU acceleration. In: IGARSS 2019–2019 IEEE International Geoscience and Remote Sensing Symposium pp 1180–1183. IEEE
Mendez J, Molina M, Rodriguez N, Cuellar MP, Morales DP (2021) Camera-LiDAR multi-level sensor fusion for target detection at the network edge. Sensors 21(12):3992
Srivastava S, Vani B, Sadistap S (2020) Machine-vision based handheld embedded system to extract quality parameters of citrus cultivars. J Food Measure Char 14(5):2746–2759
Jiahua C, Yunzhou Z, Zheng W, Jiwei L (2019) Light-weight object detection networks for embedded platform. Acta Optica Sinica 39(4):0415006
Dai J, Li Y, He K, Sun J (2016) R-fcn: Object detection via region-based fully convolutional networks. Adv Neural Inform Proc Syst, 29
Girshick R (2015) Fast r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision pp 1440–1448
Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 28
Mask RCNN, He K, Gkioxari G, Dollár P, Girshick R (2017) Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer pp 2961–2969
Lin TY, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition pp 2117–2125.
Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition pp 779–788
Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767
Yu J, Zhang W (2021) Face mask wearing detection algorithm based on improved YOLO-v4. Sensors 21(9):3263
Yang G, Feng W, Jin J, Lei Q, Li X, Gui G, & Wang W (2020) Face mask recognition system with YOLOV5 based on image recognition. In: 2020 IEEE 6th International Conference on Computer and Communications (ICCC) pp 1398–1404 IEEE
Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, Berg AC (2016) Ssd: Single shot multibox detector. In European Conference on Computer Vision pp 21–37. Springer, Cham
Zhai S, Shang D, Wang S, Dong S (2020) DF-SSD: An improved SSD object detection algorithm based on DenseNet and feature fusion. IEEE access 8:24344–24357
Lin TY, Goyal P, Girshick R, He K, Dollár P (2017) Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, pp 2980–2988
Chen X, Gupta A (2017) Spatial memory for context reasoning in object detection. In ICCV
Liu Y, Han J, Zhang Q, Shan C (2019) Deep salient object detection with contextual information guidance. IEEE Trans Image Process 29:360–374
Tang S, Li Y, Deng L, Zhang Y (2017) Object localization based on proposal fusion. IEEE Trans Multimedia 19(9):2105–2116
Ma W, Wu Y, Cen F, Wang G (2020) Mdfn: Multi-scale deep feature learning network for object detection. Pattern Recogn 100:107149
Liu S, Huang D (2018) Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV) pp 385–400
Li J, Wang Y, Wang C, Tai Y, Qian J, Yang J, Huang F (2019) DSFD: dual shot face detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition pp 5060–5069
Liu JJ, Hou Q, Cheng MM, Feng J, Jiang J (2019). A simple pooling-based design for real-time salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition pp 3917–3926
Wang T, Anwer RM, Cholakkal H, Khan FS, Pang Y, Shao L (2019) Learning rich features at high-speed for single-shot object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision pp 1971–1980
Chen S, Zhan R, Wang W, Zhang J (2020) Learning slimming SAR ship object detector through network pruning and knowledge distillation. IEEE J Selected Topics Appl Earth Observations Remote Sensing 14:1267–1282
Zhou D, Fang J, Song X, Guan C, Yin J, Dai Y, Yang R (2019). Iou loss for 2d/3d object detection. In 2019 International Conference on 3D Vision (3DV), pp 85–94, IEEE
Zheng Z, Wang P, Liu W, Li J, Ye R, Ren D (2020) Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 34, No. 07, pp. 12993–13000)
Rezatofighi H, Tsoi N, Gwak J, Sadeghian A, Reid I, Savarese S (2019) Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 658–666
Liu Z, Li J, Shen Z, Huang G, Yan S, Zhang C (2017) Learning efficient convolutional networks through network slimming. In Proceedings of the IEEE International Conference on Computer Vision (pp. 2736–2744)
Funding
Science and Technology Cooperation Project of The Xinjiang Production and Construction Corps (No. 2019BC008).
Author information
Authors and Affiliations
Contributions
Huanhuan Wu contributed to conceptualization, methodology, validation, investigation, writing. Yuantao Hua contributed to methodology, validation, model compressed. Hua Zou contributed to conceptualization, methodology, formal analysis, visualization, funding acquisition, validation, writing—review, editing, supervision. Gang Ke contributed to data curation, validation.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Ethical approval
No need ethical approval Science and Technology Project of Tarim University (No. TDZKZD202104).
Informed Consent
No need informed consent.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Wu, H., Hua, Y., Zou, H. et al. A lightweight network for vehicle detection based on embedded system. J Supercomput 78, 18209–18224 (2022). https://doi.org/10.1007/s11227-022-04596-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11227-022-04596-z