Skip to main content
Log in

A lightweight network for vehicle detection based on embedded system

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

The target detection algorithm has a high accuracy rate for the detection of general objects. However, for specific kinds of objects, especially those with complex application scenarios and high requirements on recognition rates, the accuracy and real-time performance are still far behind the actual application requirements. Meanwhile, the high real-time performance of most current target detection algorithm depends on high-performance computers or Graphics Processing Units (GPUs), which limits their application in low-performance embedded system. In this paper, an embedded vehicle detection algorithm was proposed. Compared to the existing general detection algorithms, the network structure of the proposed algorithm has been improved in many aspects. Firstly, the overall network structure of YOLOv3-tiny has been improved, and the way of feature extraction has been modified to increase the network in terms of detection speed and recognition accuracy. Secondly, a spatial pyramid pooling structure is introduced into the network to enhance the learning capability, and the spatial pyramid pooling structure is also improved by increasing the number of pooling layers in the structure and cascading multiple groups of pooling layers of different sizes. Thirdly, to speed up target detection process, the original clustering algorithm is replaced by Kmean++ clustering algorithm, and the original IoU (The Intersection over Union) loss function is replaced by GIoU (Generalized IoU). Fourthly, according to the scale factor of batch normalization layer, the model is compressed and trimmed to remove redundant channels, and a small model more suitable for embedded system application is obtained. Finally, Raspberry Pi 4B and Inter NCS2 neural computing stick are used as the embedded system with the input image size of 416 × 416. Experimental tests on UA-DETRAC dataset obtain higher mAP than other advanced existing methods, indicating that the network model proposed in this paper could effectively improve the real time and accuracy of multi-vehicle target detection.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  1. Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision And Pattern Recognition pp 580–587

  2. Min R, Quan H, Cui Z, Cao Z, Pi Y, Xu, Z (2019) SAR Target detection using AdaBoost via GPU acceleration. In: IGARSS 2019–2019 IEEE International Geoscience and Remote Sensing Symposium pp 1180–1183. IEEE

  3. Mendez J, Molina M, Rodriguez N, Cuellar MP, Morales DP (2021) Camera-LiDAR multi-level sensor fusion for target detection at the network edge. Sensors 21(12):3992

    Article  Google Scholar 

  4. Srivastava S, Vani B, Sadistap S (2020) Machine-vision based handheld embedded system to extract quality parameters of citrus cultivars. J Food Measure Char 14(5):2746–2759

    Article  Google Scholar 

  5. Jiahua C, Yunzhou Z, Zheng W, Jiwei L (2019) Light-weight object detection networks for embedded platform. Acta Optica Sinica 39(4):0415006

    Article  Google Scholar 

  6. Dai J, Li Y, He K, Sun J (2016) R-fcn: Object detection via region-based fully convolutional networks. Adv Neural Inform Proc Syst, 29

  7. Girshick R (2015) Fast r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision pp 1440–1448

  8. Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 28

  9. Mask RCNN, He K, Gkioxari G, Dollár P, Girshick R (2017) Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer pp 2961–2969

  10. Lin TY, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition pp 2117–2125.

  11. Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition pp 779–788

  12. Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767

  13. Yu J, Zhang W (2021) Face mask wearing detection algorithm based on improved YOLO-v4. Sensors 21(9):3263

    Article  Google Scholar 

  14. Yang G, Feng W, Jin J, Lei Q, Li X, Gui G, & Wang W (2020) Face mask recognition system with YOLOV5 based on image recognition. In: 2020 IEEE 6th International Conference on Computer and Communications (ICCC) pp 1398–1404 IEEE

  15. Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, Berg AC (2016) Ssd: Single shot multibox detector. In European Conference on Computer Vision pp 21–37. Springer, Cham

  16. Zhai S, Shang D, Wang S, Dong S (2020) DF-SSD: An improved SSD object detection algorithm based on DenseNet and feature fusion. IEEE access 8:24344–24357

    Article  Google Scholar 

  17. Lin TY, Goyal P, Girshick R, He K, Dollár P (2017) Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, pp 2980–2988

  18. Chen X, Gupta A (2017) Spatial memory for context reasoning in object detection. In ICCV

  19. Liu Y, Han J, Zhang Q, Shan C (2019) Deep salient object detection with contextual information guidance. IEEE Trans Image Process 29:360–374

    Article  MathSciNet  MATH  Google Scholar 

  20. Tang S, Li Y, Deng L, Zhang Y (2017) Object localization based on proposal fusion. IEEE Trans Multimedia 19(9):2105–2116

    Article  Google Scholar 

  21. Ma W, Wu Y, Cen F, Wang G (2020) Mdfn: Multi-scale deep feature learning network for object detection. Pattern Recogn 100:107149

    Article  Google Scholar 

  22. Liu S, Huang D (2018) Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV) pp 385–400

  23. Li J, Wang Y, Wang C, Tai Y, Qian J, Yang J, Huang F (2019) DSFD: dual shot face detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition pp 5060–5069

  24. Liu JJ, Hou Q, Cheng MM, Feng J, Jiang J (2019). A simple pooling-based design for real-time salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition pp 3917–3926

  25. Wang T, Anwer RM, Cholakkal H, Khan FS, Pang Y, Shao L (2019) Learning rich features at high-speed for single-shot object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision pp 1971–1980

  26. Chen S, Zhan R, Wang W, Zhang J (2020) Learning slimming SAR ship object detector through network pruning and knowledge distillation. IEEE J Selected Topics Appl Earth Observations Remote Sensing 14:1267–1282

    Article  Google Scholar 

  27. Zhou D, Fang J, Song X, Guan C, Yin J, Dai Y, Yang R (2019). Iou loss for 2d/3d object detection. In 2019 International Conference on 3D Vision (3DV), pp 85–94, IEEE

  28. Zheng Z, Wang P, Liu W, Li J, Ye R, Ren D (2020) Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 34, No. 07, pp. 12993–13000)

  29. Rezatofighi H, Tsoi N, Gwak J, Sadeghian A, Reid I, Savarese S (2019) Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 658–666

  30. Liu Z, Li J, Shen Z, Huang G, Yan S, Zhang C (2017) Learning efficient convolutional networks through network slimming. In Proceedings of the IEEE International Conference on Computer Vision (pp. 2736–2744)

Download references

Funding

Science and Technology Cooperation Project of The Xinjiang Production and Construction Corps (No. 2019BC008).

Author information

Authors and Affiliations

Authors

Contributions

Huanhuan Wu contributed to conceptualization, methodology, validation, investigation, writing. Yuantao Hua contributed to methodology, validation, model compressed. Hua Zou contributed to conceptualization, methodology, formal analysis, visualization, funding acquisition, validation, writing—review, editing, supervision. Gang Ke contributed to data curation, validation.

Corresponding author

Correspondence to Hua Zou.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Ethical approval

No need ethical approval Science and Technology Project of Tarim University (No. TDZKZD202104).

Informed Consent

No need informed consent.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wu, H., Hua, Y., Zou, H. et al. A lightweight network for vehicle detection based on embedded system. J Supercomput 78, 18209–18224 (2022). https://doi.org/10.1007/s11227-022-04596-z

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-022-04596-z

Keywords

Navigation