Abstract
In recent years, target detection framework based on deep learning has made brilliant achievements. However, real-life traffic sign detection remains a great challenge for most of the state-of-the-art object detection methods. The existing deep learning models are inadequate to effectively extract the features of small traffic signs from large images in real-world conditions. In this paper, we address the small traffic sign detection challenge by proposing a novel small traffic sign detection method based on a highly efficient end-to-end deep network model. The proposed method features fast speed and high precision as it attaches three key insights to the established You Only Look Once (YOLOv3) architecture and other correlated algorithms. Besides, network pruning is appropriately introduced to minimize network redundancy and model size while keeping a competitive detection accuracy. Furtherly, four scale prediction branches are also adopted to significantly enrich the feature maps of multi-scales prediction. In our method, we adjust the loss function to balance the contribution of error source to the total loss. The effectiveness, and robustness of the network is further proved with experiments on Tsinghua-Tencent 100 K traffic sign dataset. The experimental results indicate that our proposed method has achieved better accuracy than that of the original YOLOv3 model. Compared with the schemes in relevant literatures our proposed method not only emerges performance superiors in detection recall and accuracy, but also achieves 1.9–2.7x improvement in detection speed.
Similar content being viewed by others
References
Liu, H., Liu, Y., & Sun, F. (2014). Traffic sign recognition using group sparse coding. Information Sciences, 266, 75–89.
Abdi, L., & Meddeb, A. (2018). Spatially enhanced bags of visual words representation to improve traffic signs recognition. Journal of Signal Processing Systems, 90(12), 1729–1741.
Chen, Y., Zhao, D., Lv, L., & Zhang, Q. (2018). Multi-task learning for dangerous object detection in autonomous driving. Information Sciences, 432, 559–571.
Maldonado-Bascón, S., Lafuente-Arroyo, S., Gil-Jimenez, P., Gómez-Moreno, H., & López-Ferreras, F. (2007). Road-sign detection and recognition based on support vector machines. IEEE Transactions on Intelligent Transportation Systems, 8(2), 264–278.
Jang, C., Kim, C., Kim, D., Lee, M., & Sunwoo, M. (2014). Multiple exposure images based traffic light recognition. In 2014 IEEE Intelligent Vehicles Symposium Proceedings, (pp. 1313-1318).
De Charette, R., & Nashashibi, F. (2009). Real time visual traffic lights recognition based on spot light detection and adaptive traffic lights templates. In 2009 IEEE Intelligent Vehicles Symposium, (pp. 358-363).
Cai, Z., Gu, M., & Li, Y. (2012). Real-time arrow traffic light recognition system for intelligent vehicle. In Proceedings of the International Conference on Image Processing, Computer Vision, and Pattern Recognition (IPCV), (pp. 1).
Bangquan, X., & Xiong, W. X. (2019). Real-time embedded traffic sign recognition using efficient convolutional neural network. IEEE Access, 7, 53330–53346.
Liu, Z., Du, J., Tian, F., & Wen, J. (2019). MR-CNN: A multi-scale region-based convolutional neural network for small traffic sign recognition. IEEE Access, 7, 57120–57128.
Girshick, R., Donahue, J., Darrell, T., & Malik, J. (2014). Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, (pp. 580-587).
Girshick, R. (2015). Fast r-cnn. In Proceedings of the IEEE international conference on computer vision, (pp. 1440-1448).
Ren, S., He, K., Girshick, R., & Sun, J. (2017). Faster R-CNN: Towards real-time object detection with region proposal networks. In Advances in neural information processing systems, (Vol. 39, pp. 1137-1149, Vol. 6).
Dai, J., Li, Y., He, K., & Sun, J. (2016). R-fcn: Object detection via region-based fully convolutional networks. In Advances in neural information processing systems, (pp. 379-387).
Everingham, M., Van Gool, L., Williams, C. K., Winn, J., & Zisserman, A. J. I. J. O. C. V. (2010). The pascal visual object classes (voc) challenge. International Journal of Computer Vision, 88(2), 303–338.
Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., Berg, A. C., & Fei-Fei, L. (2015). Imagenet large scale visual recognition challenge. International Journal of Computer Vision, 115(3), 211–252.
Meng, Z., Fan, X., Chen, X., Chen, M., & Tong, Y. (2017). Detecting small signs from large images. In 2017 IEEE International Conference on Information Reuse and Integration (IRI), (pp. 217-224).
Yang, T. T., Long, X., Sangaiah, A. K., Zheng, Z. G., & Tong, C. (2018). Deep detection network for real-life traffic sign in vehicular networks. Computer Networks, 136, 95–104.
Tian, Y., Gelernter, J., Wang, X., Li, J., & Yu, Y. (2019). Traffic sign detection using a multi-scale recurrent attention network. IEEE Transactions on Intelligent Transportation Systems., 20, 4466–4475.
Zhu, Z., Liang, D., Zhang, S., Huang, X., Li, B., & Hu, S. (2016). Traffic-sign detection and classification in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (pp. 2110-2118).
Song, S., Que, Z., Hou, J., Du, S., & Song, Y. (2019). An efficient convolutional neural network for small traffic sign detection. Journal of Systems Architecture., 97, 269–277.
Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., et al. (2016). SSD: Single shot MultiBox detector. European conference on computer vision, 21–37.
Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognitionn, (pp. 779-788).
Redmon, J., & Farhadi, A. (2017). YOLO9000: Better, faster, stronger. In Proceedings of the IEEE conference on computer vision and pattern recognition, (pp. 6517-6525).
Redmon, J., & Farhadi, A. (2018). Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767, doi:1804.02767.
Huval, B., Wang, T., Tandon, S., Kiske, J., Song, W., Pazhayampallil, J., et al. (2015). An empirical evaluation of deep learning on highway driving. arXiv preprint arXiv:1504.01716.
Li, J., Liang, X., Wei, Y., Xu, T., Feng, J., & Yan, S. (2017). Perceptual generative adversarial networks for small object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (pp. 1222-1230).
Lu, Y., Lu, J., Zhang, S., & Hall, P. (2018). Traffic signal detection and classification in street views using an attention model. Computational Visual Media, 4(3), 253–266.
Jain, A., Mishra, A., Shukla, A., & Tiwari, R. (2019). A novel genetically optimized convolutional neural network for traffic sign recognition: A new benchmark on Belgium and Chinese traffic sign datasets. Neural Processing Letters, 50(3), 3019-3043.
Kim, J., Lee, S., Oh, T.-H., & Kweon, I. S. (2018). Co-domain embedding using deep quadruplet networks for unseen traffic sign recognition. In Thirty-Second AAAI Conference on Artificial Intelligence,
Acknowledgments
This work was supported by National Key Research and Development Project under grant 2019YFC0117302, National Natural Science Foundation of China under grant 62004201, Key Projects of Bureau of International Cooperation Chinese Academy of Sciences under grant 184131KYSB20160018, National Natural Science Foundation of China under grant No. U1831118, and NSFC Youth Fund under grant No. 61704179.
Author information
Authors and Affiliations
Corresponding authors
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Wan, J., Ding, W., Zhu, H. et al. An Efficient Small Traffic Sign Detection Method Based on YOLOv3. J Sign Process Syst 93, 899–911 (2021). https://doi.org/10.1007/s11265-020-01614-2
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11265-020-01614-2