Skip to main content

Traffic Sign Classification Using Embedding Learning Approach for Self-driving Cars

  • Conference paper
  • First Online:
Human Interaction, Emerging Technologies and Future Applications II (IHIET 2020)

Abstract

Image classification is one of the most popular and important problems in computer vision. In self-driving cars image classification is used to classify detected traffic signs. Modern state-of-the-art algorithms based on deep neural networks use softmax function to interpret the output of the network as the probability that the input data belongs to a certain class. This approach works well, however it has several disadvantages. More precisely, it is necessary to know the number of classes in advance, and if one wants to add a new class, then it is necessary to retrain the network. Moreover, a large number of images of each class are required. In the case of road signs, datasets may contain only the most frequent signs while ignoring rarely used ones. Thus, the traffic signs recognition module in autonomous cars will not recognize traffic signs not included into training dataset, which can lead to accidents. In this paper we put forward another approach that does not have disadvantages of networks with softmax. The approach is based on learning image embeddings in which models are trained to bring closer objects of one class and to move away objects of other classes in embeddings space. Therefore, having even a small number of images of rare classes it becomes possible to create a working classification system. In this work, we test the applicability of these algorithms in the traffic signs classification problem, and also compare its accuracy with neural networks with softmax and with networks pre-trained on softmax. We developed publicly available toolbox for training and testing embedding networks with different loss functions, backbone models, training strategies and other configuration parameters and embedding space visualization tools. All our experiments were carried out on the russian road signs dataset. To simplify the process of conducting training experiments, a framework for embedding learning based neural networks making was created. The framework can be found at https://github.com/RocketFlash/EmbeddingNet.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 349.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 449.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)

    Google Scholar 

  2. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)

  3. Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9 (2015)

    Google Scholar 

  4. Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097–1105 (2012)

    Google Scholar 

  5. Koch, G., Zemel, R., Salakhutdinov, R.: Siamese neural networks for one-shot image recognition. In: ICML Deep Learning Workshop, vol. 2 (2015)

    Google Scholar 

  6. Chopra, S., Hadsell, R., LeCun, Y.: Learning a similarity metric discriminatively, with application to face verification. In: CVPR (1), pp. 539–546 (2005)

    Google Scholar 

  7. Schroff, F., Kalenichenko, D., Philbin, J.: FaceNet: a unified embedding for face recognition and clustering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 815–823 (2015)

    Google Scholar 

  8. Hoffer, E., Ailon, N.: Deep metric learning using triplet network. In: International Workshop on Similarity-Based Pattern Recognition, pp. 84–92. Springer, Cham (2015)

    Google Scholar 

  9. Haralick, R.M., Shanmugam, K., Dinstein, I.H.: Textural features for image classification. IEEE Trans. Syst. Man Cybern. 6, 610–621 (1973)

    Article  Google Scholar 

  10. Wu, C.Y., Manmatha, R., Smola, A.J., Krahenbuhl, P.: Sampling matters in deep embedding learning. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2840–2848 (2017)

    Google Scholar 

  11. Shakhuro, V.I., Konouchine, A.S.: Russian traffic sign images dataset. Comput. Opt. 40(2), 294–300 (2016)

    Article  Google Scholar 

  12. Liu, L., Jiang, H., He, P., Chen, W., Liu, X., Gao, J., Han, J.: On the variance of the adaptive learning rate and beyond. arXiv preprint arXiv:1908.03265 (2019)

Download references

Acknowledgments

This research was supported by the Russian Education Ministry under the Agreement No. 075-10-2018-017 on granting a subsidy from 11.26.2018 “Development of a commercial urban transport with the intellectual driver assistance system “City Pilot”. The customer is a federal target program. The unique identifier of the project RFMEFI60918X0005. Industrial partner – KAMAZ.

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Rauf Yagfarov , Vladislav Ostankovich or Aydar Akhmetzyanov .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yagfarov, R., Ostankovich, V., Akhmetzyanov, A. (2020). Traffic Sign Classification Using Embedding Learning Approach for Self-driving Cars. In: Ahram, T., Taiar, R., Gremeaux-Bader, V., Aminian, K. (eds) Human Interaction, Emerging Technologies and Future Applications II. IHIET 2020. Advances in Intelligent Systems and Computing, vol 1152. Springer, Cham. https://doi.org/10.1007/978-3-030-44267-5_27

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-44267-5_27

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-44266-8

  • Online ISBN: 978-3-030-44267-5

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics