Abstract
Image classification is one of the most popular and important problems in computer vision. In self-driving cars image classification is used to classify detected traffic signs. Modern state-of-the-art algorithms based on deep neural networks use softmax function to interpret the output of the network as the probability that the input data belongs to a certain class. This approach works well, however it has several disadvantages. More precisely, it is necessary to know the number of classes in advance, and if one wants to add a new class, then it is necessary to retrain the network. Moreover, a large number of images of each class are required. In the case of road signs, datasets may contain only the most frequent signs while ignoring rarely used ones. Thus, the traffic signs recognition module in autonomous cars will not recognize traffic signs not included into training dataset, which can lead to accidents. In this paper we put forward another approach that does not have disadvantages of networks with softmax. The approach is based on learning image embeddings in which models are trained to bring closer objects of one class and to move away objects of other classes in embeddings space. Therefore, having even a small number of images of rare classes it becomes possible to create a working classification system. In this work, we test the applicability of these algorithms in the traffic signs classification problem, and also compare its accuracy with neural networks with softmax and with networks pre-trained on softmax. We developed publicly available toolbox for training and testing embedding networks with different loss functions, backbone models, training strategies and other configuration parameters and embedding space visualization tools. All our experiments were carried out on the russian road signs dataset. To simplify the process of conducting training experiments, a framework for embedding learning based neural networks making was created. The framework can be found at https://github.com/RocketFlash/EmbeddingNet.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9 (2015)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097–1105 (2012)
Koch, G., Zemel, R., Salakhutdinov, R.: Siamese neural networks for one-shot image recognition. In: ICML Deep Learning Workshop, vol. 2 (2015)
Chopra, S., Hadsell, R., LeCun, Y.: Learning a similarity metric discriminatively, with application to face verification. In: CVPR (1), pp. 539–546 (2005)
Schroff, F., Kalenichenko, D., Philbin, J.: FaceNet: a unified embedding for face recognition and clustering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 815–823 (2015)
Hoffer, E., Ailon, N.: Deep metric learning using triplet network. In: International Workshop on Similarity-Based Pattern Recognition, pp. 84–92. Springer, Cham (2015)
Haralick, R.M., Shanmugam, K., Dinstein, I.H.: Textural features for image classification. IEEE Trans. Syst. Man Cybern. 6, 610–621 (1973)
Wu, C.Y., Manmatha, R., Smola, A.J., Krahenbuhl, P.: Sampling matters in deep embedding learning. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2840–2848 (2017)
Shakhuro, V.I., Konouchine, A.S.: Russian traffic sign images dataset. Comput. Opt. 40(2), 294–300 (2016)
Liu, L., Jiang, H., He, P., Chen, W., Liu, X., Gao, J., Han, J.: On the variance of the adaptive learning rate and beyond. arXiv preprint arXiv:1908.03265 (2019)
Acknowledgments
This research was supported by the Russian Education Ministry under the Agreement No. 075-10-2018-017 on granting a subsidy from 11.26.2018 “Development of a commercial urban transport with the intellectual driver assistance system “City Pilot”. The customer is a federal target program. The unique identifier of the project RFMEFI60918X0005. Industrial partner – KAMAZ.
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Yagfarov, R., Ostankovich, V., Akhmetzyanov, A. (2020). Traffic Sign Classification Using Embedding Learning Approach for Self-driving Cars. In: Ahram, T., Taiar, R., Gremeaux-Bader, V., Aminian, K. (eds) Human Interaction, Emerging Technologies and Future Applications II. IHIET 2020. Advances in Intelligent Systems and Computing, vol 1152. Springer, Cham. https://doi.org/10.1007/978-3-030-44267-5_27
Download citation
DOI: https://doi.org/10.1007/978-3-030-44267-5_27
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-44266-8
Online ISBN: 978-3-030-44267-5
eBook Packages: EngineeringEngineering (R0)