Skip to main content
Log in

Global attention-assisted representation learning for vehicle re-identification

  • Original Paper
  • Published:
Signal, Image and Video Processing Aims and scope Submit manuscript

Abstract

Like pedestrian re-identification, vehicle re-identification (re-id) is an important part of building smart cities, and its purpose is to identify the same vehicle in vehicle images captured by multiple cameras. Vehicle re-id is more challenging than pedestrian re-id because many vehicles have similar colors and shapes, and their visual differences are usually very subtle. Existing vehicle re-id methods often rely on additional, expensive annotations to distinguish different vehicles. In contrast, we propose a two-branch network based on global attention mechanisms (MultiAttention-Net), which distinguishes subtle differences through adaptive learning. We introduce a global attention mechanism to highlight the differences between similar vehicles; however, compared with global appearance features, local features are more discriminant. Therefore, we propose combining global and local features to train the network to further improve the performance of vehicle re-id. During testing, only global features are used to measure the similarity between vehicle images. The experimental results show that the proposed MultiAttention-Net re-id method performs well on the challenging VeRi and VehicleID datasets.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Notes

  1. https://github.com/slpluan/vehicle_reid

References

  1. Bai, Y., Lou, Y., Dai, Y., Liu, J., Chen, Z., Duan, L.Y., Pillar, I.: Disentangled feature learning network for vehicle re-identification. In: IJCAI, pp. 474–480 (2020)

  2. Chen, X., Xiang, S., Liu, C.L., Pan, C.H.: Vehicle detection in satellite images by hybrid deep convolutional neural networks. IEEE Geosci. Remote Sens. Lett. 11(10), 1797–1801 (2014). https://doi.org/10.1109/lgrs.2014.2309695

    Article  Google Scholar 

  3. Chu, R., Sun, Y., Li, Y., Liu, Z., Zhang, C., Wei, Y.: Vehicle re-identification with viewpoint-aware metric learning. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8282–8291 (2019). https://doi.org/10.1109/iccv.2019.00837

  4. Cocchia, A.: Smart and digital city: a systematic literature review. Smart City 13–43 (2014). https://doi.org/10.1007/978-3-319-06160-3_2

  5. Dai, Y., Li, X., Liu, J., Tong, Z., Duan, L.Y.: Generalizable person re-identification with relevance-aware mixture of experts. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16145–16154 (2021)

  6. Dai, Y., Liu, J., Bai, Y., Tong, Z., Duan, L.Y.: Dual-refinement: joint label and feature refinement for unsupervised domain adaptive person re-identification. Preprint arXiv:2012.13689 (2020)

  7. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), vol. 1, pp. 886–893. IEEE (2005)

  8. Fang, F., Wang, H., Tang, P.: Image captioning with word level attention. In: 2018 25th IEEE International Conference on Image Processing (ICIP), pp. 1278–1282. IEEE (2018). https://doi.org/10.1109/icip.2018.8451558

  9. Fu, J., Zheng, H., Mei, T.: Look closer to see better: recurrent attention convolutional neural network for fine-grained image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4438–4446 (2017). https://doi.org/10.1109/cvpr.2017.476

  10. Ge, Y., Chen, D., Li, H.: Mutual mean-teaching: pseudo label refinery for unsupervised domain adaptation on person re-identification. Preprint arXiv:2001.01526 (2020)

  11. Ge, Y., Zhu, F., Chen, D., Zhao, R., Li, H.: Self-paced contrastive learning with hybrid memory for domain adaptive object re-id. Preprint arXiv:2006.02713 (2020)

  12. He, B., Li, J., Zhao, Y., Tian, Y.: Part-regularized near-duplicate vehicle re-identification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3997–4005 (2019). https://doi.org/10.1109/cvpr.2019.00412

  13. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016). https://doi.org/10.1109/cvpr.2016.90

  14. He, L., Liao, X., Liu, W., Liu, X., Cheng, P., Mei, T.: Fastreid: a pytorch toolbox for general instance re-identification. Preprint arXiv:2006.02631 (2020)

  15. He, Z., Lei, Y., Bai, S., Wu, W.: Multi-camera vehicle tracking with powerful visual features and spatial-temporal cue. In: CVPR Workshops, pp. 203–212 (2019)

  16. Itti, L., Koch, C.: Computational modelling of visual attention. Nat. Rev. Neurosci. 2(3), 194–203 (2001)

    Article  Google Scholar 

  17. Khorramshahi, P., Kumar, A., Peri, N., Rambhatla, S.S., Chen, J.C., Chellappa, R.: A dual-path model with adaptive attention for vehicle re-identification. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6132–6141 (2019). https://doi.org/10.1109/iccv.2019.00623

  18. Khorramshahi, P., Peri, N., Chen, J.c., Chellappa, R.: The devil is in the details: self-supervised attention for vehicle re-identification. In: European Conference on Computer Vision, pp. 369–386. Springer (2020)

  19. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. Preprint arXiv:1412.6980 (2014)

  20. Kuma, R., Weill, E., Aghdasi, F., Sriram, P.: Vehicle re-identification: an efficient baseline using triplet embedding. In: 2019 International Joint Conference on Neural Networks (IJCNN), pp. 1–9. IEEE (2019). https://doi.org/10.1109/ijcnn.2019.8852059

  21. Li, M., Huang, X., Zhang, Z.: Discovering discriminative geometric features with self-supervised attention for vehicle re-identification and beyond. Preprint arXiv:2010.09221 (2020)

  22. Li, Y., Wang, Y.: A multi-label image classification algorithm based on attention model. In: 2018 IEEE/ACIS 17th International Conference on Computer and Information Science (ICIS), pp. 728–731. IEEE (2018). https://doi.org/10.1109/icis.2018.8466472

  23. Liu, H., Tian, Y., Yang, Y., Pang, L., Huang, T.: Deep relative distance learning: tell the difference between similar vehicles. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2167–2175 (2016). https://doi.org/10.1109/cvpr.2016.238

  24. Liu, X., Liu, W., Ma, H., Fu, H.: Large-scale vehicle re-identification in urban surveillance videos. In: 2016 IEEE International Conference on Multimedia and Expo (ICME), pp. 1–6. IEEE (2016). https://doi.org/10.1109/icme.2016.7553002

  25. Liu, X., Liu, W., Mei, T., Ma, H.: A deep learning-based approach to progressive vehicle re-identification for urban surveillance. In: European Conference on Computer Vision, pp. 869–884. Springer (2016). https://doi.org/10.1007/978-3-319-46475-6_53

  26. Lowe, D.G.: Object recognition from local scale-invariant features. In: Proceedings of the Seventh IEEE International Conference on Computer Vision, vol. 2, pp. 1150–1157. IEEE (1999)

  27. Luo, H., Gu, Y., Liao, X., Lai, S., Jiang, W.: Bag of tricks and a strong baseline for deep person re-identification. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2019). https://doi.org/10.1109/cvprw.2019.00190

  28. Meng, D., Li, L., Liu, X., Li, Y., Yang, S., Zha, Z.J., Gao, X., Wang, S., Huang, Q.: Parsing-based view-aware embedding network for vehicle re-identification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7103–7112 (2020)

  29. Ojala, T., Pietikainen, M., Maenpaa, T.: Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 24(7), 971–987 (2002)

    Article  Google Scholar 

  30. Peng, J., Wang, H., Zhao, T., Fu, X.: Learning multi-region features for vehicle re-identification with context-based ranking method. Neurocomputing 359, 427–437 (2019). https://doi.org/10.1016/j.neucom.2019.06.013

    Article  Google Scholar 

  31. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. Preprint arXiv:1409.1556 (2014)

  32. Sun, Y., Zheng, L., Li, Y., Yang, Y., Tian, Q., Wang, S.: Learning part-based convolutional features for person re-identification. IEEE Trans. Pattern Anal. Mach. Intell. 43(3), 902–917 (2021)

    Article  Google Scholar 

  33. Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9 (2015). https://doi.org/10.1109/cvpr.2015.7298594

  34. Wang, G., Yuan, Y., Chen, X., Li, J., Zhou, X.: Learning discriminative features with multiple granularities for person re-identification. In: Proceedings of the 26th ACM International Conference on Multimedia, pp. 274–282 (2018)

  35. Wang, H., Hou, J., Chen, N.: A survey of vehicle re-identification based on deep learning. IEEE Access 7, 172443–172469 (2019). https://doi.org/10.1109/access.2019.2956172

    Article  Google Scholar 

  36. Wei, X.S., Zhang, C.L., Liu, L., Shen, C., Wu, J.: Coarse-to-fine: a RNN-based hierarchical attention model for vehicle re-identification. In: Asian Conference on Computer Vision, pp. 575–591. Springer (2018). https://doi.org/10.1007/978-3-030-20890-5_37

  37. Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: CBAM: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3–19 (2018)

  38. Wu, F., Yan, S., Smith, J.S., Zhang, B.: Vehicle re-identification in still images: application of semi-supervised learning and re-ranking. Signal Process. Image Commun. 76, 261–271 (2019). https://doi.org/10.1016/j.image.2019.04.021

    Article  Google Scholar 

  39. Zhang, F., Ma, Y., Yuan, G., Zhang, H., Ren, J.: Multiview image generation for vehicle reidentification. Appl. Intell. 51, 5665–5682 (2021)

    Article  Google Scholar 

  40. Zhang, Y., Liu, D., Zha, Z.J.: Improving triplet-wise training of convolutional neural network for vehicle re-identification. In: 2017 IEEE International Conference on Multimedia and Expo (ICME), pp. 1386–1391. IEEE (2017). https://doi.org/10.1109/icme.2017.8019491

  41. Zheng, A., Lin, X., Li, C., He, R., Tang, J.: Attributes guided feature learning for vehicle re-identification. Preprint arXiv:1905.08997 (2019)

  42. Zhou, Y., Shao, L.: Cross-view GAN based vehicle generation for re-identification. In: BMVC, vol. 1, pp. 1–12 (2017). https://doi.org/10.5244/c.31.186

  43. Zhu, C., Zhao, Y., Huang, S., Tu, K., Ma, Y.: Structured attentions for visual question answering. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1291–1300 (2017). https://doi.org/10.1109/iccv.2017.145

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yuanyuan Chen.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

This work was supported by Department of Science and Technology of Sichuan Province, China (Grant Nos. 20ZDYF2060 and 2021YFQ0010)

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Song, L., Zhou, X. & Chen, Y. Global attention-assisted representation learning for vehicle re-identification. SIViP 16, 807–815 (2022). https://doi.org/10.1007/s11760-021-02021-1

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11760-021-02021-1

Keywords

Navigation