Skip to main content
Log in

Domain-invariant feature extraction and fusion for cross-domain person re-identification

  • Original article
  • Published:
The Visual Computer Aims and scope Submit manuscript

Abstract

There are notable style differences among person re-identification (ReID) datasets, such as brightness, tone, resolution, background, and clothing style, that result in serious challenges for cross-domain person ReID. Two methods are usually used to solve these problems. One is to remove style differences between datasets by applying specific modules, such as instance normalization (IN). However, this method will filter out large amounts of valuable information for ReID. The other is to use person attributes as auxiliary information, but this method does not deeply explore the relationship between attribute features and global features, resulting in underutilized attribute information. We propose the domain-invariant feature extraction and fusion (DFEF), which consists of the attention and style normalization (ASN) and the attribute feature extraction and fusion (AFEF). The ASN module integrates spatial and channel attention modules on the basis of the IN layer to effectively remove the style differences between datasets and recovers the filtered-out information, which is useful for ReID. The AFEF module includes the attribute branch and the feature fusion module. For the attribute branch, we embed the convolutional block attention module (CBAM) into the attribute branch and adopt the multi-label focal loss (MLFL) to improve the accuracy of attribute recognition. For the feature fusion module, we propose the dispersion reweighting strategy to explore the correlation between attribute features and global features. The proposed DFEF method achieves 30.1% and 35.0% mAP on Market-1501 \(\rightarrow \) DukeMTMC-reID and DukeMTMC-reID \(\rightarrow \) Market-1501, respectively.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  1. Ye, M., Shen, J., Lin, G., Xiang, T., Shao, L., Hoi, S.C.H.: Deep learning for person re-identification: a survey and outlook. IEEE Trans. Pattern Anal. Mach. Intell. (2021). https://doi.org/10.1109/TPAMI.2021.3054775

  2. Fan, X., Jiang, W., Luo, H., Mao, W.: Modality-transfer generative adversarial network and dual-level unified latent representation for visible thermal person re-identification. Vis. Comput. 38, 1–16 (2020)

  3. Xie, J., Ge, Y., Zhang, J., Huang, S., Chen, F., Wang, H.: Low-resolution assisted three-stream network for person re-identification. Vis. Comput. 1–11 (2021) https://doi.org/10.1007/s00371-021-02127-0

  4. Chen, Z., Lv, X., Sun, T., Zhao, C., Chen, W.: FLAG: feature learning with additional guidance for person search. Vis. Comput. 37(4), 685–693 (2021)

    Article  Google Scholar 

  5. Chen, G., Lin, C., Ren, L., Lu, J., Zhou, J.: Self-critical attention learning for person re-identification. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV) (2019)

  6. Muandet, K., Balduzzi, D., Schölkopf, B.: Domain generalization via invariant feature representation. In: Proceedings of the 30th International Conference on International Conference on Machine Learning—Volume 28, ICML’13, pp. I-10–I-18. JMLR.org (2013)

  7. Jia, J., Ruan, Q., Hospedales, T.M.: Frustratingly easy person re-identification: generalizing person re-identification in practice. CoRR (2019). http://arxiv.org/abs/1905.03422

  8. Ulyanov, D., Vedaldi, A., Lempitsky, V.: Instance normalization: the missing ingredient for fast stylization. arXiv preprint arXiv:1607.08022 (2016)

  9. Lin, Y., Zheng, L., Zheng, Z., Wu, Y., Hu, Z., Yan, C., Yang, Y.: Improving person re-identification by attribute and identity learning. Pattern Recognit. 95, 151–161 (2019)

    Article  Google Scholar 

  10. Schumann, A., Stiefelhagen, R.: Person re-identification by deep learning attribute-complementary information. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1435–1443 (2017)

  11. Su, C., Zhang, S., Xing, J., Gao, W., Tian, Q.: Multi-type attributes driven multi-camera person re-identification. Pattern Recognit. 75, 77–89 (2018)

    Article  Google Scholar 

  12. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June (2016)

  13. Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: Convolutional_Block_Attention.pdf. ECCV (2018)

  14. Li, Y., Shi, F., Hou, S., Li, J., Li, C., Yin, G.: Feature pyramid attention model and multi-label focal loss for pedestrian attribute recognition. IEEE Access 8, 164570–164579 (2020)

    Article  Google Scholar 

  15. Song, J., Yang, Y., Song, Y., Xiang, T., Hospedales, T.M.: Generalizable person re-identification by domain-invariant mapping network. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 719–728 (2019)

  16. Kumar, D., Siva, P., Marchwica, P., Wong, A.: Fairest of them all: establishing a strong baseline for cross-domain person reid. arXiv:1907.12016 (2019)

  17. Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 1510–1519 (2017)

  18. Ulyanov, D., Vedaldi, A., Lempitsky, V.: Instance normalization: the missing ingredient for fast stylization. 07 (2016)

  19. Pan, X., Luo, P., Shi, J., Tang, X.: Two at once: enhancing learning and generalization capacities via ibn-net. In: ECCV (2018)

  20. Zhou, K., Yang, Y., Cavallaro, A., Xiang, T.: Omni-scale feature learning for person re-identification. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 3701–3711 (2019)

  21. Su, C., Zhang, S., Yang, F., Zhang, G., Tian, Q., Gao, W., Davis, L.S.: Attributes driven tracklet-to-tracklet person re-identification using latent prototypes space mapping. Pattern Recognit. 66, 4–15 (2017)

    Article  Google Scholar 

  22. Su, C., Yang, F., Zhang, S., Tian, Q., Davis, L.S., Gao, W.: Multi-task learning with low rank attribute embedding for multi-camera person re-identification. IEEE Trans. Pattern Anal. Mach. Intell. 40(5), 1167–1181 (2018)

    Article  Google Scholar 

  23. Layne, R., Hospedales, T.M., Gong, S.: Person re-identification by attributes. In: BMVC (2012)

  24. Liu, X., Song, M., Zhao, Q., Tao, D., Chen, C., Bu, J.: Attribute-restricted latent topic model for person re-identification. Pattern Recognit. 45, 4204–4213 (2012)

    Article  Google Scholar 

  25. Layne, R., Hospedales, T., Gong, S.: Re-id: Hunting attributes in the wild. In: Proceedings of the British Machine Vision Conference. BMVA Press (2014)

  26. Peng, P., Tian, Y., Xiang, T., Wang, Y., Huang, T.: Joint learning of semantic and latent attributes. In: ECCV (2016)

  27. Franco, A., Oliveira, L.: Convolutional covariance features: conception, integration and performance in person re-identification. Pattern Recognit. 61, 593–609 (2017)

    Article  Google Scholar 

  28. Yin, Z., Zheng, W., Wu, A., Yu, H.-X., Wan, H., Guo, X., Huang, F., Lai, J.: Adversarial attribute-image person re-identification. In: IJCAI (2018)

  29. Wang, J., Zhu, X., Gong, S., Li, W.: Transferable joint attribute-identity deep learning for unsupervised person re-identification. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2275–2284 (2018)

  30. Li, Y., Shi, F., Hou, S., Li, J., Li, C., Yin, G.: Feature pyramid attention model and multi-label focal loss for pedestrian attribute recognition. IEEE Access 8, 164570–164579 (2020)

    Article  Google Scholar 

  31. Zheng, L., Shen, L., Tian, L., Wang, S., Wang, J., Tian, Q.: Scalable person re-identification: a benchmark. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1116–1124 (2015)

  32. Zheng, Z., Zheng, L., Yang, Y.: Unlabeled samples generated by GAN improve the person re-identification baseline in vitro. In: Proceedings of the IEEE International Conference on Computer Vision (2017)

  33. Li, W., Zhao, R., Xiao, T., Wang, X.: DeepReID: deep filter pairing neural network for person re-identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 152–159 (2014)

  34. Zhong, Z., Zheng, L., Cao, D., Li, S.: Re-ranking person re-identification with k-reciprocal encoding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1318–1327 (2017)

  35. Wei, L., Zhang, S., Gao, W., Tian, Q.: Person transfer GAN to bridge domain gap for person re-identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 79–88 (2018)

  36. Jia, J., Ruan, Q., Hospedales, T.M.: Frustratingly easy person re-identification: generalizing person Re-ID in practice. arXiv preprint arXiv:1905.03422 (2019)

  37. Deng, W., Zheng, L., Ye, Q., Kang, G., Yang, Y., Jiao, J.: Image-image domain adaptation with preserved self-similarity and domain-dissimilarity for person re-identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June (2018)

  38. Zhong, Z., Zheng, L., Li, S., Yang, Y.: Generalizing a person retrieval model hetero- and homogeneously. In: Proceedings of the European Conference on Computer Vision (ECCV), September (2018)

  39. Chang, X., Yang, Y., Xiang, T., Hospedales, T.M.: Disjoint label space transfer learning with common factorised space. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 3288–3295 (2019)

  40. Lin, Y., Dong, X., Zheng, L., Yan, Y., Yang, Y.: A bottom-up clustering approach to unsupervised person re-identification. In: National Conference on Artificial Intelligence (2019)

  41. Liu, J., Zha, Z.-J., Chen, D., Hong, R., Wang, M.: Adaptive transfer network for cross-domain person re-identification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June (2019)

  42. Liang, W., Wang, G., Lai, J., Zhu, J.-Y.: M2M-GAN: many-to-many generative adversarial transfer learning for person re-identification. CoRR, arXiv:1811.03768 (2018)

  43. Chen, Y., Zhu, X., Gong, S.: Instance-guided context rendering for cross-domain person re-identification. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October (2019)

  44. Liao, S., Shao, L.: Interpretable and generalizable person re-identification with query-adaptive convolution and temporal lifting. In: Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XI 16, pp. 456–474. Springer (2020)

Download references

Acknowledgements

We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work, and there is no professional or other personal interest of any nature or kind in any product, service or company that could be construed as influencing the review of the manuscript. The data that support the findings of this study are available online. These datasets were derived from the following public resources: [Market-1501, DukeMTMC-reID, CUHK03-NP, MSMT17].

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Guangqiang Yin.

Ethics declarations

Conflict of interest

The authors have no financial or proprietary interests in any material discussed in this article.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jia, Z., Li, Y., Tan, Z. et al. Domain-invariant feature extraction and fusion for cross-domain person re-identification. Vis Comput 39, 1205–1216 (2023). https://doi.org/10.1007/s00371-022-02398-1

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00371-022-02398-1

Keywords

Navigation