skip to main content
research-article

Hybrid Modality Metric Learning for Visible-Infrared Person Re-Identification

Authors Info & Claims
Published:25 January 2022Publication History
Skip Abstract Section

Abstract

Visible-infrared person re-identification (Re-ID) has received increasing research attention for its great practical value in night-time surveillance scenarios. Due to the large variations in person pose, viewpoint, and occlusion in the same modality, as well as the domain gap brought by heterogeneous modality, this hybrid modality person matching task is quite challenging. Different from the metric learning methods for visible person re-ID, which only pose similarity constraints on class level, an efficient metric learning approach for visible-infrared person Re-ID should take both the class-level and modality-level similarity constraints into full consideration to learn sufficiently discriminative and robust features. In this article, the hybrid modality is divided into two types, within modality and cross modality. We first fully explore the variations that hinder the ranking results of visible-infrared person re-ID and roughly summarize them into three types: within-modality variation, cross-modality modality-related variation, and cross-modality modality-unrelated variation. Then, we propose a comprehensive metric learning framework based on four kinds of paired-based similarity constraints to address all the variations within and cross modality. This framework focuses on both class-level and modality-level similarity relationships between person images. Furthermore, we demonstrate the compatibility of our framework with any paired-based loss functions by giving detailed implementation of combing it with triplet loss and contrastive loss separately. Finally, extensive experiments of our approach on SYSU-MM01 and RegDB demonstrate the effectiveness and superiority of our proposed metric learning framework for visible-infrared person Re-ID.

REFERENCES

  1. [1] A Yuanxin Zhu, A Zhao Yang, A Li Wang, A Sai Zhao, A Xiao Hu, and B Dapeng Tao. 2020. Hetero-Center loss for cross-modality person Re-identification. Neurocomputing 386 (2020), 97109.Google ScholarGoogle ScholarCross RefCross Ref
  2. [2] Chen Weihua, Chen Xiaotang, Zhang Jianguo, and Huang Kaiqi. 2017. Beyond triplet loss: A deep quadruplet network for person re-identification. In Proceedings of the Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref
  3. [3] Cheng De, Gong Yihong, Zhou Sanping, Wang Jinjun, and Zheng Nanning. 2016. Person re-identification by multi-channel parts-based CNN with improved triplet loss function. In Computer Vision and Pattern Recognition.Google ScholarGoogle Scholar
  4. [4] Choi Seokeon, Lee Sumin, Kim Youngeun, Kim Taekyung, and Kim Changick. 2020. Hi-CMD: Hierarchical cross-modality disentanglement for visible-infrared person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR’20). IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Dai Pingyang, Ji Rongrong, Wang Haibin, Wu Qiong, and Huang Yuyu. 2018. Cross-modality person re-identification with generative adversarial training. In Proceedings of the 27th International Joint Conference on Artificial Intelligence IJCAI-18. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. [6] Dalal N.. 2005. Histograms of oriented gradients for human detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR’05). Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. [7] Dat Nguyen, Hyung Hong, Ki Kim, and Kang Park. 2017. Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors 17, 3 (2017), 605. DOI: DOI: https://doi.org/10.3390/s17030605Google ScholarGoogle ScholarCross RefCross Ref
  8. [8] Weijian Deng, Liang Zheng, Guoliang Kang, Yi Yang, Qixiang Ye, and Jianbin Jiao. 2018. Image-image domain adaptation with preserved self-similarity and domain-dissimilarity for person re-identification. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE.Google ScholarGoogle Scholar
  9. [9] Zhangxiang Feng, Jianhuang Lai, and Xiaohua Xie. 2019. Learning modality-specific representations for visible-infrared person re-identification. IEEE Transactions on Image Processing 29 (2019), 579–590.Google ScholarGoogle Scholar
  10. [10] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Y. Bengio. 2014. Generative adversarial nets. In Advances in Neural Information Processing Systems (NIPS’14). 2672–2680. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. [11] Wang Guan’An, Zhang Tianzhu, Cheng Jian, Liu Si, Yang Yang, and Hou Zengguang. 2020. RGB-infrared cross-modality person re-identification via joint pixel and feature alignment. In IEEE/CVF International Conference on Computer Vision (ICCV’19). IEEE.Google ScholarGoogle Scholar
  12. [12] Hao Lup, Wei Jiang, Xing Fan, and Sipeng Zhang. 2019. A survey on deep learning based person re-identification. Acta Automat. Sin. 45, 11 (2019), 20322049. DOI: DOI: https://doi.org/10.16383/j.aas.c180154Google ScholarGoogle Scholar
  13. [13] Hao Yi, Wang Nannan, Li Jie, and Gao Xinbo. 2019. HSME: Hypersphere manifold embedding for visible thermal person re-identification. Proc. AAAI Conf. Artif. Intell. 33 (2019), 83858392. DOI: DOI: https://doi.org/10.1609/aaai.v33i01.33018385 Google ScholarGoogle ScholarCross RefCross Ref
  14. [14] Hermans Alexander, Beyer Lucas, and Leibe Bastian. 2017. In defense of the triplet loss for person re-identification. CoRR abs/1703.07737 (2017). DOI: DOI: https://doi.org/1703.07737Google ScholarGoogle Scholar
  15. [15] Yan Huang, Jingsong Xu, Qiang Wu, Zhedong Zheng, Zhaoxiang Zhang, and Jian Zhang. 2018. Multi-pseudo regularized label for generated data in person re-identification. IEEE Transactions on Image Processing PP (2018), 1–1. https://doi.org/10.1109/TIP.2018.2874715Google ScholarGoogle Scholar
  16. [16] Jia Mengxi, Zhai Yunpeng, Lu Shijian, Ma Siwei, and Zhang Jian. 2020. A similarity inference metric for RGB-infrared cross-modality person re-identification. In Proceedings of the International Joint Conferences on Artificial Intelligence Organization. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. [17] Li Diangang, Wei Xing, Hong Xiaopeng, and Gong Yihong. 2020. Infrared-visible cross-modal person re-identification with an X modality. Proc. AAAI Conf. Artif. Intell. 34, 4 (2020), 46104617.Google ScholarGoogle Scholar
  18. [18] Shengcai Liao, Yang Hu, Xiangyu Zhu, and Stan Z. Li. 2015. Person re-identification by Local Maximal Occurrence representation and metric learning. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR’15). IEEE.Google ScholarGoogle Scholar
  19. [19] Liu Haijun, Cheng Jian, Wang Wen, Su Yanzhou, and Bai Haiwei. 2020. Enhancing the discriminative feature learning for visible-thermal cross-modality person re-identification. Neurocomputing 398 (2020), 1119.Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Liu Hao, Feng Jiashi, Qi Meibin, Jiang Jianguo, and Yan Shuicheng. 2017. End-to-end comparative attention networks for person re-identification. IEEE Trans. Image Process. 26, 99 (2017), 34923506.Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. [21] Lu Yan, Wu Yue, Liu Bin, Zhang Tianzhu, Li Baopu, Chu Qi, and Yu Nenghai. 2020. Cross-modality person re-identification with shared-specific feature transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR’20). IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Ye Mang, Shen Jianbing, Lin Gaojie, Xiang Tao, Shao Ling, and Steven C. H. Hoi. 2021. Deep learning for person re-identification: a survey and outlook. IEEE Transactions on Pattern Analysis and Machine Intelligence PP, 99 (2021), 1–1.Google ScholarGoogle Scholar
  23. [23] Ye Mang, Lan Xiangyuan, and Leng Qingming. 2019. Modality-aware Collaborative Learning for Visible Thermal Person Re-Identification. In 27th ACM International Conference. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. [24] Xuelin Qian, Yanwei Fu, Wenxuan Wang, Tao Xiang, Yang Wu, Yu-Gang Jiang, and Xiangyang Xue. 2018. Pose-normalized image generation for person re-identification. In Proceedings of the 15th European Conference, Munich, Germany. Springer, Cham.Google ScholarGoogle Scholar
  25. [25] Ergys Ristani and Carlo Tomasi. 2018. Features for multi-target multi-camera tracking and re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 6036–6046.Google ScholarGoogle Scholar
  26. [26] Hailin Shi, Yang Yang, Xiangyu Zhu, Shengcai Liao, Zhen Lei, and Stan Z. Li. 2016. Embedding deep metric for person re-identification: A study against large variations. In European Conference on Computer Vision. Springer, Cham, 732–748.Google ScholarGoogle Scholar
  27. [27] Sohn Kihyuk. 2016. Improved deep metric learning with multi-class n-pair loss objective. In Advances in Neural Information Processing Systems, Lee D., Sugiyama M., Luxburg U., Guyon I., and Garnett R. (Eds.), Vol. 29. Curran Associates, Inc., 18571865. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. [28] Varior Rahul Rama, Haloi Mrinal, and Wang Gang. 2016. Gated siamese convolutional neural network architecture for human re-identification. In Proceedings of the European Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Varior Rahul Rama, Shuai Bing, Lu Jiwen, Xu Dong, and Wang Gang. 2016. A siamese long short-term memory architecture for human re-identification. In Proceedings of the European Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Wang Guan’An, Zhang Tianzhu, Yang Yang, Cheng Jian, Chang Jianlong, Liang Xu, and Hou Zengguang. 2020. Cross-modality paired-images generation for RGB-infrared person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence 34, 7 (2020), 12144–12151.Google ScholarGoogle Scholar
  31. [31] Wang Yicheng, Chen Zhenzhong, Wu Feng, and Wang Gang. 2018. Person re-identification with cascaded pairwise convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Reconigtion.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Longhui Wei, Shiliang Zhang, Wen Gao, and Qi Tian. 2018. Person transfer gan to bridge domain gap for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 79–88.Google ScholarGoogle Scholar
  33. [33] Wu Ancong, Zheng Wei Shi, Yu Hong Xing, Gong Shaogang, and Lai Jianhuang. 2017. RGB-infrared cross-modality person re-identification. In Proceedings of the International Conference on Computer Vision (ICCV’17). IEEE, Los Alamitos, CA.Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Ye Mang, Lan Xiangyuan, Li Jiawei, and Yuen Pong. 2018. Hierarchical discriminative learning for visible thermal person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence. AAAI.Google ScholarGoogle ScholarCross RefCross Ref
  35. [35] Ye Mang, Wang Zheng, Lan Xiangyuan, and Yuen Pong C.. 2018. Visible thermal person re-identification via dual-constrained top-ranking. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI’18). Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. [36] Shizhou Zhang, Yifei Yang, Peng Wang, Xiuwei Zhang, and Yanning Zhang. 2021. Attend to the difference: Cross-modality person re-identification via contrastive correlation. IEEE Transactions on Image Processing 30 (2021), 8861–8872.Google ScholarGoogle Scholar
  37. [37] Zhao Yun Bo, Lin Jian Wu, Xuan Qi, and Xi Xugang. 2020. HPILN: A feature learning framework for cross-modality person re-identification. IET Image Process. 13, 14 (2020), 28972904. DOI: DOI: https://doi.org/10.1049/iet-ipr.2019.0699Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Zheng Liang, Yang Yi, and Hauptmann Alexander G.. 2016. Person re-identification: Past, present and future.Google ScholarGoogle Scholar
  39. [39] Zhixiang Wang, Zheng Wang, Yinqiang Zheng, Yu Chuang Yung, and Shin’Ich Satoh. 2019. Learning to reduce dual-level discrepancy for infrared-visible person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR’19).Google ScholarGoogle Scholar
  40. [40] Zhong Zhun, Zheng Liang, Zheng Zhedong, Li Shaozi, and Yang Yi. 2018. Camera style adaptation for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Hybrid Modality Metric Learning for Visible-Infrared Person Re-Identification

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Multimedia Computing, Communications, and Applications
      ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 18, Issue 1s
      February 2022
      352 pages
      ISSN:1551-6857
      EISSN:1551-6865
      DOI:10.1145/3505206
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 25 January 2022
      • Accepted: 1 June 2021
      • Revised: 1 May 2021
      • Received: 1 January 2020
      Published in tomm Volume 18, Issue 1s

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Refereed

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text

    HTML Format

    View this article in HTML Format .

    View HTML Format