skip to main content
10.1145/3573428.3573775acmotherconferencesArticle/Chapter ViewAbstractPublication PageseitceConference Proceedingsconference-collections
research-article

Multi-loss joint cross-modal pedestrian re-identification method fused with grayscale and RGA

Published: 15 March 2023 Publication History

Abstract

Cross-modal pedestrian re-identification is to match the color image and infrared image of the pedestrian to determine whether there is the same pedestrian. The images have large inter-modal differences and large intra-modal variations, resulting in low recognition accuracy. This paper proposes a multi-loss joint cross-modal pedestrian re-identification method that fuses grayscale images and relation-aware global attention (RGA). First, grayscale the color images in the cross-modal dataset to reduce the network's dependence on color information; second, use the trained generative adversarial network to convert the color images and infrared images in the dataset to each other to reduce pedestrian poses and modality changes; then, the RGA is embedded into a shared-weight dual-stream ResNet50 to capture more robust features; finally, the hard sample triplet loss is improved and converted into a cross-modal hard sample triplet loss. The cross-entropy loss of the smooth label is combined with the hard sample triplet loss and the cross-modal hard sample triplet loss, respectively, to form joint functions L1 and L2, and then supervised training of the network. Experiments are carried out on the RegDB and SYSU-MM01 datasets, and the mAP reaches 75.85% and 69.56%, respectively, which are better than many current methods, indicating that the proposed method has better recognition accuracy.

References

[1]
Wu A, Zheng W S, Yu H X, RGB-infrared cross-modality person re-identification [C]// Proceedings of the IEEE international conference on computer vision. 2017: 5380-5389.
[2]
Ye M, Lan X, Li J, Hierarchical discriminative learning for visible thermal person re-identification [C]// Proceedings of the AAAI Conference on Artificial Intelligence. 2018, 32(1).
[3]
Du Peng, Song Yonghong, Zhang Xinyao. Research on cross-modal pedestrian re-recognition method based on self-attention modal fusion network [J/OL]. Acta Automatica Sinica: 1-12, 2020, 09, 20. https://doi. org/10.16383/j.aas.c190340. (in Chinese)
[4]
Zhu J Y, Park T, Isola P, Unpaired image-to-image translation using cycle-consistent adversarial networks [C]// Proceedings of the IEEE international conference on computer vision. 2017: 2223-2232.
[5]
Dai P, Ji R, Wang H, Cross-modality person re-identification with generative adversarial training [C]// IJCAI. 2018, 1(3): 6.
[6]
Ye M, Lan X, Wang Z, Bi-directional center-constrained top-ranking for visible thermal person re-identification [J]. IEEE Transactions on Information Forensics and Security, 2019, 15: 407-419.
[7]
Feng Z, Lai J, Xie X. Learning modality-specific representations for visible-infrared person re-identification [J]. IEEE Transactions on Image Processing, 2019, 29: 579-590.
[8]
Hao Y, Wang N, Li J, HSME: Hypersphere manifold embedding for visible thermal person re-identification [C]// Proceedings of the AAAI conference on artificial intelligence. 2019, 33(01): 8385-8392.
[9]
Lin J W, Li H. HPILN: A feature learning framework for cross-modality person re-identification [J]. arXiv preprint arXiv: 1906.03142, 2019.
[10]
Wang G, Zhang T, Cheng J, Rgb-infrared cross-modality person re-identification via joint pixel and feature alignment [C]// Proceedings of the IEEE/CVF International Conference on Computer Vision. 2019: 3623-3632.
[11]
Li D, Wei X, Hong X, Infrared-visible cross-modal person re-identification with an x modality[C]//Proceedings of the AAAI Conference on Artificial Intelligence. 2020, 34(04): 4610-4617.
[12]
LI T W, ZENG Z Y. Cross-modal person re-identification model based on dynamic dual-attention mechanism [J]. Journal of Computer Applications, 2022: 0. LI Dawei, ZENG Zhiyong. Cross-modal person re-identification model based on dynamic dual-attention mechanism [J/OL]. Journal of Computer Applications: 1-10, 2022, 06, 07. (in Chinese)
[13]
Ye M, Shen J, Shao L. Visible-infrared person re-identification via homogeneous augmented tri-modal learning [J]. IEEE Transactions on Information Forensics and Security, 2020, 16: 728-739.
[14]
Lu Y, Wu Y, Liu B, Cross-modality person re-identification with shared-specific feature transfer[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020: 13379-13389.
[15]
Chen K F, Pan Z S, Wang J B, Collaborative Learning Method for Cross Modality Person Re-identification [J]. Computer Engineering and Applications. 2021, 57 (12): 115-125.
[16]
Huang P, Zhu S H, Liang Z W. Cross-modal person re-identification with triple- attention feature aggregation [J]. Journal of Nanjing University of Posts and Telecommuncations (Natural Science Edition), 2021, 41 (05): 101-112.
[17]
Wei Z Y, Ynag X, Wang N N, Reciprocal bi-directional generative adversarial networkfor cross-modal pedestrian re-identification [J]. JOURNAL OF XIDIAN UNIVERSITY.,2021, 48(02):205-212.
[18]
Cheng D, Hao Y, Zhou J Y, Cross-modality person re-identification utility utility thehybrid two-stream neural networks[J]. JOURNAL OF XIDIAN UNIVERSITY, 2021,48(05):190-200.
[19]
Zhang Z, Lan C, Zeng W, Relation-aware global attention for person re-identification[C]//Proceedings of the ieee/cvf conference on computer vision and pattern recognition. 2020: 3186-3195.
[20]
Chen R, Huang W, Huang B, Reusing discriminators for encoding: Towards unsupervised image-to-image translation[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2020: 8168-8177.
[21]
Nguyen D T, Hong H G, Kim K W, Person recognition system based on a combination of body images from visible light and thermal cameras [J]. Sensors, 2017, 17(3): 605.
[22]
Basaran E, Gökmen M, Kamasak M E. An efficient framework for visible–infrared cross modality person re-identification[J]. Signal Processing: Image Communication, 2020, 87: 115933.
[23]
Zhu Y, Yang Z, Wang L, Hetero-center loss for cross-modality person re-identification[J]. Neurocomputing, 2020, 386: 97-109.
[24]
Ye M, Shen J, J Crandall D, Dynamic dual-attentive aggregation learning for visible-infrared person re-identification[C]//European Conference on Computer Vision. Springer, Cham, 2020: 229-247.
[25]
Zhang Y, Xiang Xu, Tang Jun, Cross-modality person re-identification algorithm using symmetric network[J]. Journal of National University of Defense Technology, 2022, 44 (1): 122-128(in Chinese)
[26]
Ye M, Lan X, Leng Q. Modality-aware collaborative learning for visible thermal person re-identification [C]// Proceedings of the 27th ACM International Conference on Multimedia. 2019: 347-355.
[27]
Liu H, Cheng J, Wang W, Enhancing the discriminative feature learning for visible-thermal cross-modality person re-identification [J]. Neurocomputing, 2020, 398: 11-19.
[28]
Hao Y, Wang N, Gao X, Dual-alignment feature embedding for cross-modality person re-identification [C]// Proceedings of the 27th ACM International Conference on Multimedia. 2019: 57-65.
[29]
Ye M, Shen J, Lin G, Deep learning for person re-identification: A survey and outlook[J]. IEEE transactions on pattern analysis and machine intelligence, 2021, 44 (6): 2872-2893.

Index Terms

  1. Multi-loss joint cross-modal pedestrian re-identification method fused with grayscale and RGA

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    EITCE '22: Proceedings of the 2022 6th International Conference on Electronic Information Technology and Computer Engineering
    October 2022
    1999 pages
    ISBN:9781450397148
    DOI:10.1145/3573428
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 15 March 2023

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Cross-modal person re-identification
    2. Multi-loss joint function
    3. RGA
    4. generative adversarial networks
    5. grayscale

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    EITCE 2022

    Acceptance Rates

    Overall Acceptance Rate 508 of 972 submissions, 52%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 28
      Total Downloads
    • Downloads (Last 12 months)12
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 28 Feb 2025

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media