Loading [a11y]/accessibility-menu.js
Cross Modal Person Re-identification with Visual-Textual Queries | IEEE Conference Publication | IEEE Xplore

Cross Modal Person Re-identification with Visual-Textual Queries


Abstract:

Classical person re-identification approaches assume that a person of interest has appeared across different cameras and can be queried by one of the existing images. How...Show More

Abstract:

Classical person re-identification approaches assume that a person of interest has appeared across different cameras and can be queried by one of the existing images. However, in real-world surveillance scenarios, frequently no visual information will be available about the queried person. In such scenarios, a natural language description of the person by a witness will provide the only source of information for retrieval. In this work, person re-identification using both vision and language information is addressed under all possible gallery and query scenarios. A two stream deep convolutional neural network framework supervised by identity based cross entropy loss is presented. Canonical Correlation Analysis is performed to enhance the correlation between the two modalities in a joint latent embedding space. To investigate the benefits of the proposed approach, a new testing protocol under a multi modal ReID setting is proposed for the test split of the CUHK-PEDES and CUHK-SYSU benchmarks. The experimental results verify that the learnt visual representations are more robust and perform 20% better during retrieval as compared to a single modality system.
Date of Conference: 28 September 2020 - 01 October 2020
Date Added to IEEE Xplore: 06 January 2021
ISBN Information:

ISSN Information:

Conference Location: Houston, TX, USA

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.