Processing math: 33%
Semantics-Aware Spatial-Temporal Binaries for Cross-Modal Video Retrieval | IEEE Journals & Magazine | IEEE Xplore

Semantics-Aware Spatial-Temporal Binaries for Cross-Modal Video Retrieval


Abstract:

With the current exponential growth of video-based social networks, video retrieval using natural language is receiving ever-increasing attention. Most existing approache...Show More

Abstract:

With the current exponential growth of video-based social networks, video retrieval using natural language is receiving ever-increasing attention. Most existing approaches tackle this task by extracting individual frame-level spatial features to represent the whole video, while ignoring visual pattern consistencies and intrinsic temporal relationships across different frames. Furthermore, the semantic correspondence between natural language queries and person-centric actions in videos has not been fully explored. To address these problems, we propose a novel binary representation learning framework, named Semantics-aware Spatial-temporal Binaries (\text{S}^{2} Bin), which simultaneously considers spatial-temporal context and semantic relationships for cross-modal video retrieval. By exploiting the semantic relationships between two modalities, \text{S}^{2} Bin can efficiently and effectively generate binary codes for both videos and texts. In addition, we adopt an iterative optimization scheme to learn deep encoding functions with attribute-guided stochastic training. We evaluate our model on three video datasets and the experimental results demonstrate that \text{S}^{2} Bin outperforms the state-of-the-art methods in terms of various cross-modal video retrieval tasks.
Published in: IEEE Transactions on Image Processing ( Volume: 30)
Page(s): 2989 - 3004
Date of Publication: 09 February 2021

ISSN Information:

PubMed ID: 33560984

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.