ABSTRACT
With the rapid growth of multimedia data, it is very desirable to effectively and efficiently search objects of interest across different modalities from large scale databases. Cross-modal hashing provides a very promising way to address such problem. In this paper, we propose a two-step cross-modal hashing approach to obtain compact hash codes and learn hash functions from multimodal data. Our approach decomposes the cross-modal hashing problem into two steps: generating hash code and learning hash function. In the first step, we obtain the hash codes for all modalities of data via a joint multi-modal graph, which takes into consideration both the intra-modality and inter-modality similarity. In the second step, learning hashing function is formulated as a binary classification problem. We train binary classifiers to predict the hash code for any data object unseen before. Experimental results on two cross-modal datasets show the effectiveness of our proposed approach.
- M. M. Bronstein, A. M. Bronstein, F. Michel, and N. Paragios. Data fusion through cross-modality metric learning using similarity-sensitive hashing. In CVPR, pages 3594--3601, 2010.Google ScholarCross Ref
- T.-S. Chua, J. Tang, R. Hong, H. Li, Z. Luo, , and Y.-T. Zheng. NUS-WIDE: A real-world web image database from national university of singapore. In ACM International Conference on Image and Video Retrieval, 2009. Google ScholarDigital Library
- R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin. LIBLINEAR: A library for large linear classification. JMLR, 9:1871--1874, 2008. Google ScholarDigital Library
- Y. Gong and S. Lazebnik. Iterative quantization: A procrustean approach to learning binary codes. In CVPR, pages 817--824, 2011. Google ScholarDigital Library
- S. Kumar and R. Udupa. Learning hash functions for cross-view similarity search. In IJCAI, pages 1360--1365, 2011. Google ScholarDigital Library
- N. Quadrianto and C. H. Lampert. Learning multi-view neighborhood preserving projections. In ICML, pages 425--432, 2011.Google Scholar
- N. Rasiwasia, P. J. Moreno, and N. Vasconcelos. Bridging the gap: query by semantic example. IEEE TMM, 9(5):923--938, 2007. Google ScholarDigital Library
- N. Rasiwasia, J. C. Pereira, E. Coviello, G. Doyle, G. Lanckriet, R. Levy, and N. Vasconcelos. A new approach to cross-modal multimedia retrieval. In ACM MM, pages 251--260, 2010. Google ScholarDigital Library
- M. Rastegari, J. Choi, S. Fakhraei, H. D. III, and L. S. Davis. Predictable dual-view hashing. In ICML, 2013.Google ScholarDigital Library
- J. Song, Y. Yang, Z. Huang, H. Shen, and R. Hong. Multiple feature hashing for real-time large scale near-duplicate video retrieval. In ACM MM, pages 423--432, 2011. Google ScholarDigital Library
- Y. Weiss, A. Torralba, and R. Fergus. Spectral hashing. In NIPS, pages 1753--1760, 2008.Google ScholarDigital Library
- D. Zhang, F. Wang, and L. Si. Composite hashing with multiple information sources. In ACM SIGIR, pages 225--234, 2011. Google ScholarDigital Library
- D. Zhang, J. Wang, D. Cai, and J. S. Liu. Self-taught hashing for fast similarity search. In SIGIR, pages 18--25, 2010. Google ScholarDigital Library
- Y. Zhen and D.-Y. Yeung. A probabilistic model for multimodal hash function learning. In SIGKDD, pages 940--948, 2012. Google ScholarDigital Library
Index Terms
- A Two-step Approach to Cross-modal Hashing
Recommendations
Discriminant Cross-modal Hashing
ICMR '16: Proceedings of the 2016 ACM on International Conference on Multimedia RetrievalHashing based methods have attracted considerable attention for efficient cross-modal retrieval on large-scale multimedia data. The core problem of cross-modal hashing is how to effectively integrate heterogeneous features from different modalities to ...
A Label Noise Robust Cross-Modal Hashing Approach
Knowledge Science, Engineering and ManagementAbstractCross-modal hashing has attracted more and more research interest for its high speed and low storage cost in solving cross-modal approximate nearest neighbor search problem. With the rapid growth of social networks, a large amount of information ...
Deep fused two-step cross-modal hashing with multiple semantic supervision
AbstractExisting cross-modal hashing methods ignore the informative multimodal joint information and cannot fully exploit the semantic labels. In this paper, we propose a deep fused two-step cross-modal hashing (DFTH) framework with multiple semantic ...
Comments