Skip to main content

M3LH: Multi-modal Multi-label Hashing for Large Scale Data Search

  • Conference paper
  • First Online:
  • 3361 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 10132))

Abstract

Recently, hashing based technique has attracted much attention in media search community. In many applications, data have multiple modalities and multiple labels. Many hashing methods have been proposed for multi-modal data; however, they seldom consider the scenario of multiple labels or only use such information to build a simple similarity matrix, e.g., the corresponding value is 1 when two samples share at least one same label. Apparently, such methods cannot make full use of the information contained in multiple labels. Thus, a model is expected to have good performance if it can make full use of information in multi-modal and multi-label data. Motivated by this, in this paper, we propose a new method, multi-modal multi-label hashing-M3LH, which can not only work on multi-modal data, but also make full use of information contained in multiple labels. Specifically, in M3LH, we assume every label is associated with a binary code in Hamming space, and the binary code of a sample can be generated by combining the binary codes of its labels. While minimizing the Hamming distance between similar pairs and maximizing the Hamming distance between dissimilar pairs, we also learn a project matrix which can be used to generate binary codes for out-of-samples. Experimental results on three widely used data sets show that M3LH outperforms or is comparable to several state-of-the-art hashing methods.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. http://press.liacs.nl/mirflickr/

  2. Chua, T.S., Tang, J., Hong, R., Li, H., Luo, Z., Zheng, Y.: NUS-WIDE: a real-world web image database from National University of Singapore. In: Proceedings of CIVR, article no. 48 (2009)

    Google Scholar 

  3. Datar, M., Immorlica, N., Indyk, P., Mirrokni, V.S.: Locality-sensitive hashing scheme based on p-stable distributions. In: Proceedings of SCG, pp. 253–262 (2004)

    Google Scholar 

  4. Ding, G., Guo, Y., Zhou, J.: Collective matrix factorization hashing for multimodal data. In: Proceedings of CVPR, pp. 2075–2082 (2014)

    Google Scholar 

  5. Ding, K., Huo, C., Fan, B., Pan, C.: kNN hashing with factorized neighborhood representation. In: Proceedings of ICCV, pp. 1098–1106 (2015)

    Google Scholar 

  6. Gong, Y., Lazebnik, S.: Iterative quantization: a procrustean approach to learning binary codes. In: Proceedings of CVPR, pp. 817–824 (2011)

    Google Scholar 

  7. He, K., Wen, F., Sun, J.: K-means hashing: an affinity-preserving quantization method for learning binary compact codes. In: Proceedings of CVPR, pp. 2938–2945 (2013)

    Google Scholar 

  8. Kim, S., Choi, S.: Multi-view anchor graph hashing. In: Proceedings of ICASSP, pp. 3123–3127 (2013)

    Google Scholar 

  9. Kumar, S., Udupa, R.: Learning hash functions for cross-view similarity search. In: IJCAI Proceedings-International Joint Conference on Artificial Intelligence, vol. 22, p. 1360 (2011)

    Google Scholar 

  10. Lin, Z., Ding, G., Hu, M., Wang, J.: Semantics-preserving hashing for cross-view retrieval. In: Proceedings of CVPR, pp. 3864–3872 (2015)

    Google Scholar 

  11. Rasiwasia, N., Costa Pereira, J., Coviello, E., Doyle, G., Lanckriet, G.R., Levy, R., Vasconcelos, N.: A new approach to cross-modal multimedia retrieval. In: Proceedings of MM, pp. 251–260 (2010)

    Google Scholar 

  12. Song, J., Yang, Y., Yang, Y., Huang, Z., Shen, H.T.: Inter-media hashing for large-scale retrieval from heterogeneous data sources. In: Proceedings of SIGMOD, pp. 785–796 (2013)

    Google Scholar 

  13. Tang, J., Li, Z., Wang, M., Zhao, R.: Neighborhood discriminant hashing for large-scale image retrieval. IEEE Trans. Image Process. 24(9), 2827–2840 (2015)

    Article  MathSciNet  Google Scholar 

  14. Wang, J., Kumar, S., Chang, S.F.: Sequential projection learning for hashing with compact codes. In: Proceedings of ICML, pp. 1127–1134 (2010)

    Google Scholar 

  15. Wang, J., Xu, X.S., Guo, S., Cui, L., Wang, X.: Linear unsupervised hashing for ANN search in Euclidean space. Neurocomputing 171(c), 283–292 (2016)

    Article  Google Scholar 

  16. Wang, S.S., Huang, Z., Xu, X.S.: A multi-label least-squares hashing for scalable image search. In: Proceedings of SDM, pp. 954–962 (2015)

    Google Scholar 

  17. Weinberger, K.Q., Saul, L.K.: Distance metric learning for large margin nearest neighbor classification. J. Mach. Learn. Res. 10, 207–244 (2009)

    MATH  Google Scholar 

  18. Weiss, Y., Torralba, A., Fergus, R.: Spectral hashing. In: NIPS, pp. 1753–1760 (2009)

    Google Scholar 

  19. Wu, B., Yang, Q., Zheng, W.-S., Wang, Y., Wang, J.: Quantized correlation hashing for fast cross-modal search. In: Proceedings of IJCAI, pp. 3946–3952 (2015)

    Google Scholar 

  20. Xu, X.-S.: Dictionary learning based hashing for cross-modal retrieval. In: Proceedings of MM, pp. 177–181 (2016)

    Google Scholar 

  21. Yan, T.-K., Xu, X.-S., Guo, S., Huang, Z., Wang, X.-L.: Supervised robust discrete multimodal hashing for cross-media retrieval. In: Proceedings of CIKM, pp. 1271–1280 (2016)

    Google Scholar 

  22. Yang, Y., Shen, F., Shen, H.T., Li, H., Li, X.: Robust discrete spectral hashing for large-scale image semantic indexing. IEEE Trans. Big Data 1(4), 162–171 (2015)

    Article  Google Scholar 

  23. Yang, Y., Zha, Z.-J., Gao, Y., Zhu, X., Chua, T.-S.: Exploiting web images for robust semantic video indexing via sample-specific loss. IEEE Trans. Multimed. 16(6), 1677–1689 (2014)

    Article  Google Scholar 

  24. Zhang, D., Li, W.J.: Large-scale supervised multimodal hashing with semantic correlation maximization. In: Proceedings of AAAI, pp. 2177–2183 (2014)

    Google Scholar 

  25. Zhou, J., Ding, G., Guo, Y.: Latent semantic sparse hashing for cross-modal similarity search. In: Proceedings of SIGIR, pp. 415–424 (2014)

    Google Scholar 

  26. Zou, F., Liu, C., Ling, H., Feng, H., Yan, L., Li, D.: Least square regularized spectral hashing for similarity search. Signal Process. 93(8), 2265–2273 (2013)

    Article  Google Scholar 

Download references

Acknowledgements

This work was partially supported by National Natural Science Foundation of China (61173068, 61573212, 91546203), Program for New Century Excellent Talents in University of the Ministry of Education, Independent Innovation Foundation of Shandong Province (2014CGZH1106), Key Research and Development Program of Shandong Province (2016GGX101044, 2015GGE27033).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xin-Shun Xu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Yang, GQ., Xu, XS., Guo, S., Wang, XL. (2017). M3LH: Multi-modal Multi-label Hashing for Large Scale Data Search. In: Amsaleg, L., Guðmundsson, G., Gurrin, C., Jónsson, B., Satoh, S. (eds) MultiMedia Modeling. MMM 2017. Lecture Notes in Computer Science(), vol 10132. Springer, Cham. https://doi.org/10.1007/978-3-319-51811-4_17

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-51811-4_17

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-51810-7

  • Online ISBN: 978-3-319-51811-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics