Skip to main content

Critical Separation Hashing for Cross-Modal Retrieval

  • Conference paper
  • First Online:
6GN for Future Wireless Networks (6GN 2022)

Abstract

With the development of Internet technology, unimodal retrieval techniques are no longer suitable for the current environment, and mutual retrieval between multiple modalities is needed to obtain more complete information. Deep hashing has clearly become a simpler and faster method in cross-modal hashing. In recent years, unsupervised cross-modal hashing has received increasing attention. However, existing methods fail to exploit the common information across modalities, thus resulting in information wastage. In this paper, we propose a new critical separation cross-modal hashing (CSCH) for unsupervised cross-modal retrieval, which explores the similarity information across modalities by highlighting the similarity between instances to help the network learn the hash function, and we carefully design the loss function by introducing the likelihood loss commonly used in supervised learning into the loss function. Extensive experiments on two cross-modal retrieval datasets show that CSCH has better performance.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Lin, Z., Ding, G., Hu, M., Wang, J.: Semantics preserving hashing for cross-view retrieval. In: CVPR, pp. 3864–3872 (2015)

    Google Scholar 

  2. Zhou, J., Ding, G., Guo, Y.: Latent semantic sparse hashing for cross-modal similarity search. In: Proceedings of the 37th international ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 415–424. ACM, 2014

    Google Scholar 

  3. Gu, W., Gu, X., Gu, J., Li, B., Xiong, Z., Wang, W.: Adversary guided asymmetric hashing for cross-modal retrieval. In: ICMR, pp. 159–167. ACM (2019)

    Google Scholar 

  4. Jiang, Q.Y., Li, W.J.: Deep cross-modal hashing. In: IEEE CVPR, pp. 3232–3240 (2017)

    Google Scholar 

  5. Cao, Z., Long, M., Yang, Q.: Transitive hashing network for heterogeneous multimedia retrieval. CoRR,2016,abs/1608.04307

    Google Scholar 

  6. Li, C., Deng, C., Li, N., Liu, W., Gao, X., Tao, D.: Self-supervised adversarial hashing networks for cross-modal retrieval. In: Proceedings of IEEE CVPR, pp. 4242–4251 (2018)

    Google Scholar 

  7. Jiang, Q., Li, W.: Discrete latent factor model for cross-modal hashing. In: IEEE Transactions on Image Processing, pp. 3490–3501 (2019)

    Google Scholar 

  8. Zhou, X., et al.: Graph convolutional network hashing. IEEE Trans. Cybern. 1460–1472 (2020)

    Google Scholar 

  9. Kumar, S., Udupa, R.: Learning hash functions for cross-view similarity search. In: Twenty-Second International Joint Conference on Artificial Intelligence (2011)

    Google Scholar 

  10. Zhou, J., Ding, G., Guo, Y.: Latent semantic sparse hashing for cross-modal similarity search. In: Proceedings of the 37th International ACM SIGIR Conference on Research Development in Information Retrieval, pp. 415–424 (2014)

    Google Scholar 

  11. Ding, G., Guo, Y., Zhou, J.: Collective matrix factorization hashing for multimodal data. In: Proceedings of the IEEE Conference on Computer Vision And Pattern Recognition, pp. 2075–2082 (2014)

    Google Scholar 

  12. Zhang, J., Peng, Y., Yuan, M.: Unsupervised generative adversarial cross-modal hashing. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)

    Google Scholar 

  13. Su, S., Zhong, Z., Zhang, C.: Deep joint-semantics reconstructing hashing for large-scale unsupervised cross-modal retrieval. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3027–3035 (2019)

    Google Scholar 

  14. Yang, D., Wu, D., Zhang, W., Zhang, H., Li, B., Wang, W.: Deep semantic-alignment hashing for unsupervised cross-modal retrieval. In: Proceedings of the 2020 International Conference on Multimedia Retrieval, pp. 44–52 (2020)

    Google Scholar 

  15. Liu, S., Qian, S., Guan, Y., Zhan, J., Ying, L.: Joint-modal distribution-based similarity hashing for large-scale unsupervised deep cross-modal retrieval. In: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1379–1388 (2020)

    Google Scholar 

  16. Chua, T.-S., Tang, J., Hong, R., Li, H., Luo, Z., Zheng, Y.: NUS-WIDE: a real-world web image database from National University of Singapore. In: Proceedings of the ACM International Conference on Image and Video Retrieval, pp. 1–9 (2009)

    Google Scholar 

  17. Huiskes, M.J., Lew, M.J.: The MIR flickr retrieval evaluation. In: Multimedia Information Retrieval. ACM, pp. 39–43 (2008)

    Google Scholar 

Download references

Acknowledgement

This work was supported in part by the National Natural Science Foundation of China under Grant 62071157, National Key Research and Development Programme 2022YFD2000500 and Natural Science Foundation of Heilongjiang Province under Grant YQ2019F011.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zening Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Wang, Z., Sun, Y., Liu, L., Li, A. (2023). Critical Separation Hashing for Cross-Modal Retrieval. In: Li, A., Shi, Y., Xi, L. (eds) 6GN for Future Wireless Networks. 6GN 2022. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 504. Springer, Cham. https://doi.org/10.1007/978-3-031-36011-4_15

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-36011-4_15

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-36010-7

  • Online ISBN: 978-3-031-36011-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics