Abstract
In image retrieval, deep-learning-based models combing deep hashing and Bayesian learning have become one of the mainstream approaches. The choice of likelihood functions can significantly affect the performance of existing image retrieval methods that combine deep hashing and Bayesian learning, resulting in issues such as misclassification in single-label datasets and biased label association in multi-label ones. However, it remains to further explore how image retrieval performance can be reliably enhanced through proper likelihood function design. In this paper, we propose a deep adaptive-mapping-based hashing (DAMH) method that enhances image retrieval performance via adjustable likelihood function design. Through strategically re-mapping image samples with low-gradient to high-gradient regions of the likelihood functions, our method both effectively expands the ranges over which inner products used in single-label image retrieval are trained and properly delimits the likelihood functions to prevent multi-label images from being excessively mapped into Hamming sphere(s) of any single class. Furthermore, we design a batch-by-batch optimization method that treats easy and hard samples differently, preventing the gradients of hard samples from being submerged by those of the easy ones during the training process. Our experiments on general-purpose image datasets, including CIFAR10, NUSWIDE and ImageNet100, show that DAMH excels existing peer methods in overall image retrieval performance.


















Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The datasets and model parameter settings generated and/or analyzed during the current study are recorded, documented and made available in our DAMH project repository at Github: https://github.com/q878787/DAMH. In addition to our version of the NUSWIDE and ImageNet100 datasets available at https://github.com/q878787/DAMH/tree/main/data, the original version of the datasets utilized in our current study, which are all publicly available, can be accessed via the following links: 1. CIFAR10 [47] dataset: https://www.cs.toronto.edu/~kriz/cifar.html. 2. NUSWIDE [48] dataset: https://lms.comp.nus.edu.sg/wp-content/uploads/2019/research/nuswide/NUS-WIDE.html. 3. ImageNet100 [49] dataset: https://www.image-net.org/.
Notes
Our source code can be found at https://github.com/q878787/DAMH.
References
Luo X, Chen C, Zhong H, Zhang H, Deng M, Huang J, Hua X (2020) A survey on deep hashing methods. arXiv preprint arXiv:2003.03369
Hecht-Nielsen R (1992) Theory of the backpropagation neural network. Neural networks for perception. Elsevier, Amsterdam, Netherlands, pp 65–93
Neal Radford M (2012) Bayesian learning for neural networks, vol 118. Springer, Cham
Tianfeng C, Draxler Roland R (2014) Root mean square error (RMSE) or mean absolute error (MAE)?-arguments against avoiding RMSE in the literature. Geosci Model Dev 7(3):1247–1250
Kalyan D, Jiming J, Rao JNK (2004) Mean squared error of empirical predictor. Ann Stat 32(2):818–840
Hahn G, Lutz SM, Laha N, Lange C (2022) A framework to efficiently smooth L1 penalties for linear regression. bioRxiv, p 2020–09
Golik P, Doetsch P, Ney H (2013) Cross-entropy vs. squared error training: a theoretical and experimental comparison. Interspeech, vol 13. ISCA, Dublin, Ireland, pp 1756–1760
Kline DM, Berardi VL (2005) Revisiting squared-error and cross-entropy functions for training neural network classifiers. Neural Comput Appl 14(4):310–318
Li W-J, Wang S, Kang W-C (2015) Feature learning based deep supervised hashing with pairwise labels. arXiv preprint arXiv:1511.03855
Zhu H, Long M, Wang J, Cao Y (2016) Deep hashing network for efficient similarity retrieval. In: Proceedings of the AAAI conference on artificial intelligence, vol 30
Cao Z, Long M, Wang J, Yu PS (2017) Hashnet: deep learning to hash by continuation. In: Proceedings of the IEEE international conference on computer vision, p 5608–5617
Li Q, Sun Z, He R, Tan T (2017) Deep supervised discrete hashing. arXiv preprint arXiv:1705.10999
Kang R, Cao Y, Long M, Wang J, Yu PS (2019) Maximum-margin hamming hashing. In: Proceedings of the IEEE/CVF international conference on computer vision, p 8251–8260
Cao Z, Sun Z, Long M, Wang J, Yu PS (2018) Deep priority hashing. In: Proceedings of the 26th ACM international conference on multimedia, p 1653–1661
Yang E, Yao D, Cao B, Guan H, Yap P-T, Shen D, Liu M (2020) Deep disentangled hashing with momentum triplets for neuroimage search. International conference on medical image computing and computer-assisted intervention. Springer, Cham, pp 191–201
Cao Y, Long M, Liu B, Wang J (2018) Deep cauchy hashing for hamming space retrieval. In: Proceedings of the IEEE conference on computer vision and pattern recognition, p 1229–1237
Yuan L, Wang T, Zhang X, Tay FEH, Jie Z, Liu W, Feng J (2020) Central similarity quantization for efficient image and video retrieval. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, p 3080–3089
Weiss Y, Torralba A, Fergus R et al (2008) Spectral hashing. In: Proceedings of the 22nd international conference on neural information processing systems (NIPS), vol 1, p 4
Chen S, Cao L, Lin M, Wang Y, Sun X, Wu C, Qiu J, Ji R (2019) Hadamard codebook based deep hashing. arXiv preprint arXiv:1910.09182
Bengio Y, Louradour J, Collobert R, Weston J (2009) Curriculum learning. In: Proceedings of the 26th annual international conference on machine learning, p 41–48
Wang X, Han X, Huang W, Dong D, Scott MR (2019) Multi-similarity loss with general pair weighting for deep metric learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, p 5017–5025
Chen W, Liu Y, Wang W, Bakker E, Georgiou T, Fieguth P, Liu L, Lew MS (2021) Deep image retrieval: a survey. arXiv preprint arXiv:2101.11282
Noh H, Araujo A, Sim J, Weyand T, Han B (2017) Large-scale image retrieval with attentive deep local features. In: Proceedings of the IEEE international conference on computer vision, p 3456–3465
Cao B, Araujo A, Sim J (2020) Unifying deep local and global features for image search. European conference on computer vision. Springer, Cham, pp 726–743
Lai H, Pan Y, Liu Y, Yan S (2015) Simultaneous feature learning and hash coding with deep neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, p 3270–3278
Yang X, Deng C, Liu T, Tao D (2020) IEEE transactions on pattern analysis and machine intelligence. Heterogeneous graph attention network for unsupervised multiple-target domain adaptation. IEEE, New York
Wang X, Shi Y, Kitani KM (2016) Deep supervised hashing with triplet labels. Asian conference on computer vision. Springer, Cham, pp 70–84
Huang L-K, Chen J, Pan SJ (2019) Accelerate learning of deep hashing with gradient attention. In: Proceedings of the IEEE/CVF international conference on computer vision, p 5271–5280
Liu H, Wang R, Shan S, Chen X (2016) Deep supervised hashing for fast image retrieval. In: Proceedings of the IEEE conference on computer vision and pattern recognition, p 2064–2072
Ji Z, Yao W, Wei W, Song H, Pi H (2019) Deep multi-level semantic hashing for cross-modal retrieval. IEEE Access 7:23667–23674
Xia R, Pan Y, Lai H, Liu C, Yan S (2014) Supervised hashing for image retrieval via image representation learning. In: Twenty-eighth AAAI conference on artificial intelligence, p 2156–2162
Jiang Q-Y, Li W-J (2018) Asymmetric deep supervised hashing. In: Proceedings of the AAAI conference on artificial intelligence, vol 32
Zhang R, Lin L, Zhang R, Zuo W, Zhang L (2015) Bit-scalable deep hashing with regularized similarity learning for image retrieval and person re-identification. IEEE Trans Image Process 24(12):4766–4779
Zhao F, Huang Y, Wang L, Tan T (2015) Deep semantic ranking based hashing for multi-label image retrieval. In: Proceedings of the IEEE conference on computer vision and pattern recognition, p 1556–1564
Liu B, Cao Y, Long M, Wang J, Wang J (2018) Deep triplet quantization. In: Proceedings of the 26th ACM international conference on multimedia, p 755–763. Association for computing machinery
Yan X, Zhang L, Li W-J (2017) Semi-supervised deep hashing with a bipartite graph. In: Proceedings of the 26th international joint conference on artificial intelligence, p 3238–3244
Qiu Z, Pan Y, Yao T, Mei T (2017) Deep semantic hashing with generative adversarial networks. In: Proceedings of the 40th international ACM SIGIR conference on research and development in information retrieval, p 225–234
Liu X, Nie X, Yin Y (2019) Mutual linear regression-based discrete hashing. arXiv preprint arXiv:1904.00744
Li Y, Pei W, Gemert J van et al (2019) Push for quantization: deep fisher hashing. arXiv preprint arXiv:1909.00206
Li X, Mengfei X, Jiabo X, Weise T, Zou L, Sun F, Zhize W (2020) Image retrieval using a deep attention-based hash. IEEE Access 8:142229–142242
Zhang P, Zhang W, Li W-J, Guo M (2014) Supervised hashing with latent factor models. In: Proceedings of the 37th international ACM SIGIR conference on research and development in information retrieval, p 173–182
Deng C, Chen Z, Liu X, Gao X, Tao D (2018) Triplet-based deep hashing network for cross-modal retrieval. IEEE Trans Image Process 27(8):3893–3903
Yan C, Pang G, Bai X, Shen C, Zhou J, Hancock E (2019) Deep hashing by discriminating hard examples. In: Proceedings of the 27th ACM international conference on multimedia, p 1535–1542
Weisstein EW Disk line picking. https://mathworld.wolfram.com/
Su S, Zhang C, Han K, Tian Y (2018) Greedy hash: towards fast optimization for accurate hash coding in CNN. In: Proceedings of the 32nd international conference on neural information processing systems, p 806–815
Zhu H, Gao S et al (2017) Locality constrained deep supervised hashing for image retrieval. In: Proceedings of the 26th international joint conference on artificial intelligence, p 3567–3573
Krizhevsky A (2009) Learning multiple layers of features from tiny images. ON Canada, Groups at MIT and NYU, Toronto
Chua T-S, Tang J, Hong R, Li H, Luo Z, Zheng Y (2009) Nus-wide: a real-world web image database from national university of singapore. In: Proceedings of the ACM international conference on image and video retrieval, p 1–9
Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M et al (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115(3):211–252
Chatfield K, Simonyan K, Vedaldi A, Zisserman A (2014) Return of the devil in the details: delving deep into convolutional nets. arXiv preprint arXiv:1405.3531
Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097–1105
Northcutt CG, Athalye A, Mueller J (2021) Pervasive label errors in test sets destabilize machine learning benchmarks. arXiv preprint arXiv:2103.14749
Acknowledgements
This work is supported by the Guangdong Provincial Foundation for Basic and Applied Basic Research Grant No. \(2021A1515110673\) from the Department of Science and Technology of Guangdong Province, P.R. China. Any opinions, findings, conclusions, or recommendations expressed in this publication are those of the author(s) and do not necessarily reflect the view of the funding agency.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Su, H., Fang, J., Liu, W. et al. A deep hashing method of likelihood function adaptive mapping. Neural Comput & Applic 35, 5903–5921 (2023). https://doi.org/10.1007/s00521-022-07962-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-022-07962-3