Abstract
Nowadays, Deep Neural Networks (DNNs) are fundamental to many vision tasks, including large-scale visual recognition. As the primary goal of the DNNs is to characterize complex boundaries of thousands of classes in a high-dimensional space, it is critical to learn higher-order representations for enhancing nonlinear modeling capability. Recently, a novel method called Quantum-State-based Mapping (QSM) has been proposed to improve the feature calibration ability of the existing attention modules in transfer learning tasks. QSM uses the wave function describing the state of microscopic particles to map the feature vector into the probability space. In essence, QSM introduces a novel higher-order representation to improve the nonlinear capability of the network. In this paper, we extend QSM to Quantum Embedding (QE) for designing new attention modules and Self-Organizing Maps, a class of unsupervised learning methods. We also conducted experiments to validate the effectiveness of QE.
J. Zhou—Co-first author.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bausch, J.: Recurrent quantum neural networks. In: NeurIPS, pp. 1368–1379 (2020)
Chen, X., Hsieh, C.J., Gong, B.: When vision transformers outperform resnets without pre-training or strong data augmentations. In: ICLR, pp. 869 (2022)
Chen, Y., Kalantidis, Y., Li, J., Yan, S., Feng, J.: A\(\hat{\,}\)2-nets: Double attention networks. In: NeurIPS, pp. 350–359 (2018)
Cottrell, M., Olteanu, M., Rossi, F., Villa-Vialaneix, N.: Theoretical and applied aspects of the self-organizing maps. In: Merényi, E., Mendenhall, M.J., O’Driscoll, P. (eds.) Advances in Self-Organizing Maps and Learning Vector Quantization. AISC, vol. 428, pp. 3–26. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-28518-4_1
Dosovitskiy, A., et al.: An image is worth 16 \(\times \) 16 words: transformers for image recognition at scale. In: ICLR (2021)
Gao, Z., Xie, J., Wang, Q., Li, P.: Global second-order pooling convolutional networks. In: CVPR, pp. 3024–3033 (2019)
Garg, D., Ikbal, S., Srivastava, S.K., Vishwakarma, H., Karanam, H.P., Subramaniam, L.V.: Quantum embedding of knowledge for reasoning. In: NeurIPS, pp. 5595–5605 (2019)
Goyal, P., et al.: Accurate, large minibatch SGD: training imagenet in 1 hour. CoRR abs/1706.02677 (2017). http://arxiv.org/abs/1706.02677
Griffiths, D.J.: Introduction to quantum mechanics. Am. J. Phys. 63(8), 1–12 (2005)
Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: CVPR, pp. 7132–7141 (2018)
Kerenidis, I., Landman, J., Luongo, A., Prakash, A.: q-means: a quantum algorithm for unsupervised machine learning. In: NeurIPS, pp. 4136–4146 (2019)
Kerenidis, I., Landman, J., Prakash, A.: Quantum algorithms for deep convolutional neural networks. In: ICLR (2020)
Kohonen, T.: Self-organized formation of topologically correct feature maps. Biol. Cybern. 43(1), 59–69 (1982)
Lee, H., Kim, H., Nam, H.: SRM: a style-based recalibration module for convolutional neural networks. In: ICCV, pp. 1854–1862 (2019)
MMClassification Contributors: Openmmlab’s image classification toolbox and benchmark. https://github.com/open-mmlab/mmclassification (2020)
Olivas, E.S., Guerrero, J.D.M., Martinez-Sober, M., Magdalena-Benedito, J.R., Serrano, L., et al.: Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques. IGI Global (2009)
Qin, Z., Zhang, P., Wu, F., Li, X.: FCANet: frequency channel attention networks. In: ICCV, pp. 763–772 (2021)
Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., Hu, Q.: ECA-Net: efficient channel attention for deep convolutional neural networks. In: CVPR, pp. 11531–11539 (2020)
Wightman, R., Touvron, H., Jegou, H.: Resnet strikes back: an improved training procedure in TIMM. In: NeurIPS 2021 Workshop on ImageNet: Past, Present, and Future (2021)
Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_1
Yang, Z., Zhu, L., Wu, Y., Yang, Y.: Gated channel transformation for visual recognition. In: CVPR (2020)
Yosinski, J., Clune, J., Bengio, Y., Lipson, H.: How transferable are features in deep neural networks? In: Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2, pp. 3320–3328 (2014)
Zhang, J., et al.: An application of quantum mechanics to attention methods in computer vision. In: ICASSP (2023)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhang, J. et al. (2024). Applications of Quantum Embedding in Computer Vision. In: Luo, B., Cheng, L., Wu, ZG., Li, H., Li, C. (eds) Neural Information Processing. ICONIP 2023. Communications in Computer and Information Science, vol 1965. Springer, Singapore. https://doi.org/10.1007/978-981-99-8145-8_14
Download citation
DOI: https://doi.org/10.1007/978-981-99-8145-8_14
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8144-1
Online ISBN: 978-981-99-8145-8
eBook Packages: Computer ScienceComputer Science (R0)