Abstract
In the past decade, the attention mechanism has played an increasingly important role in computer vision. Such an attention mechanism can be regarded as a dynamic weight adjustment process based on features of the input image. In this paper, we propose Quantum-Effects-based Attention Networks (QEA-Net), the simple yet effective attention networks, they can be integrated into many network architectures seamlessly. QEA-Net uses quantum effects between two identical particles to enhance the global channel information representation of the attention module. Our method could consistently outperform the SENet, with a lower number of parameters and computational cost. We evaluate QEA-Net through experiments on ImageNet-1K and compare it with state-of-the-art counterparts. We also demonstrated the effect of QEA-Net in combination with pre-trained networks on small downstream transfer learning tasks.
J. Zhang and J. Zhou—Contributed equally to this work.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bausch, J.: Recurrent quantum neural networks. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H. (eds.) Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, pp. 1368–1379, December 6–12, 2020, virtual (2020)
Chen, X., Hsieh, C.J., Gong, B.: When vision transformers outperform ResNets without pre-training or strong data augmentations. In: International Conference on Learning Representations, pp. 869 (2022)
Chen, Y., Kalantidis, Y., Li, J., Yan, S., Feng, J.: A2-Nets: double attention networks. In: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, December 3–8, 2018, Montréal, Canada, pp. 350–359 (2018)
Csurka, G.: A comprehensive survey on domain adaptation for visual applications. In: Csurka, G. (ed.) Domain Adaptation in Computer Vision Applications, pp. 1–35. Springer, Advances in Computer Vision and Pattern Recognition (2017)
Dosovitskiy, A., et al.: An image is worth 16\(\times \)16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)
Gao, Z., Xie, J., Wang, Q., Li, P.: Global second-order pooling convolutional networks. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16–20, 2019, pp. 3024–3033. Computer Vision Foundation / IEEE (2019)
Garg, D., Ikbal, S., Srivastava, S.K., Vishwakarma, H., Karanam, H.P., Subramaniam, L.V.: Quantum embedding of knowledge for reasoning. In: Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R. (eds.) Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8–14, 2019, Vancouver, BC, Canada, pp. 5595–5605 (2019)
Goyal, P., et al.: Accurate, large minibatch SGD: training ImageNet in 1 hour. CoRR abs/1706.02677 (2017). http://arxiv.org/abs/1706.02677
Griffiths, D.J.: Introduction to quantum mechanics. Am. J. Phys. 63(8) (2005)
Hu, J., Shen, L., Albanie, S., Sun, G., Vedaldi, A.: Gather-excite: exploiting feature context in convolutional neural networks. In: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, pp. 9423–9433
Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18–22, 2018, pp. 7132–7141. Computer Vision Foundation / IEEE Computer Society (2018)
Kerenidis, I., Landman, J., Luongo, A., Prakash, A.: q-means: a quantum algorithm for unsupervised machine learning. In: Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R. (eds.) Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8–14, 2019, Vancouver, BC, Canada, pp. 4136–4146 (2019)
Kerenidis, I., Landman, J., Prakash, A.: Quantum algorithms for deep convolutional neural networks. In: 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26–30, 2020. OpenReview.net (2020)
Kornblith, S., Shlens, J., Le, Q.V.: Do better ImageNet models transfer better? In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16–20, 2019. pp. 2661–2671. Computer Vision Foundation / IEEE (2019)
Lee, H., Kim, H., Nam, H.: SRM: a style-based recalibration module for convolutional neural networks. In: 2019 IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea (South), October 27 - November 2, 2019, pp. 1854–1862. IEEE (2019)
Liu, H., Dai, Z., So, D.R., Le, Q.V.: Pay attention to MLPs. In: Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6–14, 2021, virtual, pp. 9204–9215 (2021)
MMClassification Contributors: Openmmlab’s image classification toolbox and benchmark. https://github.com/open-mmlab/mmclassification (2020)
Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2010)
Qin, Z., Zhang, P., Wu, F., Li, X.: FcaNet: frequency channel attention networks. In: 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 763–772 (2021)
Silver, D., Patel, T., Tiwari, D.: QUILT: effective multi-class classification on quantum computers using an ensemble of diverse quantum classifiers. In: Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, pp. 8324–8332 (2022)
Tolstikhin, I.O., et al.: MLP-mixer: an all-MLP architecture for vision. In: Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems, vol. 34, pp. 24261–24272. Curran Associates, Inc. (2021)
Touvron, H., et al.: ResMLP: feedforward networks for image classification with data-efficient training. CoRR abs/2105.03404 (2021). https://arxiv.org/abs/2105.03404
Wightman, R., Touvron, H., Jegou, H.: ResNet strikes back: an improved training procedure in timm. In: NeurIPS 2021 Workshop on ImageNet: Past, Present, and Future (2021)
Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_1
Yang, Z., Zhu, L., Wu, Y., Yang, Y.: Gated channel transformation for visual recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020)
Zhang, J., et al.: An application of quantum mechanics to attention methods in computer vision. In: ICASSP 2023–2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2023)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhang, J. et al. (2024). QEA-Net: Quantum-Effects-based Attention Networks. In: Liu, Q., et al. Pattern Recognition and Computer Vision. PRCV 2023. Lecture Notes in Computer Science, vol 14427. Springer, Singapore. https://doi.org/10.1007/978-981-99-8435-0_9
Download citation
DOI: https://doi.org/10.1007/978-981-99-8435-0_9
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8434-3
Online ISBN: 978-981-99-8435-0
eBook Packages: Computer ScienceComputer Science (R0)