skip to main content
10.1145/3590003.3590082acmotherconferencesArticle/Chapter ViewAbstractPublication PagescacmlConference Proceedingsconference-collections
research-article

Improved Convolutional Neural Networks by Integrating High-frequency Information for Image Classification

Published:29 May 2023Publication History

ABSTRACT

Deep convolutional neural networks are powerful and popular tools as deep learning emerges in recent years for image classification in computer vision. However, it is difficult to learn convolutional filters from the examples. The innate frequency property of the data has not been well considered. To address this problem, we find high-frequency information import within deep networks and therefore propose our high-pass attention method (HPA) to help the learning process. HPA explicitly generates high-frequency information via a stage-wise high-pass filter to alleviate the burden of learning such information. Strengthened by channel attention on the concatenated features, our method demonstrates consistent improvements upon ResNet-18/ResNet-50 by 1.36%/1.60% and 1.47%/1.39% on the ImageNet-1K dataset and the Food-101 dataset, respectively, as well as the effectiveness over a variety of modules.

References

  1. Yunpeng Chen, Haoqi Fan, Bing Xu, Zhicheng Yan, Yannis Kalantidis, Marcus Rohrbach, Shuicheng Yan, and Jiashi Feng. 2019. Drop an octave: Reducing spatial redundancy in convolutional neural networks with octave convolution. In Proceedings of CVPR. 3435–3444.Google ScholarGoogle ScholarCross RefCross Ref
  2. Bowen Cheng, Rong Xiao, Jianfeng Wang, Thomas Huang, and Lei Zhang. 2020. High frequency residual learning for multi-scale image classification. In 30th British Machine Vision Conference, BMVC 2019.Google ScholarGoogle Scholar
  3. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2015. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision. 1026–1034.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of CVPR. 770–778.Google ScholarGoogle ScholarCross RefCross Ref
  5. Geoffrey E Hinton, Alex Krizhevsky, and Ilya Sutskever. 2012. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 25 (2012), 1106–1114.Google ScholarGoogle Scholar
  6. Jie Hu, Li Shen, and Gang Sun. 2018. Squeeze-and-excitation networks. In Proceedings of CVPR. 7132–7141.Google ScholarGoogle ScholarCross RefCross Ref
  7. Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of Machine Learning Research, Vol. 37. 448–456.Google ScholarGoogle Scholar
  8. Qiufu Li, Linlin Shen, Sheng Guo, and Zhihui Lai. 2020. Wavelet integrated cnns for noise-robust image classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 7245–7254.Google ScholarGoogle ScholarCross RefCross Ref
  9. Pengju Liu, Hongzhi Zhang, Wei Lian, and Wangmeng Zuo. 2019. Multi-level wavelet convolutional neural networks. IEEE Access 7 (2019), 74973–74985.Google ScholarGoogle ScholarCross RefCross Ref
  10. Jongchan Park, Sanghyun Woo, Joon-Young Lee, and In-So Kweon. 2018. BAM: Bottleneck Attention Module. In British Machine Vision Conference (BMVC). British Machine Vision Association (BMVA).Google ScholarGoogle Scholar
  11. Zhinan Qiao, Xiaohui Yuan, and Mohamed Elhoseny. 2020. Urban Scene Recognition via Deep Network Integration. In Urban Intelligence and Applications. Singapore, 135–149.Google ScholarGoogle Scholar
  12. Haohan Wang, Xindi Wu, Zeyi Huang, and Eric P Xing. 2020. High-frequency Component Helps Explain the Generalization of Convolutional Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 8684–8694.Google ScholarGoogle ScholarCross RefCross Ref
  13. Sanghyun Woo, Jongchan Park, Joon-Young Lee, and In So Kweon. 2018. Cbam: Convolutional block attention module. In Proceedings of ECCV. 3–19.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Xiaohui Yuan, Zhinan Qiao, and Abolfazl Meyarian. 2022. Scale Attentive Network for Scene Recognition. Neurocomputing 492 (2022), 612–623.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Matthew D Zeiler and Rob Fergus. 2014. Visualizing and understanding convolutional networks. In Proceedings of ECCV. 818–833.Google ScholarGoogle ScholarCross RefCross Ref
  16. Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. 2018. mixup: Beyond Empirical Risk Minimization. In International Conference on Learning Representations.Google ScholarGoogle Scholar
  17. Richard Zhang. 2019. Making Convolutional Networks Shift-Invariant Again. In International Conference on Machine Learning. 7324–7334.Google ScholarGoogle Scholar

Index Terms

  1. Improved Convolutional Neural Networks by Integrating High-frequency Information for Image Classification

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      CACML '23: Proceedings of the 2023 2nd Asia Conference on Algorithms, Computing and Machine Learning
      March 2023
      598 pages
      ISBN:9781450399449
      DOI:10.1145/3590003

      Copyright © 2023 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 29 May 2023

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

      Acceptance Rates

      CACML '23 Paper Acceptance Rate93of241submissions,39%Overall Acceptance Rate93of241submissions,39%
    • Article Metrics

      • Downloads (Last 12 months)27
      • Downloads (Last 6 weeks)3

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format