skip to main content
10.1145/3590003.3590082acmotherconferencesArticle/Chapter ViewAbstractPublication PagescacmlConference Proceedingsconference-collections
research-article

Improved Convolutional Neural Networks by Integrating High-frequency Information for Image Classification

Published: 29 May 2023 Publication History

Abstract

Deep convolutional neural networks are powerful and popular tools as deep learning emerges in recent years for image classification in computer vision. However, it is difficult to learn convolutional filters from the examples. The innate frequency property of the data has not been well considered. To address this problem, we find high-frequency information import within deep networks and therefore propose our high-pass attention method (HPA) to help the learning process. HPA explicitly generates high-frequency information via a stage-wise high-pass filter to alleviate the burden of learning such information. Strengthened by channel attention on the concatenated features, our method demonstrates consistent improvements upon ResNet-18/ResNet-50 by 1.36%/1.60% and 1.47%/1.39% on the ImageNet-1K dataset and the Food-101 dataset, respectively, as well as the effectiveness over a variety of modules.

References

[1]
Yunpeng Chen, Haoqi Fan, Bing Xu, Zhicheng Yan, Yannis Kalantidis, Marcus Rohrbach, Shuicheng Yan, and Jiashi Feng. 2019. Drop an octave: Reducing spatial redundancy in convolutional neural networks with octave convolution. In Proceedings of CVPR. 3435–3444.
[2]
Bowen Cheng, Rong Xiao, Jianfeng Wang, Thomas Huang, and Lei Zhang. 2020. High frequency residual learning for multi-scale image classification. In 30th British Machine Vision Conference, BMVC 2019.
[3]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2015. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision. 1026–1034.
[4]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of CVPR. 770–778.
[5]
Geoffrey E Hinton, Alex Krizhevsky, and Ilya Sutskever. 2012. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 25 (2012), 1106–1114.
[6]
Jie Hu, Li Shen, and Gang Sun. 2018. Squeeze-and-excitation networks. In Proceedings of CVPR. 7132–7141.
[7]
Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of Machine Learning Research, Vol. 37. 448–456.
[8]
Qiufu Li, Linlin Shen, Sheng Guo, and Zhihui Lai. 2020. Wavelet integrated cnns for noise-robust image classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 7245–7254.
[9]
Pengju Liu, Hongzhi Zhang, Wei Lian, and Wangmeng Zuo. 2019. Multi-level wavelet convolutional neural networks. IEEE Access 7 (2019), 74973–74985.
[10]
Jongchan Park, Sanghyun Woo, Joon-Young Lee, and In-So Kweon. 2018. BAM: Bottleneck Attention Module. In British Machine Vision Conference (BMVC). British Machine Vision Association (BMVA).
[11]
Zhinan Qiao, Xiaohui Yuan, and Mohamed Elhoseny. 2020. Urban Scene Recognition via Deep Network Integration. In Urban Intelligence and Applications. Singapore, 135–149.
[12]
Haohan Wang, Xindi Wu, Zeyi Huang, and Eric P Xing. 2020. High-frequency Component Helps Explain the Generalization of Convolutional Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 8684–8694.
[13]
Sanghyun Woo, Jongchan Park, Joon-Young Lee, and In So Kweon. 2018. Cbam: Convolutional block attention module. In Proceedings of ECCV. 3–19.
[14]
Xiaohui Yuan, Zhinan Qiao, and Abolfazl Meyarian. 2022. Scale Attentive Network for Scene Recognition. Neurocomputing 492 (2022), 612–623.
[15]
Matthew D Zeiler and Rob Fergus. 2014. Visualizing and understanding convolutional networks. In Proceedings of ECCV. 818–833.
[16]
Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. 2018. mixup: Beyond Empirical Risk Minimization. In International Conference on Learning Representations.
[17]
Richard Zhang. 2019. Making Convolutional Networks Shift-Invariant Again. In International Conference on Machine Learning. 7324–7334.

Cited By

View all
  • (2025)Frequency Regulated Channel-Spatial Attention module for improved image classificationExpert Systems with Applications10.1016/j.eswa.2024.125463260(125463)Online publication date: Jan-2025

Index Terms

  1. Improved Convolutional Neural Networks by Integrating High-frequency Information for Image Classification

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    CACML '23: Proceedings of the 2023 2nd Asia Conference on Algorithms, Computing and Machine Learning
    March 2023
    598 pages
    ISBN:9781450399449
    DOI:10.1145/3590003
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 29 May 2023

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. attention
    2. classification
    3. deep convolutional neural networks
    4. high frequency

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    CACML 2023

    Acceptance Rates

    CACML '23 Paper Acceptance Rate 93 of 241 submissions, 39%;
    Overall Acceptance Rate 93 of 241 submissions, 39%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)30
    • Downloads (Last 6 weeks)3
    Reflects downloads up to 25 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2025)Frequency Regulated Channel-Spatial Attention module for improved image classificationExpert Systems with Applications10.1016/j.eswa.2024.125463260(125463)Online publication date: Jan-2025

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media