skip to main content
10.1145/3616901.3616909acmotherconferencesArticle/Chapter ViewAbstractPublication PagesfaimlConference Proceedingsconference-collections
research-article

BSDIB: Balanced Self-Distillation Information Bottleneck for Long-Tailed Visual Recognition

Published: 05 March 2024 Publication History

Abstract

Real-world data are usually long-tailed, which results in poor performance of visual recognition systems. Current prevailing methods decouple the training procedure into representation learning and classification. However, the two stages in such methods are independent, resulting in sub-optimal solutions. In this paper, we propose a one-stage method, namely balanced self-distillation information bottleneck (BSDIB), to take care of both representation learning and classification simultaneously. Specifically, a re-weighting strategy is used to obtain an unbiased classifier, and the class-dependent self-distillation information bottleneck is utilized to promote data representations. Extensive experiments on three commonly-used long-tailed datasets indicate that BSDIB achieves SOTA performance and the visualization results show that our approach focuses more on the target object.

References

[1]
Olga Russakovsky, Jia Deng, Hao Su (2015). Imagenet large scale visual recognition challenge. International journal of computer vision, vol. 115, pp. 211-252
[2]
Mateusz Buda, Atsuto Maki and Maciej A Mazurowski (2018). A systematic study of the class imbalance problem in convolutional neural networks. Neural networks, vol. 106, pp. 249-259
[3]
Kaidi Cao, Colin Wei, Adrien Gaidon (2019). Learning imbalanced datasets with label-distribution-aware margin loss. in NeurIPS, vol. 32
[4]
Jiawei Ren, Cunjun Yu, Shunan Sheng (2020). Balanced meta-softmax for long-tailed visual recognition. in NeurIPS, vol. 33, pp. 4175-4186
[5]
Bingyi Kang, Saining Xie, Marcus Rohrbach (2020). Decoupling representation and classifier for long-tailed recognition. in ICLR
[6]
Boyan Zhou, Quan Cui, XiuShen Wei (2020). Bbn: Bilateral-branch network with cumulative learning for long-tailed visual recognition. in CVPR, pp. 9719-9728
[7]
Zhisheng Zhong, Jiequan Cui, Shu Liu (2021). Improving calibration for long-tailed recognition. in CVPR, pp. 16489-16498
[8]
Naftali Tishby, Fernando C Pereira and William Bialek (2000). The information bottleneck method. arXiv preprint physics/0004057
[9]
Alexander A Alemi, Ian Fischer, Joshua V Dillon (2017). Deep Variational Information Bottleneck. in ICLR
[10]
Xudong Tian, Zhizhong Zhang, Shaohui Lin (2021). Farewell to mutual information: Variational distillation for cross-modal person re-identification. in CVPR, pp. 1522-1531
[11]
TsungYi Lin, Priya Goyal, Ross Girshick (2017). Focal loss for dense object detection. in ICCV, pp. 2980-2988
[12]
Bingyi Kang, Yu Li, Sa Xie (2021). Exploring balanced feature spaces for representation learning. in ICLR
[13]
Tianhong Li, Peng Cao, Yuan Yuan (2022). Targeted supervised contrastive learning for long-tailed recognition. in CVPR, pp. 6918-6928
[14]
Songyang Zhang, Zeming Li, Shipeng Yan (2021). Distribution alignment: A unified framework for long-tail visual recognition. in CVPR, pp. 2361-2370
[15]
Youngkyu Hong, Seungju Han, Kwanghee Choi (2021). Disentangling label distribution for long-tailed visual recognition. in CVPR, pp. 6626-6636
[16]
Ramprasaath R Selvaraju, Michael Cogswell, Abhishek Das (2017). Grad-cam: Visual explanations from deep networks via gradient-based localization. in ICCV, pp. 618-626

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
FAIML '23: Proceedings of the 2023 International Conference on Frontiers of Artificial Intelligence and Machine Learning
April 2023
296 pages
ISBN:9798400707544
DOI:10.1145/3616901
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 05 March 2024

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. deep learning
  2. information bottleneck
  3. long-tailed recognition
  4. self-distillation

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Funding Sources

Conference

FAIML 2023

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 19
    Total Downloads
  • Downloads (Last 12 months)19
  • Downloads (Last 6 weeks)2
Reflects downloads up to 01 Mar 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media