Loading [a11y]/accessibility-menu.js
Distribution Unified and Probability Space Aligned Teacher-Student Learning for Imbalanced Visual Recognition | IEEE Journals & Magazine | IEEE Xplore

Distribution Unified and Probability Space Aligned Teacher-Student Learning for Imbalanced Visual Recognition


Abstract:

Imbalanced label distribution is usually the case for real-world data, which poses a challenge for training unbiased recognition model. In this paper, we study two underl...Show More

Abstract:

Imbalanced label distribution is usually the case for real-world data, which poses a challenge for training unbiased recognition model. In this paper, we study two underlying mismatches, i.e., distribution mismatch and probability space mismatch, present in class-imbalanced learning. Firstly, we analyze the label distribution mismatch between imbalanced training data and balanced test data, and introduce a distribution unified framework to unify the two distributions through probability conversion. Secondly, we analyze that the utilization of cross-entropy loss under the proposed framework may lead to probability space mismatch, where the conversion of the predictive probability is implemented in softmax probability space while the comparison with one-hot label is implemented in true probability space. To alleviate this dilemma, we involve a teacher model and formulate a teacher-student learning strategy, which contains two novel techniques. The Teacher Guided Label Smoothing (TGLS) is first proposed to relax the one-hot label to smoother pseudo softmax probability, which is more aligned with the softmax probability space. Additionally, we propose Distribution Unified Knowledge Distillation (DU-KD) under the proposed framework to further reduce both the mismatches. Experiments on several benchmarks confirm the top-level performance of the proposed method.
Page(s): 2414 - 2425
Date of Publication: 01 September 2023

ISSN Information:

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.