Abstract
The class-imbalanced distributions between different classes in the visual world pose great challenges for deep learning-based classification models particularly on correct prediction of minority classes. In this study, different from existing strategies to alleviate the data imbalance issue, a novel mechanism based on the CutMix regularization technique is proposed for imbalanced image classification. The novelty is from two aspects. First, a novel sampling strategy is proposed to create the synthetic training data with a more balanced distribution. Second, labels of synthetic images were assigned with a bias toward minority classes. With the novel sampling and label assignment, more synthetic images of minority classes can be obtained to balance the class distribution of training data. Experiments on three benchmark datasets justified that the proposed method consistently outperforms commonly used strategies to alleviate the class imbalance issue.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Buda, A., Maki, A., Mazurowski. M.A.: A systematic study of the class imbalance problem in convolutional neural networks. Neural Netw. 106 (2018)
Cao, K., Wei, C., Gaidon, A., Arechiga, N., Ma, T.: Learning imbalanced datasets with label-distribution-aware margin loss. In: Proceedings of the 33rd International Conference on Advances in Neural Information Processing Systems (2019)
Chawla, N., Bowyer, K., Hall, L., Kegelmeyer. P.: SMOTE: synthetic minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357 (2002)
Chou, H.-P., Chang, S.-C., Pan, J.-Y., Wei, W., Juan, D.-C.: Remix: rebalanced Mixup. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12540, pp. 95–110. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-65414-6_9
Chu, P., Bian, X., Liu, S., Ling. H.: Feature space augmentation for long-tailed data. In: European Conference on Computer Vision (2020)
Cui, Y., Jia, M., Lin, T., Song, Y., Belongie, S.: Class-balanced loss based on effective number of samples. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2019)
Deng, J., Dong, W., Socher, R., Li, L.-J., Kai, L., Li, F.-F.: ImageNet: a large-scale hierarchical image database. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2009)
DeVries, T., Taylor, G.: Improved regularization of convolutional neural networks with cutout. http://arxiv.org/abs/1708.04552 (2017)
Galdran, A., Carneiro, G., González Ballester, M.A.: Balanced-MixUp for highly imbalanced medical image classification. In: de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C. (eds.) MICCAI 2021. LNCS, vol. 12905, pp. 323–333. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-87240-3_31
Geifman, Y., El-Yaniv, R.: Deep active learning over the long tail. In: International Conference on Learning Representations (2018)
Goyal, P., et al.: Accurate, large minibatch SGD: Training ImageNet in 1 hour. http://arxiv.org/abs/1706.02677 (2017)
Han, H., Wang, W.-Y., Mao, B.-H.: Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning. In: International Conference on Intelligent Computing (2005)
He, H., Bai, Y., Garcia, E., Li, S.: ADASYN: adaptive synthetic sampling approach for imbalanced learning. In: IEEE International Joint Conference on Neural Networks (2008)
He, H., Garcia, E.A.: Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 21, 1263–1284 (2009)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2016)
Huang, C., Li, Y., Change Loy, C., Tang, X.: Learning deep representation for imbalanced classification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2016)
Japkowicz, N., Stephen, S.: The class imbalance problem: a systematic study. Intell. Data Anal. 6, 429–449 (2002)
Kang, B., et al.: Decoupling representation and classifier for long-tailed recognition. In: International Conference on Learning Representations (2020)
Kubat, M., Matwin, S.: Addressing the curse of imbalanced training sets: one-sided selection. In: Proceedings of the International Conference on Machine Learning (1997)
Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. In: Proceedings of the IEEE (1998)
Lin, T., Goyal, P., Girshick, R., He, K., Dollar, P.: Focal loss for dense object detection. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (2017)
Liu, Z., Miao, Z., Zhan, X., Wang, J., Gong, B., Yu, S.X.: Large-scale long-tailed recognition in an open world. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2019)
Shen, L., Lin, Z., Huang, Q.: Relay backpropagation for effective learning of deep convolutional neural networks. In: European Conference on Computer Vision (2016)
Ming Ting, K.: A comparative study of cost-sensitive boosting algorithms. In: Proceedings of the International Conference on Machine Learning (2000)
Wang, Y., Ramanan, D., Hebert, M.: Learning to model the tail. In: Advances in Neural Information Processing Systems (2017)
Xie, S., Girshick, R., Dollár, P., Tu, Z., He, K.: Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2017)
Yun, S., Han, D., Joon Oh, S., Chun, S., Choe, J., Yoo, Y.: CutMix: regularization strategy to train strong classifiers with localizable features. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (2019)
Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: Mixup: beyond empirical risk minimization. In: International Conference on Learning Representations (2018)
Zhou, B., Cui, Q., Wei, X., Zhaomin Chen, X.: BBN: bilateral-branch network with cumulative learning for long-tailed visual recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2019)
Zhou, Z., Liu, X.: Training cost-sensitive neural networks with methods addressing the class imbalance problem. In: IEEE Trans. Knowl. Data Eng. 18, 63–77 (2006)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Zeng, C., Lu, H., Chen, K., Wang, R., Tao, J. (2023). Synthetic Minority with CutMix for Imbalanced Image Classification. In: Arai, K. (eds) Intelligent Systems and Applications. IntelliSys 2022. Lecture Notes in Networks and Systems, vol 543. Springer, Cham. https://doi.org/10.1007/978-3-031-16078-3_37
Download citation
DOI: https://doi.org/10.1007/978-3-031-16078-3_37
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-16077-6
Online ISBN: 978-3-031-16078-3
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)