Abstract
Exemplar-free class-incremental learning recognizes both old and new classes without saving old class exemplars because of storage limitations and privacy constraints. To address the forgetting of knowledge caused by the absence of old training data, we present a novel method that consists of two modules, multi-view prototype balance and temporary proxy constraints, which are based on feature retention and representation optimization. Specifically, multi-view prototype balance first extends the prototypes to maintain the general state of the class and then balances these prototypes combining knowledge distillation and prototype compensation to ensure the stability and plasticity of the model. To alleviate the feature overlap, the proposed temporary proxy constraint sets the temporary proxies to lightly compress the feature distribution during each mini-batch of training. Extensive experiments on five datasets with different settings demonstrate the superiority of our method against the state-of-the-art exemplar-free class-incremental learning methods.












Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability and Access
The datasets generated during and/or analyzed during the current study are available in the [CIFAR-100 dataset] with [https://www.cs.toronto.edu/~kriz/cifar.html], [Tiny-ImageNet dataset] with [https://www.kaggle.com/c/tiny-imagenet], [ImageNet-Sub dataset] with [https://www.kaggle.com/datasets/ambityga/imagenet100], and [CUB dataset, ImageNet-R] with [https://github.com/zhoudw-zdw/RevisitingCIL].
References
Touvron H, Bojanowski P, Caron M, Cord M, El-Nouby A, Grave E, Izacard G, Joulin A, Synnaeve G, Verbeek J et al (2022) Resmlp: Feedforward networks for image classification with data-efficient training. IEEE Trans Pattern Anal Mach Intell 45(4):5314–5321
He T, Zhang Z, Zhang H, Zhang Z, Xie J, Li M (2019) Bag of tricks for image classification with convolutional neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 558–567
Baek S-H, Heide F (2021) Polka lines: Learning structured illumination and reconstruction for active stereo. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 5757–5767
Long T, Liang Z, Liu Q (2019) Advanced technology of high-resolution radar: target detection, tracking, imaging, and recognition. Sci China Inf Sci 62:1–26
Devlin J, Chang M, Lee K, Toutanova K (2019) BERT: pre-training of deep bidirectional transformers for language understanding. In: Annual conference of the North American Chapter of the association for computational linguistics: human language technologies. Association for Computational Linguistics, pp 4171–4186
Wu L, Chen Y, Shen K, Guo X, Gao H, Li S, Pei J, Long B et al (2023) Graph neural networks for natural language processing: a survey. Found Trends® Mach Learn 16(2):119–328
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30
Castro FM, Marín-Jiménez MJ, Guil N, Schmid C, Alahari K (2018) End-to-end incremental learning. In: Proceedings of the European conference on computer vision, pp 233–248
Hou S, Pan X, Loy CC, Wang Z, Lin D (2019) Learning a unified classifier incrementally via rebalancing. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 831–839
De Lange M, Aljundi R, Masana M, Parisot S, Jia X, Leonardis A, Slabaugh G, Tuytelaars T (2021) A continual learning survey: defying forgetting in classification tasks. IEEE Trans Pattern Anal Mach Intell 44(7):3366–3385
Sun L, Zhang M, Wang B, Tiwari P (2024) Few-shot class-incremental learning for medical time series classification. IEEE J Biomed Health Inform 28(4):1872–1882
Singh T, Kalra R, Mishra S, Satakshi, Kumar M (2023) An efficient real-time stock prediction exploiting incremental learning and deep learning. Evol Syst 14(6):919–937
Constantinides C, Shiaeles S, Ghita B, Kolokotronis N (2019) A novel online incremental learning intrusion prevention system. In: 2019 10th IFIP International conference on new technologies, mobility and security (NTMS), pp 1–6
Rebuffi S-A, Kolesnikov A, Sperl G, Lampert CH (2017) icarl: Incremental classifier and representation learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2001–2010
Shin H, Lee JK, Kim J, Kim J (2017) Continual learning with deep generative replay. Adv Neural Inf Process Syst 30
Belouadah E, Popescu A (2019) Il2m: Class incremental learning with dual memory. In: Proceedings of the IEEE/CVF International conference on computer vision, pp 583–592
Liu Y, Parisot S, Slabaugh G, Jia X, Leonardis A, Tuytelaars T (2020) More classifiers, less forgetting: a generic multi-classifier paradigm for incremental learning. In: Proceedings of the european conference on computer vision, pp 699–716
Zhu F, Zhang X-Y, Wang C, Yin F, Liu C-L (2021) Prototype augmentation and self-supervision for incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 5871–5880
Zhu K, Zhai W, Cao Y, Luo J, Zha Z-J (2022) Self-sustaining representation expansion for non-exemplar class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9296–9305
Yu L, Twardowski B, Liu X, Herranz L, Wang K, Cheng Y, Jui S, Weijer JVD (2020) Semantic drift compensation for class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6982–6991
Hinton G, Vinyals O, Dean J Distilling the knowledge in a neural network. arXiv:1503.02531
Kim S, Kim D, Cho M, Kwak S (2020) Proxy anchor loss for deep metric learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3238–3247
Lee H, Hwang SJ, Shin J (2020) Self-supervised label augmentation via input transformations. In: International conference on machine learning, pp 5714–5724
Prideaux D, Ash J, Cottrell A (2013) Integrated learning. In: Oxford textbook of medical education. Oxford University Press Oxford, pp 63–73
Li B, Yang Y, Zhao Z, Ni X, Zhang D (2024) A novel ensemble learning approach for intelligent logistics demand management. J Internet Technol 25(4):507–515
Kirkpatrick J, Pascanu R, Rabinowitz N, Veness J, Desjardins G, Rusu AA, Milan K, Quan J, Ramalho T, Grabska-Barwinska A et al (2017) Overcoming catastrophic forgetting in neural networks. Proc Natl Acad Sci 114(13):3521–3526
Aljundi R, Babiloni F, Elhoseiny M, Rohrbach M, Tuytelaars T (2018) Memory aware synapses: learning what (not) to forget. In: Proceedings of the European conference on computer vision, pp 139–154
Yoon J, Yang E, Lee J, Hwang SJ (2018) Lifelong learning with dynamically expandable networks. In: International conference on learning representations
Yan S, Xie J, He X (2021) Der: Dynamically expandable representation for class incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3014–3023
Fu Z, Wang Z, Xu X, Li D, Yang H (2023) Knowledge aggregation networks for class incremental learning. Pattern Recognit 137:109310
Lin H, Zhang Y, Qiu Z, Niu S, Gan C, Liu Y, Tan M (2022) Prototype-guided continual adaptation for class-incremental unsupervised domain adaptation. In: Proceedings of the European conference on computer vision, pp 351–368
Zhu F, Cheng Z, Zhang X-Y, Liu C-L (2021) Class-incremental learning via dual augmentation. Adv Neural Inf Process Syst 34:14306–14318
Petit G, Popescu A, Schindler H, Picard D, Delezoide B (2023) Fetril: Feature translation for exemplar-free class-incremental learning. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp 3911–3920
Li Q, Peng Y, Zhou J (2024) Fcs: Feature calibration and separation for non-exemplar class incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 28495–28504
Peyré G, Cuturi M et al (2019) Computational optimal transport: with applications to data science. Found Trends® Mach Learn 11(5–6):355–607
Kaya M, Bilge HŞ (2019) Deep metric learning: a survey. Symmetry 11(9):1066
Li X, Yang X, Ma Z, Xue J-H (2023) Deep metric learning for few-shot image classification: a review of recent developments. Pattern Recognit 109381
Schroff F, Kalenichenko D, Philbin J (2015) Facenet: A unified embedding for face recognition and clustering. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 815–823
Movshovitz-Attias Y, Toshev A, Leung TK, Ioffe S, Singh S (2017) No fuss distance metric learning using proxies. In: Proceedings of the IEEE international conference on computer vision, pp 360–368
Wang J, Zhou F, Wen S, Liu X, Lin Y (2017) Deep metric learning with angular loss. In: Proceedings of the IEEE international conference on computer vision, pp 2593–2601
Yu B, Tao D (2019) Deep metric learning with tuplet margin loss. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 6490–6499
Elezi I, Vascon S, Torcinovich A, Pelillo M, Leal-Taixé L (2020) The group loss for deep metric learning. In: Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part VII 16. Springer, pp 277–294
Krizhevsky A, Hinton G et al (2009) Learning multiple layers of features from tiny images
Yang K, Yau JH, Fei-Fei L, Deng J, Russakovsky O (2022) A study of face obfuscation in imagenet. In: International conference on machine learning, pp 25313–25330
Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset
Hendrycks D, Basart S, Mu N, Kadavath S, Wang F, Dorundo E, Desai R, Zhu T, Parajuli S, Guo M, et al (2021) The many faces of robustness: a critical analysis of out-of-distribution generalization. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 8340–8349
Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M et al (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115:211–252
Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp 248–255
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 770–778
Kingma DP, Ba J (2015) Adam: A method for stochastic optimization. In: International conference on learning representations
Li Z, Hoiem D (2017) Learning without forgetting. IEEE Trans Pattern Anal Mach Intell 40(12):2935–2947
Shi Y, Shi D, Qiao Z, Wang Z, Zhang Y, Yang S, Qiu C (2023) Multi-granularity knowledge distillation and prototype consistency regularization for class-incremental learning. Neural Netw 164:617–630
Chatzimparmpas A, Martins RM, Kerren A (2020) t-visne: Interactive assessment and interpretation of t-sne projections. IEEE Trans Vis Comput Graph 26(8):2696–2714
Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, Sastry G, Askell A, Mishkin P, Clark J et al (2021) Learning transferable visual models from natural language supervision. In: International conference on machine learning, pp 8748–8763
Shorten C, Khoshgoftaar TM (2019) A survey on image data augmentation for deep learning. J Big Data 6(1):1–48
Feng SY, Gangal V, Wei J, Chandar S, Vosoughi S, Mitamura T, Hovy EH (2021) A survey of data augmentation approaches for NLP. In: International joint conference on natural language processing, Vol. ACL/IJCNLP 2021 of Findings of ACL. Association for Computational Linguistics, pp 968–988
Yun S, Han D, Oh SJ, Chun S, Choe J, Yoo Y (2019) Cutmix: Regularization strategy to train strong classifiers with localizable features. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 6023–6032
Zhang H, Cissé M, Dauphin, Lopez-Paz D (2018) mixup: Beyond empirical risk minimization. In: International conference on learning representations. OpenReview.net
DeVries T, Taylor GW (2017) Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552
Acknowledgements
This work is supported by Natural Science Foundation of China under Grant No. 62476087, the National Key Research and Development Program of China (2022YFB3203500), Shanghai Science and Technology Program “Federated based cross-domain and cross-task incremental learning” under Grant No. 21511100800, Natural Science Foundation of China under Grant No. 62002193, Chinese Defense Program of Science and Technology under Grant No.2021-JCJQ-JJ-0041, China Aerospace Science and Technology Corporation Industry-University-Research Cooperation Foundation of the Eighth Research Institute under Grant No.SAST2021-007.
Author information
Authors and Affiliations
Contributions
Heng Tian: Methodology, original draft review and editing, experiment; Qiang Zhang: Validation, reviewing and proofreading; Zhe Wang: Project administration, reviewing and proofreading; Yu Zhang: Validation, investigation. Xinlei Xu and Zhiling Fu: Investigation and proofreading.
Corresponding authors
Ethics declarations
Competing Interests
We declare that the manuscript has been approved by all authors for publication, and no conflict of interest exists in the submission of it.
Ethical and Informed Consent for Data Used
Written informed consent for publication of this paper was obtained from the Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, and all authors.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Tian, H., Zhang, Q., Wang, Z. et al. Multi-view prototype balance and temporary proxy constraint for exemplar-free class-incremental learning. Appl Intell 55, 344 (2025). https://doi.org/10.1007/s10489-025-06233-7
Accepted:
Published:
DOI: https://doi.org/10.1007/s10489-025-06233-7