Skip to main content

Advertisement

Log in

Multi-view prototype balance and temporary proxy constraint for exemplar-free class-incremental learning

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Exemplar-free class-incremental learning recognizes both old and new classes without saving old class exemplars because of storage limitations and privacy constraints. To address the forgetting of knowledge caused by the absence of old training data, we present a novel method that consists of two modules, multi-view prototype balance and temporary proxy constraints, which are based on feature retention and representation optimization. Specifically, multi-view prototype balance first extends the prototypes to maintain the general state of the class and then balances these prototypes combining knowledge distillation and prototype compensation to ensure the stability and plasticity of the model. To alleviate the feature overlap, the proposed temporary proxy constraint sets the temporary proxies to lightly compress the feature distribution during each mini-batch of training. Extensive experiments on five datasets with different settings demonstrate the superiority of our method against the state-of-the-art exemplar-free class-incremental learning methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Algorithm 1
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data Availability and Access

The datasets generated during and/or analyzed during the current study are available in the [CIFAR-100 dataset] with [https://www.cs.toronto.edu/~kriz/cifar.html], [Tiny-ImageNet dataset] with [https://www.kaggle.com/c/tiny-imagenet], [ImageNet-Sub dataset] with [https://www.kaggle.com/datasets/ambityga/imagenet100], and [CUB dataset, ImageNet-R] with [https://github.com/zhoudw-zdw/RevisitingCIL].

References

  1. Touvron H, Bojanowski P, Caron M, Cord M, El-Nouby A, Grave E, Izacard G, Joulin A, Synnaeve G, Verbeek J et al (2022) Resmlp: Feedforward networks for image classification with data-efficient training. IEEE Trans Pattern Anal Mach Intell 45(4):5314–5321

  2. He T, Zhang Z, Zhang H, Zhang Z, Xie J, Li M (2019) Bag of tricks for image classification with convolutional neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 558–567

  3. Baek S-H, Heide F (2021) Polka lines: Learning structured illumination and reconstruction for active stereo. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 5757–5767

  4. Long T, Liang Z, Liu Q (2019) Advanced technology of high-resolution radar: target detection, tracking, imaging, and recognition. Sci China Inf Sci 62:1–26

    Article  MATH  Google Scholar 

  5. Devlin J, Chang M, Lee K, Toutanova K (2019) BERT: pre-training of deep bidirectional transformers for language understanding. In: Annual conference of the North American Chapter of the association for computational linguistics: human language technologies. Association for Computational Linguistics, pp 4171–4186

  6. Wu L, Chen Y, Shen K, Guo X, Gao H, Li S, Pei J, Long B et al (2023) Graph neural networks for natural language processing: a survey. Found Trends® Mach Learn 16(2):119–328

    Article  MATH  Google Scholar 

  7. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30

  8. Castro FM, Marín-Jiménez MJ, Guil N, Schmid C, Alahari K (2018) End-to-end incremental learning. In: Proceedings of the European conference on computer vision, pp 233–248

  9. Hou S, Pan X, Loy CC, Wang Z, Lin D (2019) Learning a unified classifier incrementally via rebalancing. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 831–839

  10. De Lange M, Aljundi R, Masana M, Parisot S, Jia X, Leonardis A, Slabaugh G, Tuytelaars T (2021) A continual learning survey: defying forgetting in classification tasks. IEEE Trans Pattern Anal Mach Intell 44(7):3366–3385

    Google Scholar 

  11. Sun L, Zhang M, Wang B, Tiwari P (2024) Few-shot class-incremental learning for medical time series classification. IEEE J Biomed Health Inform 28(4):1872–1882

    Article  MATH  Google Scholar 

  12. Singh T, Kalra R, Mishra S, Satakshi, Kumar M (2023) An efficient real-time stock prediction exploiting incremental learning and deep learning. Evol Syst 14(6):919–937

  13. Constantinides C, Shiaeles S, Ghita B, Kolokotronis N (2019) A novel online incremental learning intrusion prevention system. In: 2019 10th IFIP International conference on new technologies, mobility and security (NTMS), pp 1–6

  14. Rebuffi S-A, Kolesnikov A, Sperl G, Lampert CH (2017) icarl: Incremental classifier and representation learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2001–2010

  15. Shin H, Lee JK, Kim J, Kim J (2017) Continual learning with deep generative replay. Adv Neural Inf Process Syst 30

  16. Belouadah E, Popescu A (2019) Il2m: Class incremental learning with dual memory. In: Proceedings of the IEEE/CVF International conference on computer vision, pp 583–592

  17. Liu Y, Parisot S, Slabaugh G, Jia X, Leonardis A, Tuytelaars T (2020) More classifiers, less forgetting: a generic multi-classifier paradigm for incremental learning. In: Proceedings of the european conference on computer vision, pp 699–716

  18. Zhu F, Zhang X-Y, Wang C, Yin F, Liu C-L (2021) Prototype augmentation and self-supervision for incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 5871–5880

  19. Zhu K, Zhai W, Cao Y, Luo J, Zha Z-J (2022) Self-sustaining representation expansion for non-exemplar class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9296–9305

  20. Yu L, Twardowski B, Liu X, Herranz L, Wang K, Cheng Y, Jui S, Weijer JVD (2020) Semantic drift compensation for class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6982–6991

  21. Hinton G, Vinyals O, Dean J Distilling the knowledge in a neural network. arXiv:1503.02531

  22. Kim S, Kim D, Cho M, Kwak S (2020) Proxy anchor loss for deep metric learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3238–3247

  23. Lee H, Hwang SJ, Shin J (2020) Self-supervised label augmentation via input transformations. In: International conference on machine learning, pp 5714–5724

  24. Prideaux D, Ash J, Cottrell A (2013) Integrated learning. In: Oxford textbook of medical education. Oxford University Press Oxford, pp 63–73

  25. Li B, Yang Y, Zhao Z, Ni X, Zhang D (2024) A novel ensemble learning approach for intelligent logistics demand management. J Internet Technol 25(4):507–515

    Article  MATH  Google Scholar 

  26. Kirkpatrick J, Pascanu R, Rabinowitz N, Veness J, Desjardins G, Rusu AA, Milan K, Quan J, Ramalho T, Grabska-Barwinska A et al (2017) Overcoming catastrophic forgetting in neural networks. Proc Natl Acad Sci 114(13):3521–3526

    Article  MathSciNet  MATH  Google Scholar 

  27. Aljundi R, Babiloni F, Elhoseiny M, Rohrbach M, Tuytelaars T (2018) Memory aware synapses: learning what (not) to forget. In: Proceedings of the European conference on computer vision, pp 139–154

  28. Yoon J, Yang E, Lee J, Hwang SJ (2018) Lifelong learning with dynamically expandable networks. In: International conference on learning representations

  29. Yan S, Xie J, He X (2021) Der: Dynamically expandable representation for class incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3014–3023

  30. Fu Z, Wang Z, Xu X, Li D, Yang H (2023) Knowledge aggregation networks for class incremental learning. Pattern Recognit 137:109310

    Article  MATH  Google Scholar 

  31. Lin H, Zhang Y, Qiu Z, Niu S, Gan C, Liu Y, Tan M (2022) Prototype-guided continual adaptation for class-incremental unsupervised domain adaptation. In: Proceedings of the European conference on computer vision, pp 351–368

  32. Zhu F, Cheng Z, Zhang X-Y, Liu C-L (2021) Class-incremental learning via dual augmentation. Adv Neural Inf Process Syst 34:14306–14318

    MATH  Google Scholar 

  33. Petit G, Popescu A, Schindler H, Picard D, Delezoide B (2023) Fetril: Feature translation for exemplar-free class-incremental learning. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp 3911–3920

  34. Li Q, Peng Y, Zhou J (2024) Fcs: Feature calibration and separation for non-exemplar class incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 28495–28504

  35. Peyré G, Cuturi M et al (2019) Computational optimal transport: with applications to data science. Found Trends® Mach Learn 11(5–6):355–607

    Article  MATH  Google Scholar 

  36. Kaya M, Bilge HŞ (2019) Deep metric learning: a survey. Symmetry 11(9):1066

    Article  MATH  Google Scholar 

  37. Li X, Yang X, Ma Z, Xue J-H (2023) Deep metric learning for few-shot image classification: a review of recent developments. Pattern Recognit 109381

  38. Schroff F, Kalenichenko D, Philbin J (2015) Facenet: A unified embedding for face recognition and clustering. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 815–823

  39. Movshovitz-Attias Y, Toshev A, Leung TK, Ioffe S, Singh S (2017) No fuss distance metric learning using proxies. In: Proceedings of the IEEE international conference on computer vision, pp 360–368

  40. Wang J, Zhou F, Wen S, Liu X, Lin Y (2017) Deep metric learning with angular loss. In: Proceedings of the IEEE international conference on computer vision, pp 2593–2601

  41. Yu B, Tao D (2019) Deep metric learning with tuplet margin loss. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 6490–6499

  42. Elezi I, Vascon S, Torcinovich A, Pelillo M, Leal-Taixé L (2020) The group loss for deep metric learning. In: Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part VII 16. Springer, pp 277–294

  43. Krizhevsky A, Hinton G et al (2009) Learning multiple layers of features from tiny images

  44. Yang K, Yau JH, Fei-Fei L, Deng J, Russakovsky O (2022) A study of face obfuscation in imagenet. In: International conference on machine learning, pp 25313–25330

  45. Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset

  46. Hendrycks D, Basart S, Mu N, Kadavath S, Wang F, Dorundo E, Desai R, Zhu T, Parajuli S, Guo M, et al (2021) The many faces of robustness: a critical analysis of out-of-distribution generalization. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 8340–8349

  47. Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M et al (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115:211–252

    Article  MathSciNet  Google Scholar 

  48. Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp 248–255

  49. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 770–778

  50. Kingma DP, Ba J (2015) Adam: A method for stochastic optimization. In: International conference on learning representations

  51. Li Z, Hoiem D (2017) Learning without forgetting. IEEE Trans Pattern Anal Mach Intell 40(12):2935–2947

    Article  Google Scholar 

  52. Shi Y, Shi D, Qiao Z, Wang Z, Zhang Y, Yang S, Qiu C (2023) Multi-granularity knowledge distillation and prototype consistency regularization for class-incremental learning. Neural Netw 164:617–630

    Article  MATH  Google Scholar 

  53. Chatzimparmpas A, Martins RM, Kerren A (2020) t-visne: Interactive assessment and interpretation of t-sne projections. IEEE Trans Vis Comput Graph 26(8):2696–2714

    Article  MATH  Google Scholar 

  54. Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, Sastry G, Askell A, Mishkin P, Clark J et al (2021) Learning transferable visual models from natural language supervision. In: International conference on machine learning, pp 8748–8763

  55. Shorten C, Khoshgoftaar TM (2019) A survey on image data augmentation for deep learning. J Big Data 6(1):1–48

    Article  MATH  Google Scholar 

  56. Feng SY, Gangal V, Wei J, Chandar S, Vosoughi S, Mitamura T, Hovy EH (2021) A survey of data augmentation approaches for NLP. In: International joint conference on natural language processing, Vol. ACL/IJCNLP 2021 of Findings of ACL. Association for Computational Linguistics, pp 968–988

  57. Yun S, Han D, Oh SJ, Chun S, Choe J, Yoo Y (2019) Cutmix: Regularization strategy to train strong classifiers with localizable features. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 6023–6032

  58. Zhang H, Cissé M, Dauphin, Lopez-Paz D (2018) mixup: Beyond empirical risk minimization. In: International conference on learning representations. OpenReview.net

  59. DeVries T, Taylor GW (2017) Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552

Download references

Acknowledgements

This work is supported by Natural Science Foundation of China under Grant No. 62476087, the National Key Research and Development Program of China (2022YFB3203500), Shanghai Science and Technology Program “Federated based cross-domain and cross-task incremental learning” under Grant No. 21511100800, Natural Science Foundation of China under Grant No. 62002193, Chinese Defense Program of Science and Technology under Grant No.2021-JCJQ-JJ-0041, China Aerospace Science and Technology Corporation Industry-University-Research Cooperation Foundation of the Eighth Research Institute under Grant No.SAST2021-007.

Author information

Authors and Affiliations

Authors

Contributions

Heng Tian: Methodology, original draft review and editing, experiment; Qiang Zhang: Validation, reviewing and proofreading; Zhe Wang: Project administration, reviewing and proofreading; Yu Zhang: Validation, investigation. Xinlei Xu and Zhiling Fu: Investigation and proofreading.

Corresponding authors

Correspondence to Qian Zhang or Zhe Wang.

Ethics declarations

Competing Interests

We declare that the manuscript has been approved by all authors for publication, and no conflict of interest exists in the submission of it.

Ethical and Informed Consent for Data Used

Written informed consent for publication of this paper was obtained from the Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, and all authors.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Tian, H., Zhang, Q., Wang, Z. et al. Multi-view prototype balance and temporary proxy constraint for exemplar-free class-incremental learning. Appl Intell 55, 344 (2025). https://doi.org/10.1007/s10489-025-06233-7

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10489-025-06233-7

Keywords