Abstract
Generalized Few-Shot Learning (GFSL) aims to recognize novel classes with limited training samples without forgetting knowledge of auxiliary data (base classes). Most current approaches re-engage the base classes after initial training to balance the predictive bias between the base and novel classes. However, re-using the auxiliary data might not always be possible due to privacy or ethical constraints. Consequently, the zero-base GFSL paradigm emerges, where models trained on the base classes are directly fine-tuned on the novel classes without revisiting the auxiliary data, avoiding the re-balancing of prediction biases. We believe that solving this paradigm relies on a critical yet often overlooked issue: feature overlap between the base and novel classes in the embedding space. To tackle this issue, we propose the Concept Agent Network, a novel framework that interprets visual features as affinity features, thereby effectively diminishing feature overlap by aggregating feature embeddings of the novel classes according to their similarity with the base classes. Additionally, we present the Concept Catena Generator, which creates multiple concepts per base class, improving understanding of the feature distribution of the base classes and clarifying the relationships between the base and novel concepts. To prevent the catastrophic forgetting of the base classes when adapting to the novel ones, we propose an Active Training Regularization strategy, promoting the preservation of base class knowledge. Extensive experimental results on two benchmarks, mini-ImageNet and tiered-ImageNet, have demonstrated the effectiveness of our framework. The potential utility of our framework spans several real-world applications, including autonomous driving, medical image analysis, and real-time surveillance, where the ability to rapidly learn from a few examples without forgetting previously acquired knowledge is critical.









Similar content being viewed by others
Explore related subjects
Discover the latest articles and news from researchers in related subjects, suggested using machine learning.Data Availability Statement
The data that support the findings of this study are openly available at https://lyy.mpi-inf.mpg.de/mtl/download/Lmzjm9tX.html
References
Ji Z, An P, Liu X, Gao C, Pang Y, Shao L (2023) Semantic-aware dynamic generation networks for few-shot human-object interaction recognition. IEEE Trans Neural Netw Learn Syst 1–12
Ji Z, Hou Z, Liu X, Pang Y, Han J (2022) Information symmetry matters: a modal-alternating propagation network for few-shot learning. IEEE Trans Image Process 31:1520–1531
Lim JY, Lim KM, Lee CP, Tan YX (2023) Scl: Self-supervised contrastive learning for few-shot image classification. Neural Netw 165:19–30
Liu X, Ji Z, Pang Y, Han Z (2023) Dual distillation discriminator networks for domain adaptive few-shot learning. Neural Netw 165:625–633
He K, Pu N, Lao M, Bakker EM, Lew MS (2023) Dual selective knowledge transfer for few-shot classification. Appl Intell 53(22):27779–27789
Yoon SW, Kim D-Y, Seo J, Moon J (2020) Xtarnet: Learning to extract task-adaptive representation for incremental few-shot learning. In: Proceedings of the international conference on machine learning, pp 10852–10860
Kim S-W, Choi D-W (2023) Better generalized few-shot learning even without base data. In: Proceedings of the AAAI conference on artificial intelligence, vol 37, pp 8282–8290
Radhika T, Chandrasekar A, Vijayakumar V, Zhu Q (2023) Analysis of markovian jump stochastic cohen-grossberg bam neural networks with time delays for exponential input-to-state stability. Neural Process Lett 55(8):11055–11072
Cao Y, Chandrasekar A, Radhika T, Vijayakumar V (2024) Input-to-state stability of stochastic markovian jump genetic regulatory networks. Math Comput Simul 222:174–187
Gidaris S, Komodakis N (2018) Dynamic few-shot visual learning without forgetting. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4367–4375
Ren M, Liao R, Fetaya E, Zemel R (2019) Incremental few-shot learning with attention attractor networks. In: Advances in neural information processing systems, pp 5275–5285
Kemker R, McClure M, Abitino A, Hayes T, Kanan C (2018) Measuring catastrophic forgetting in neural networks. In: Proceedings of the AAAI conference on artificial intelligence, pp 3390–3398
Shmelkov K, Schmid C, Alahari K (2017) Incremental learning of object detectors without catastrophic forgetting. In: Proceedings of the IEEE international conference on computer vision, pp 3400–3409
Lee K, Lee K, Shin J, Lee H (2019) Overcoming catastrophic forgetting with unlabeled data in the wild. In: Proceedings of the IEEE international conference on computer vision, pp 312–321
Ge L, Hou Y, Jiang Y, Yao S, Yang C (2020) Veca: A method for detecting overfitting in neural networks. In: Proceedings of the AAAI conference on artificial intelligence, pp 13791–13792
Jiang M, Li F, Liu L (2022) Continual meta-learning algorithm. Appl Intell 52(4):4527–4542
Fan W, Huang H, Liang C, Liu X, Peng S-J (2023) Unsupervised meta-learning via spherical latent representations and dual vae-gan. Appl Intell 53(19):22775–22788
Zhang B, Li X, Ye Y, Huang Z, Zhang L (2021) Prototype completion with primitive knowledge for few-shot learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3754–3762
Zheng Z, Feng X, Yu H, Li X, Gao M (2023) Bdla: Bi-directional local alignment for few-shot learning. Appl Intell 53(1):769–785
Shi X, Salewski L, Schiegg M, Akata Z, Welling M (2019) Relational generalized few-shot learning. arXiv:1907.09557
Akyürek AF, Akyürek E, Wijaya D, Andreas J (2022) Subspace regularizers for few-shot class incremental learning. In: Proceedings of the conference on international conference on learning representations, pp 1–10
Zemel RS, Mozer MC (2001) Localist attractor networks. Neural Comput 13(5):1045–1064
Cheraghian A, Rahman S, Fang P, Roy SK, Petersson L, Harandi M (2021) Semantic-aware knowledge distillation for few-shot class-incremental learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2534–2543
Kukleva A, Kuehne H, Schiele B (2021) Generalized and incremental few-shot learning by explicit learning and calibration without forgetting. In: Proceedings of the IEEE international conference on computer vision, pp 9020–9029
Qi H, Brown M, Lowe DG (2018) Low-shot learning with imprinted weights. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5822–5830
Tao X, Hong X, Chang X, Dong S, Wei X, Gong Y (2020) Few-shot class-incremental learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 12183–12192
Wang X, Ji Z, Pang Y, Yu Y (2024) A cognition-driven framework for few-shot class-incremental learning. Neurocomputing 600:128118
Li J, Dong S, Gong Y, He Y, Wei X (2024) Analogical learning-based few-shot class-incremental learning. IEEE Trans Circuits Syst Video Technol 34(7):5493–5504
Wang QW, Zhou DW, Zhang YK, Zhan DC, Ye HJ (2023) Few-shot class-incremental learning via training-free prototype calibration. Adv Neural Inf Process Syst 36:15060–15076
Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M et al (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115(3):211–252
Ren M, Triantafillou E, Ravi S, Snell J, Swersky K, Tenenbaum JB, Larochelle H, Zemel RS (2018) Meta-learning for semi-supervised few-shot classification. arXiv:1803.00676
Zhang C, Song N, Lin G, Zheng Y, Pan P, Xu Y (2021) Few-shot incremental learning with continually evolved classifiers. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 12455–12464
Krizhevsky A, Hinton G et al (2009) Learning multiple layers of features from tiny images
Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset
Mazumder P, Singh P, Rai P (2021) Few-shot lifelong learning. In: Proceedings of the AAAI conference on artificial intelligence, vol 35, pp 2337–2345
Wang X, Ji Z, Liu X, Pang Y, Han J (2025) On the approximation risk of few-shot class-incremental learning. In: European conference on computer vision. Springer, pp 162–178
Wang X, Ji Z, Yu Y, Pang Y, Han J (2024) Model attention expansion for few-shot class-incremental learning. IEEE Trans Image Process 33:4419–4431
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Snell J, Swersky K, Zemel R (2017) Prototypical networks for few-shot learning. In: Advances in neural information processing systems, pp 1–11
Ji Z, Hou Z, Liu X, Pang Y, Li X (2023) Memorizing complementation network for few-shot class-incremental learning. IEEE Trans Image Process 32:937–948
Zhou D-W, Wang F-Y, Ye H-J, Ma L, Pu S, Zhan D-C (2022) Forward compatible few-shot class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9046–9056
Peng C, Zhao K, Wang T, Li M, Lovell BC (2022) Few-shot class-incremental learning from an open-set perspective. In: Proceedings of the European Conference on Computer Vision (ECCV). Springer, pp 382–397
Song Z, Zhao Y, Shi Y, Peng P, Yuan L, Tian Y (2023) Learning with fantasy: semantic-aware virtual contrastive constraint for few-shot class-incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 24183–24192
Yang Y, Yuan H, Li X, Lin Z, Torr P, Tao D (2023) Neural collapse inspired feature-classifier alignment for few-shot class-incremental learning. In: International Conference on Learning Representations (ICLR), pp 1–13
Huang W, Hu Y, Hu S, Liu J (2021) Learning a linear classifier by transforming feature vectors for few-shot image classification. In: Proceedings of the international conference on control, robotics and intelligent system, pp 158–162
Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: Teh YW, Titterington M (eds) Proceedings of the thirteenth international conference on artificial intelligence and statistics. Proceedings of machine learning research, vol 9, pp 249–256. PMLR, Chia Laguna Resort, Sardinia, Italy
He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV)
Funding
This work was supported in part by the National Key Research and Development Program of China under Grant 2022ZD0160403 and in part by the National Natural Science Foundation of China (NSFC) under Grant 62176178.
Author information
Authors and Affiliations
Contributions
Xuan Wang: Conceptualization, Methodology, Writing, Software. Zhong Ji: Conceptualization, Supervision, Writing, Funding acquisition. Xiyao Liu: Methodology, Writing. Yanwei Pang: Methodology, Writing – review & Editing. Xuelong Li: Conceptualization, Supervision, Writing – review & Editing.
Corresponding author
Ethics declarations
Competing Interests
The authors have no competing interests to declare that are relevant to the content of this article.
Ethics Approval
Not applicable.
Consent to Participate
Not applicable.
Consent for Publication
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, X., Ji, Z., Liu, X. et al. Concept agent network for zero-base generalized few-shot learning. Appl Intell 55, 414 (2025). https://doi.org/10.1007/s10489-025-06331-6
Accepted:
Published:
DOI: https://doi.org/10.1007/s10489-025-06331-6