Skip to main content

Advertisement

Log in

Concept agent network for zero-base generalized few-shot learning

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Generalized Few-Shot Learning (GFSL) aims to recognize novel classes with limited training samples without forgetting knowledge of auxiliary data (base classes). Most current approaches re-engage the base classes after initial training to balance the predictive bias between the base and novel classes. However, re-using the auxiliary data might not always be possible due to privacy or ethical constraints. Consequently, the zero-base GFSL paradigm emerges, where models trained on the base classes are directly fine-tuned on the novel classes without revisiting the auxiliary data, avoiding the re-balancing of prediction biases. We believe that solving this paradigm relies on a critical yet often overlooked issue: feature overlap between the base and novel classes in the embedding space. To tackle this issue, we propose the Concept Agent Network, a novel framework that interprets visual features as affinity features, thereby effectively diminishing feature overlap by aggregating feature embeddings of the novel classes according to their similarity with the base classes. Additionally, we present the Concept Catena Generator, which creates multiple concepts per base class, improving understanding of the feature distribution of the base classes and clarifying the relationships between the base and novel concepts. To prevent the catastrophic forgetting of the base classes when adapting to the novel ones, we propose an Active Training Regularization strategy, promoting the preservation of base class knowledge. Extensive experimental results on two benchmarks, mini-ImageNet and tiered-ImageNet, have demonstrated the effectiveness of our framework. The potential utility of our framework spans several real-world applications, including autonomous driving, medical image analysis, and real-time surveillance, where the ability to rapidly learn from a few examples without forgetting previously acquired knowledge is critical.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Algorithm 1
Algorithm 2
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Explore related subjects

Discover the latest articles and news from researchers in related subjects, suggested using machine learning.

Data Availability Statement

The data that support the findings of this study are openly available at https://lyy.mpi-inf.mpg.de/mtl/download/Lmzjm9tX.html

References

  1. Ji Z, An P, Liu X, Gao C, Pang Y, Shao L (2023) Semantic-aware dynamic generation networks for few-shot human-object interaction recognition. IEEE Trans Neural Netw Learn Syst 1–12

  2. Ji Z, Hou Z, Liu X, Pang Y, Han J (2022) Information symmetry matters: a modal-alternating propagation network for few-shot learning. IEEE Trans Image Process 31:1520–1531

    Article  MATH  Google Scholar 

  3. Lim JY, Lim KM, Lee CP, Tan YX (2023) Scl: Self-supervised contrastive learning for few-shot image classification. Neural Netw 165:19–30

    Article  MATH  Google Scholar 

  4. Liu X, Ji Z, Pang Y, Han Z (2023) Dual distillation discriminator networks for domain adaptive few-shot learning. Neural Netw 165:625–633

    Article  MATH  Google Scholar 

  5. He K, Pu N, Lao M, Bakker EM, Lew MS (2023) Dual selective knowledge transfer for few-shot classification. Appl Intell 53(22):27779–27789

    Article  Google Scholar 

  6. Yoon SW, Kim D-Y, Seo J, Moon J (2020) Xtarnet: Learning to extract task-adaptive representation for incremental few-shot learning. In: Proceedings of the international conference on machine learning, pp 10852–10860

  7. Kim S-W, Choi D-W (2023) Better generalized few-shot learning even without base data. In: Proceedings of the AAAI conference on artificial intelligence, vol 37, pp 8282–8290

  8. Radhika T, Chandrasekar A, Vijayakumar V, Zhu Q (2023) Analysis of markovian jump stochastic cohen-grossberg bam neural networks with time delays for exponential input-to-state stability. Neural Process Lett 55(8):11055–11072

    Article  MATH  Google Scholar 

  9. Cao Y, Chandrasekar A, Radhika T, Vijayakumar V (2024) Input-to-state stability of stochastic markovian jump genetic regulatory networks. Math Comput Simul 222:174–187

    Article  MathSciNet  MATH  Google Scholar 

  10. Gidaris S, Komodakis N (2018) Dynamic few-shot visual learning without forgetting. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4367–4375

  11. Ren M, Liao R, Fetaya E, Zemel R (2019) Incremental few-shot learning with attention attractor networks. In: Advances in neural information processing systems, pp 5275–5285

  12. Kemker R, McClure M, Abitino A, Hayes T, Kanan C (2018) Measuring catastrophic forgetting in neural networks. In: Proceedings of the AAAI conference on artificial intelligence, pp 3390–3398

  13. Shmelkov K, Schmid C, Alahari K (2017) Incremental learning of object detectors without catastrophic forgetting. In: Proceedings of the IEEE international conference on computer vision, pp 3400–3409

  14. Lee K, Lee K, Shin J, Lee H (2019) Overcoming catastrophic forgetting with unlabeled data in the wild. In: Proceedings of the IEEE international conference on computer vision, pp 312–321

  15. Ge L, Hou Y, Jiang Y, Yao S, Yang C (2020) Veca: A method for detecting overfitting in neural networks. In: Proceedings of the AAAI conference on artificial intelligence, pp 13791–13792

  16. Jiang M, Li F, Liu L (2022) Continual meta-learning algorithm. Appl Intell 52(4):4527–4542

    Article  MATH  Google Scholar 

  17. Fan W, Huang H, Liang C, Liu X, Peng S-J (2023) Unsupervised meta-learning via spherical latent representations and dual vae-gan. Appl Intell 53(19):22775–22788

    Article  Google Scholar 

  18. Zhang B, Li X, Ye Y, Huang Z, Zhang L (2021) Prototype completion with primitive knowledge for few-shot learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3754–3762

  19. Zheng Z, Feng X, Yu H, Li X, Gao M (2023) Bdla: Bi-directional local alignment for few-shot learning. Appl Intell 53(1):769–785

    Article  MATH  Google Scholar 

  20. Shi X, Salewski L, Schiegg M, Akata Z, Welling M (2019) Relational generalized few-shot learning. arXiv:1907.09557

  21. Akyürek AF, Akyürek E, Wijaya D, Andreas J (2022) Subspace regularizers for few-shot class incremental learning. In: Proceedings of the conference on international conference on learning representations, pp 1–10

  22. Zemel RS, Mozer MC (2001) Localist attractor networks. Neural Comput 13(5):1045–1064

    Article  MATH  Google Scholar 

  23. Cheraghian A, Rahman S, Fang P, Roy SK, Petersson L, Harandi M (2021) Semantic-aware knowledge distillation for few-shot class-incremental learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2534–2543

  24. Kukleva A, Kuehne H, Schiele B (2021) Generalized and incremental few-shot learning by explicit learning and calibration without forgetting. In: Proceedings of the IEEE international conference on computer vision, pp 9020–9029

  25. Qi H, Brown M, Lowe DG (2018) Low-shot learning with imprinted weights. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5822–5830

  26. Tao X, Hong X, Chang X, Dong S, Wei X, Gong Y (2020) Few-shot class-incremental learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 12183–12192

  27. Wang X, Ji Z, Pang Y, Yu Y (2024) A cognition-driven framework for few-shot class-incremental learning. Neurocomputing 600:128118

    Article  MATH  Google Scholar 

  28. Li J, Dong S, Gong Y, He Y, Wei X (2024) Analogical learning-based few-shot class-incremental learning. IEEE Trans Circuits Syst Video Technol 34(7):5493–5504

    Article  Google Scholar 

  29. Wang QW, Zhou DW, Zhang YK, Zhan DC, Ye HJ (2023) Few-shot class-incremental learning via training-free prototype calibration. Adv Neural Inf Process Syst 36:15060–15076

    Google Scholar 

  30. Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M et al (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115(3):211–252

    Article  MathSciNet  Google Scholar 

  31. Ren M, Triantafillou E, Ravi S, Snell J, Swersky K, Tenenbaum JB, Larochelle H, Zemel RS (2018) Meta-learning for semi-supervised few-shot classification. arXiv:1803.00676

  32. Zhang C, Song N, Lin G, Zheng Y, Pan P, Xu Y (2021) Few-shot incremental learning with continually evolved classifiers. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 12455–12464

  33. Krizhevsky A, Hinton G et al (2009) Learning multiple layers of features from tiny images

  34. Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset

  35. Mazumder P, Singh P, Rai P (2021) Few-shot lifelong learning. In: Proceedings of the AAAI conference on artificial intelligence, vol 35, pp 2337–2345

  36. Wang X, Ji Z, Liu X, Pang Y, Han J (2025) On the approximation risk of few-shot class-incremental learning. In: European conference on computer vision. Springer, pp 162–178

  37. Wang X, Ji Z, Yu Y, Pang Y, Han J (2024) Model attention expansion for few-shot class-incremental learning. IEEE Trans Image Process 33:4419–4431

    Article  MATH  Google Scholar 

  38. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778

  39. Snell J, Swersky K, Zemel R (2017) Prototypical networks for few-shot learning. In: Advances in neural information processing systems, pp 1–11

  40. Ji Z, Hou Z, Liu X, Pang Y, Li X (2023) Memorizing complementation network for few-shot class-incremental learning. IEEE Trans Image Process 32:937–948

    Article  MATH  Google Scholar 

  41. Zhou D-W, Wang F-Y, Ye H-J, Ma L, Pu S, Zhan D-C (2022) Forward compatible few-shot class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9046–9056

  42. Peng C, Zhao K, Wang T, Li M, Lovell BC (2022) Few-shot class-incremental learning from an open-set perspective. In: Proceedings of the European Conference on Computer Vision (ECCV). Springer, pp 382–397

  43. Song Z, Zhao Y, Shi Y, Peng P, Yuan L, Tian Y (2023) Learning with fantasy: semantic-aware virtual contrastive constraint for few-shot class-incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp 24183–24192

  44. Yang Y, Yuan H, Li X, Lin Z, Torr P, Tao D (2023) Neural collapse inspired feature-classifier alignment for few-shot class-incremental learning. In: International Conference on Learning Representations (ICLR), pp 1–13

  45. Huang W, Hu Y, Hu S, Liu J (2021) Learning a linear classifier by transforming feature vectors for few-shot image classification. In: Proceedings of the international conference on control, robotics and intelligent system, pp 158–162

  46. Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: Teh YW, Titterington M (eds) Proceedings of the thirteenth international conference on artificial intelligence and statistics. Proceedings of machine learning research, vol 9, pp 249–256. PMLR, Chia Laguna Resort, Sardinia, Italy

  47. He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV)

Download references

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant 2022ZD0160403 and in part by the National Natural Science Foundation of China (NSFC) under Grant 62176178.

Author information

Authors and Affiliations

Authors

Contributions

Xuan Wang: Conceptualization, Methodology, Writing, Software. Zhong Ji: Conceptualization, Supervision, Writing, Funding acquisition. Xiyao Liu: Methodology, Writing. Yanwei Pang: Methodology, Writing – review & Editing. Xuelong Li: Conceptualization, Supervision, Writing – review & Editing.

Corresponding author

Correspondence to Zhong Ji.

Ethics declarations

Competing Interests

The authors have no competing interests to declare that are relevant to the content of this article.

Ethics Approval

Not applicable.

Consent to Participate

Not applicable.

Consent for Publication

Not applicable.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, X., Ji, Z., Liu, X. et al. Concept agent network for zero-base generalized few-shot learning. Appl Intell 55, 414 (2025). https://doi.org/10.1007/s10489-025-06331-6

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10489-025-06331-6

Keywords