Skip to main content

Preliminary Study on Adapting ProtoPNet to Few-Shot Learning Using MAML

  • Conference paper
  • First Online:

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1628))

Abstract

ProtoPNet proposed by Chen et al. is able to provide interpretability that conforms to human intuition, but it requires many iterations of training to learn class-specific prototypes and does not support few-shot learning. We propose the few-shot learning version of ProtoPNet by using MAML, enabling it to converge quickly on different classification tasks. We test our model on the Omniglot and MiniImagenet datasets and evaluate their prototype interpretability. Our experiments show that MAML-ProtoPNet is a transparent model that can achieve or even exceed the baseline accuracy, and its prototype can learn class-specific features, which are consistent with our human recognition.

Yue Wang is the corresponding author of this paper (yuelwang@163.com). Yapu Zhao proposes the idea of combing MAML and ProtoPNet. And Yue Wang proposes the consistency of features (CoF) indicator in this paper. This work is supported by: National Defense Science and Technology Innovation Special Zone Project (No. 18-163-11-ZT-002-045-04); Engineering Research Center of State Financial Security, Ministry of Education, Central University of Finance and Economics, Beijing, 102206, China; Program for Innovation Research in Central University of Finance and Economics; National College Students’ Innovation and Entrepreneurship Training Program “Research and development of interpretable algorithms and prototype system for small sample image recognition”.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Wang, Y., Yao, Q., Kwok, J.T., Ni, L.M.: Generalizing from a few examples: a survey on few-shot learning. arXiv: Learning (2019)

    Google Scholar 

  2. Xue, Z., Duan, L., Li, W., Chen, L., Luo, J.: Region comparison network for interpretable few-shot image classification. arXiv preprint arXiv:2009.03558 (2020)

  3. Mehrotra, A., Dukkipati, A.: Generative adversarial residual pairwise networks for one shot learning. ArXiv, abs/1703.08033 (2017)

    Google Scholar 

  4. Luo, Z., Zou, Y., Hoffman, J., Fei-Fei, L.F.: Label efficient learning of transferable representations across domains and tasks. In: Advances in Neural Information Processing Systems, 30 (2017)

    Google Scholar 

  5. Gidaris, S., Komodakis, N. Dynamic few-shot visual learning without forgetting. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4367–4375 (2018)

    Google Scholar 

  6. Suárez, J.L., García, S., Herrera, F.: A tutorial on distance metric learning: mathematical foundations, algorithms, experimental analysis, prospects and challenges. Neurocomputing 425, 300–322 (2021)

    Article  Google Scholar 

  7. Snell, J., Swersky, K., Zemel, R.: Prototypical networks for few-shot learning. In: Advances in Neural Information Processing Systems, 30 (2017)

    Google Scholar 

  8. Chen, C., Li, O., Barnett, A., Su, J., Rudin, C. This looks like that: deep learning for interpretable image recognition. NeurIPS (2019)

    Google Scholar 

  9. Gao, T., Han, X., Liu, Z., Sun, M. Hybrid attention-based prototypical networks for noisy few-shot relation classification. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 1, pp. 6407–6414 (2019)

    Google Scholar 

  10. Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: ICML 2017, pp. 1126–1135 (2017)

    Google Scholar 

  11. Jiang, X., et al.: On the importance of attention in meta-learning for few-shot text classification. ArXiv, abs/1806.00852 (2018)

    Google Scholar 

  12. Jamal, M., Qi, G., Shah, M.: Task agnostic meta-learning for few-shot learning. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11711–11719 (2019)

    Google Scholar 

  13. Phillips, P.J., Hahn, C.A., Fontana, P.C., Broniatowski, D.A., Przybocki, M.A.: Four Principles of Explainable Artificial Intelligence (2020)

    Google Scholar 

  14. Bastani, O., Kim, C., Bastani, H.: Interpreting blackbox models via model extraction. ArXiv, abs/1705.08504 (2017)

    Google Scholar 

  15. Zhou, B., Bau, D., Oliva, A., Torralba, A.: Interpreting deep visual representations via network dissection. IEEE Trans. Pattern Anal. Mach. Intell. 41, 2131–2145 (2019)

    Article  Google Scholar 

  16. Cheng, X., Rao, Z., Chen, Y., Zhang, Q.: Explaining knowledge distillation by quantifying the knowledge. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12922–12932 (2020)

    Google Scholar 

  17. Zhang, Q., Yang, Y., Wu, Y.N., Zhu, S.: Interpreting CNNs via decision trees. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6254–6263 (2019)

    Google Scholar 

  18. Brahimi, M., Mahmoudi, S., Boukhalfa, K., Moussaoui, A. (2019, September). Deep interpretable architecture for plant diseases classification. In: 2019 Signal Processing: Algorithms, Architectures, Arrangements, and Applications (SPA), pp. 111–116. IEEE (2019)

    Google Scholar 

  19. Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1(5), 206–215 (2019)

    Article  Google Scholar 

  20. Qader, W.A., Ameen, M.M., Ahmed, B.I.: An overview of bag of words; importance, implementation, applications, and challenges. In: 2019 International Engineering Conference (IEC), pp. 200–204. IEEE (2019)

    Google Scholar 

  21. Melekhov, I., Kannala, J., Rahtu, E.: Siamese network features for image matching. In: 2016 23rd International Conference on Pattern Recognition (ICPR), pp. 378–383. IEEE (2016)

    Google Scholar 

  22. Lake, B.M., Salakhutdinov, R., Tenenbaum, J.B.: The Omniglot challenge: a 3-year progress report. Curr. Opin. Behav. Sci. 29, 97–104 (2019)

    Article  Google Scholar 

  23. Kulis, B.: Metric learning: a survey. Found. Trends Mach. Learn. 5(4), 287–364 (2013)

    Google Scholar 

  24. Vinyals, O., Blundell, C., Lillicrap, T., Wierstra, D.: Matching networks for one shot learning. In: Advances in Neural Information Processing Systems, 29 (2016)

    Google Scholar 

  25. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yue Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhao, Y., Wang, Y., Zhai, X. (2022). Preliminary Study on Adapting ProtoPNet to Few-Shot Learning Using MAML. In: Wang, Y., Zhu, G., Han, Q., Wang, H., Song, X., Lu, Z. (eds) Data Science. ICPCSEE 2022. Communications in Computer and Information Science, vol 1628. Springer, Singapore. https://doi.org/10.1007/978-981-19-5194-7_11

Download citation

  • DOI: https://doi.org/10.1007/978-981-19-5194-7_11

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-19-5193-0

  • Online ISBN: 978-981-19-5194-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics