Skip to main content
Log in

A PID controller method for meta-learning about aerospace target classification

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Traditional deep learning algorithms cannot accomplish few-shot aerospace targets classification task because the aerospace target classification dataset is scarcity. Meta-learning has strong generalization ability and quickly learn new tasks based on the acquisition of existing knowledge, which has greater advantages than traditional machine learning to solve few-shot aerospace targets classification task. Model-agnostic meta-learning (MAML) was the most classical optimization based meta-learning algorithm, while not easy to converge, and prone to oscillate at the end of training. Thus, we proposed a PID based meta-learning algorithm, and named as PIDMAML, which has good generalization and stable convergence. In the PIDMAML algorithm, the update of the model not only depends on the query set gradient (MAML), but the integration and differential of the support set gradient and query set gradient, which can retain more information on each task, and obtain a better initial model. The integration and differential of the support set gradient and query set gradient can accelerate model convergence and can apply to other MAML-based method. We also proposed an aerospace target dataset for few-shot aerospace target classification task. We verified the effectiveness of the PIDMAML algorithm on the Omniglot and MiniImagenet dataset, and the versatility of the PIDMAML on the Aerospace dataset. PIDMAML obtain much better results than classical meta-learning method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Algorithm 1
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Data availability statement

The Aerospace datasets generated for our paper can be requested from the corresponding author.

References

  1. Rezende D, Danihelka I, Gregor K, Wierstra D et al (2016) One-shot generalization in deep generative models. In: International conference on machine learning, pp 1521–1529. PMLR

  2. Santoro A, Bartunov S, Botvinick M, Wierstra D, Lillicrap T (2016) Meta-learning with memory-augmented neural networks. In: International conference on machine learning, pp 1842–1850. PMLR

  3. Munkhdalai T, Yu H (2017) Meta networks. In: International conference on machine learning, pp 2554–2563. PMLR

  4. Duan Y, Schulman J, Chen X, Bartlett PL, Sutskever I, Abbeel P (2016) R12: fast reinforcement learning via slow reinforcement learning. arXiv preprint arXiv:1611.02779

  5. Wang JX, Kurth-Nelson Z, Tirumala D, Soyer H, Leibo JZ, Munos R, Blundell C, Kumaran D, Botvinick M (2016) Learning to reinforcement learn. arXiv preprint arXiv:1611.05763

  6. Oreshkin BN, Rodriguez P, Lacoste A (2018) Tadam: task dependent adaptive metric for improved few-shot learning. arXiv preprint arXiv:1805.10123

  7. Snell J, Swersky K, Zemel RS (2017) Prototypical networks for few-shot learning. arXiv preprint arXiv:1703.05175

  8. Sung F, Yang Y, Zhang L, Xiang T, Torr PH, Hospedales TM (2018) Learning to compare: relation network for few-shot learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1199–1208

  9. Vinyals O, Blundell C, Lillicrap T, Wierstra D et al (2016) Matching networks for one shot learning. Adv Neural Inf Process Syst 29:3630–3638

    Google Scholar 

  10. Koch G, Zemel R, Salakhutdinov R et al (2015) Siamese neural networks for one-shot image recognition. In: ICML Deep learning workshop vol 2. Lille

  11. Shyam P, Gupta S, Dukkipati A (2017) Attentive recurrent comparators. In: International conference on machine learning, pp 3173–3181. PMLR

  12. Antoniou A, Edwards H, Storkey A (2018) How to train your maml. arXiv, 1–11 arXiv:1810.09502

  13. Finn C, Xu K, Levine S (2018) Probabilistic model-agnostic meta-learning. Adv Neural Inf Process Syst 2018-Decem:9516–9527

  14. Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. CoRR arXiv:1703.03400

  15. Yoon J, Kim T, Dia O, Kim S, Bengio Y, Ahn S (2018) Bayesian model-agnostic meta-learning. In: Proceedings of the 32nd international conference on neural information processing systems, pp 7343–7353

  16. Li Z, Zhou F, Chen F, Li H (2017) Meta-sgd: Learning to learn quickly for few shot learning. CoRR arXiv:1707.09835

  17. Rusu AA, Rao D, Sygnowski J, Vinyals O, Pascanu R, Osindero S, Hadsell R (2018) Meta-learning with latent embedding optimization. arXiv preprint arXiv:1807.05960

  18. Ravi S, Beatson A (2018) Amortized bayesian meta-learning. In: International conference on learning representations

  19. Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. 34th Int Conf Mach Learn, ICML 2017 3:1856–1868. arXiv:1703.03400

  20. Baik S, Hong S, Lee KM (2020) Learning to forget for meta-learning. In: CVPR

  21. Flennerhag S, Rusu AA, Pascanu R, Visin F, Yin H, Hadsell R (2019) Meta-learning with warped gradient descent. arXiv preprint arXiv:1909.00025

  22. Jamal MA, Qi G-J (2019) Task agnostic meta-learning for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11719–11727

  23. Jiang X, Havaei M, Varno F, Chartrand G, Chapados N, Matwin S (2018) Learning to learn with conditional class dependencies. In: International conference on learning representations

  24. Park E, Oliva JB (2019) Meta-curvature. arXiv preprint arXiv:1902.03356

  25. Raghu A, Raghu M, Bengio S, Vinyals O (2019) Rapid learning or feature reuse? towards understanding the effectiveness of maml. arXiv preprint arXiv:1909.09157

  26. Rajeswaran A, Finn C, Kakade S, Levine S (2019) Meta-learning with implicit gradients

  27. Triantafillou E, Zhu T, Dumoulin V, Lamblin P, Evci U, Xu K, Goroshin R, Gelada C, Swersky K, Manzagol P-A et al (2019) Metadataset: a dataset of datasets for learning to learn from few examples. arXiv preprint arXiv:1903.03096

  28. Vuorio R, Sun S-H, Hu H, Lim JJ (2019) Multimodal modelagnostic meta-learning via task-aware modulation. arXiv preprint arXiv:1910.13616

  29. Yao H, Wei Y, Huang J, Li Z (2019) Hierarchically structured metalearning. In: International conference on machine learning, pp 7045–7054. PMLR

  30. Yin M, Tucker G, Zhou M, Levine S, Finn C (2019) Meta-learning without memorization. arXiv preprint arXiv:1912.03820

  31. Zintgraf LM, Shiarlis K, Kurin V, Hofmann K, Whiteson S (2018) CAML: fast context adaptation via meta-learning. CoRR arXiv:1810.03642

  32. Ghiasi G, Lin T-Y, Le QV (2018) Dropblock: a regularization method for convolutional networks. arXiv preprint arXiv:1810.12890

  33. Larsson G, Maire M, Shakhnarovich G (2016) Fractalnet: ultra-deep neural networks without residuals. arXiv preprint arXiv:1605.07648

  34. Tompson J, Goroshin R, Jain A, LeCun Y, Bregler C (2015) Efficient object localization using convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 648–656

  35. Wan L, Zeiler M, Zhang S, Le Cun Y, Fergus R (2013) Regularization of neural networks using dropconnect. In: International conference on machine learning, pp 1058–1066. PMLR

  36. Goodfellow I, Warde-Farley D, Mirza M, Courville A, Bengio Y (2013) Maxout networks. In: International conference on machine learning, pp 1319–1327. PMLR

  37. Gupta G, Yadav K, Paull L (2020) La-MAML: look-ahead meta learning for continual learning. arXiv arXiv:2007.13904

  38. Zheng Y, Xiang J, Su K, Shlizerman E (2020) BI-MAML: balanced incremental approach for meta learning. CoRR arXiv:2006.07412

  39. Ased EN, Odels MEM (2020) Towards fast adaptation of neural architectures with meta learning (1):1–22

  40. Baik S, Choi M, Choi J, Kim H, Lee KM (2020) Meta-learning with adaptive hyperparameters. Adv Neural Inf Process Syst 33:20755–20765

    Google Scholar 

  41. Tseng HY, Chen YW, Tsai YH, Liu S, Lin YY, Yang MH (2021) Regularizing Meta-learning via Gradient Dropout. Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics) 12625 LNCS, 218–234 arXiv:2004.05859. https://doi.org/10.1007/978-3-030-69538-5_14

  42. Mandal D, Medya S, Uzzi B, Aggarwal C (2021) Meta-learning with graph neural networks: methods and applications arXiv:2103.00137

  43. An W, Wang H, Sun Q, Xu J, Dai Q, Zhang L (2018) A pid controller approach for stochastic optimization of deep networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8522–8531

  44. Baik S, Choi J, Kim H, Cho D, Min J, Lee KM (2021) Meta-learning with task-adaptive loss function for few-shot learning. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 9465–9474

Download references

Acknowledgements

This work was supported in part by the National Key R &D Program of China (No. 2019YFB1312001)

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhongliang Yu.

Ethics declarations

Conflict of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yu, Z., Lv, J. A PID controller method for meta-learning about aerospace target classification. Multimed Tools Appl 83, 39217–39233 (2024). https://doi.org/10.1007/s11042-023-17022-0

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-023-17022-0

Keywords