Abstract
When used for small-scale datasets classification tasks, deep neural networks are difficult to train, which results in the network not extracting useful features and the low accuracy of network. This paper proposes a method called Enhanced Adversarial Autoencoders (EAAE) to improve the accuracy of classification in small-scale datasets scenarios. We reuse the decoder of the Adversarial Autoencoders to generate data for the specified category, which expands the training space of the samples. In order to capture the features of the generated data, we also reuse the encoder to generate the corresponding latent label variables and train them in a supervised manner. To maintain and improve the generative power of the decoder, we impose conditional adversarial training on the generated data. Finally, we use the encoder with latent feature extraction capability to complete the classification by transfer learning. To evaluate the performance of the proposed method, we validate on a real Gastric Cancer dataset and five widely used UCI public datasets. Extensive experiments show that EAAE improves significantly in classification accuracy on small-scale datasets and performs better on large-scale datasets compared to popular classification methods. Furthermore, we evaluate the time cost of EAAE and conduct ablation experiments to demonstrate the effectiveness of the various modules of EAAE.
Similar content being viewed by others
References
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444
Shinde PP, Shah S (2018) A review of machine learning and deep learning applications. In: Proceedings of the Fourth International Conference on Computing Communication Control and Automation, pp 1–6
Pandey SK, Janghel RR (2019) Recent deep learning techniques, challenges and its applications for medical healthcare system: a review. Neural Process Lett 50(2):1907–1935
Yu J, Tan M, Zhang H et al (2019) Hierarchical deep click feature prediction for fine-grained image recognition. IEEE Trans Pattern Anal Mach Intell 44(2):563–578
Yu J, Zhu C, Zhang J et al (2019) Spatial pyramid-enhanced netvlad with weighted triplet loss for place recognition. IEEE Transactions on Neural Networks and Learning Systems 31(2):661–674
Zhao ZQ, Zheng P, Xu S et al (2019) Object detection with deep learning: A review. IEEE Transactions on Neural Networks and Learning Systems 30(11):3212–3232
Ayachi R, Afif M, Said Y et al (2020) Traffic signs detection for realworld application of an advanced driving assisting system using deep learning. Neural Process Lett 51(1):837–851
Hong C, Yu J, Zhang J et al (2018) Multimodal face-pose estimation with multitask manifold deep learning. IEEE Trans Industr Inf 15(7):3952–3961
Wang P, Di J (2018) Deep learning-based object classification through multimode fiber via a cnn-architecture specklenet. Appl Opt 57(28):8258–8263
Sharma S, Rana V, Kumar V (2021) Deep learning based semantic personalized recommendation system. International Journal of Information Management Data Insights 1(2):100028
Choi H, Cho K, Bengio Y (2018) Fine-grained attention mechanism for neural machine translation. Neurocomputing 284:171–176
Sun C, Shrivastava A, Singh S, et al (2017) Revisiting unreasonable effectiveness of data in deep learning era. In: Proceedings of the IEEE International Conference on Computer Vision, pp 843–852
Wang Y, Yao Q, Kwok JT et al (2020) Generalizing from a few examples: A survey on few-shot learning. ACM Computing Surveys (CSUR) 53(3):1–34
Tan C, Sun F, Kong T, et al (2018) A survey on deep transfer learning. In: Proceedings of the Conference on Artificial Neural Networks, pp 270–279
He Y, Ding G (2020) Deep transfer learning for image emotion analysis: Reducing marginal and joint distribution discrepancies together. Neural Process Lett 51(3):2077–2086
Hospedales T, Antoniou A, Micaelli P, et al (2021) Meta-learning in neural networks: A survey. arXiv preprint arXiv:2004.05439
Snell J, Swersky K, Zemel R (2017) Prototypical networks for few-shot learning. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp 4080–4090
Wang YX, Hebert M (2016) Learning to learn: Model regression networks for easy small sample learning. In: Proceedings of the European Conference on Computer Vision, pp 616–634
Salamon J, Bello JP (2017) Deep convolutional neural networks and data augmentation for environmental sound classification. IEEE Signal Process Lett 24(3):279–283
Cubuk ED, Zoph B, Mane D, et al (2019) AutoAugment: Learning augmentation strategies from data. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 113–123
Sung F, Yang Y, Zhang L, et al (2018) Learning to compare: Relation network for few-shot learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1199–1208
Gao R, Hou X, Qin J et al (2020) Zero-VAE-GAN: Generating unseen features for generalized and transductive zero-shot learning. IEEE Trans Image Process 29:3665–3680
Hong Y, Niu L, Zhang J, et al (2020) F2GAN: Fusing-and-filling gan for few-shot image generation. In: Proceedings of the 28th ACM International Conference on Multimedia, pp 2535–2543
Hong Y, Niu L, Zhang J, et al (2020) MatchingGAN: Matching-based few-shot image generation. In: Proceedings of the IEEE International Conference on Multimedia and Expo, pp 1–6
Goodfellow IJ, Pouget-Abadie J, Mirza M, et al (2014) Generative adversarial nets. In: Proceedings of the Conference on Neural Information Processing Systems, pp 2672–2680
Larsen ABL, Sønderby SK, Larochelle H, et al (2016) Autoencoding beyond pixels using a learned similarity metric. In: Proceedings of the Conference on Machine Learning, pp 1558–1566
Kim JY, Cho SB (2021) Deep cnn transferred from vae and gan for classifying irritating noise in automobile. Neurocomputing 452:395–403
Yu S, Principe JC (2019) Understanding autoencoders with information theoretic concepts. Neural Netw 117:104–123
Kingma DP, Welling M (2019) An introduction to variational autoencoders. Foundations and Trends in Machine Learning 12(4):307–392
Hoffman MD, Blei DM, Wang C et al (2013) Stochastic variational inference. J Mach Learn Res 14(1):1303–1347
Mirza M, Osindero S (2014) Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784
Madani A, Moradi M, Karargyris A, et al (2018) Semi-supervised learning with generative adversarial networks for chest x-ray classification with ability of data domain adaptation. In: Proceedings of the International Symposium on Biomedical Imaging, pp 1038–1042
Makhzani A, Shlens J, Jaitly N, et al (2016) Adversarial autoencoders. arXiv preprint arXiv:1511.05644
Li J, Liu W, Zhou Y et al (2020) Domain adaptation with few labeled source samples by graph regularization. Neural Process Lett 51(1):23–39
Liu W, Li J, Liu B et al (2021) Unified crossdomain classification via geometric and statistical adaptations. Pattern Recogn 110:107658
Kornblith S, Shlens J, Le QV (2019) Do better imagenet models transfer better? In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 2661–2671
Girshick R, Donahue J, Darrell T, et al (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 580–587
Long M, Cao Y, Wang J, et al (2015) Learning transferable features with deep adaptation networks. In: Proceedings of the International Conference on Machine Learning, pp 97–105
Sharif RA, Azizpour H, Sullivan J, et al (2014) CNN features off-the-shelf: an astounding baseline for recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp 806–813
Guo Y, Shi H, Kumar A, et al (2019) SpotTune: transfer learning through adaptive fine-tuning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 4805–4814
Blake CL, Merz CJ (1998) Uci repository of machine learning databases
Abu Alfeilat HA, Hassanat AB, Lasassmeh O et al (2019) Effects of distance measure choice on k-nearest neighbor classifier performance: A review. Big Data 7(4):221–248
Grandini M, Bagli E, Visani G (2020) Metrics for multi-class classification: an overview. arXiv preprint arXiv:2008.05756
Elaidi H, Benabbou Z, Abbar H (2018) A comparative study of algorithms constructing decision trees: Id3 and c4. 5. In: Proceedings of the International Conference on Learning and Optimization Algorithms: Theory and Applications, pp 1–5
Patel HH, Prajapati P (2018) Study and analysis of decision tree based classification algorithms. International Journal of Computer Sciences and Engineering 6(10):74–78
Zhou ZH, Feng J (2017) Deep forest: Towards an alternative to deep neural networks. In: Proceedings of the International Joint Conference on Artificial Intelligence, pp 1038–1042
Nalepa J, Kawulok M (2019) Selecting training sets for support vector machines: a review. Artif Intell Rev 52(2):857–900
Rumelhart DE, Hinton GE, Williams RJ (1985) Learning internal representations by error propagation. Parallel Distributed Processing: Explorations in the Microstructure of Cognition 1:318–362
He K, Zhang X, Ren S, et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 770–778
Arık SO, Pfister T (2021) Tabnet: Attentive interpretable tabular learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 6679–6687
Ioffe S, Szegedy C (2015) Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: Proceedings of the International Conference on Machine Learning, pp 448–456
Kingma DP, Ba J (2015) Adam: A method for stochastic optimization. In: Proceedings of the Conference on Learning Representations
Mukkamala MC, Hein M (2017) Variants of rmsprop and adagrad with logarithmic regret bounds. In: Proceedings of the International Conference on Machine Learning, pp 2545–2553
Müller R, Kornblith S, Hinton G (2019) When does label smoothing help? In: Proceedings of the Conference on Neural Information Processing Systems, pp 4696–4705
Acknowledgements
This work is sponsored by the Science and Technology Planning Project of Guangzhou under Grant No.202103000036, the Guangdong Basic and Applied Basic Research Foundation under Grant No.2021B1515120048, the National Natural Science Foundation of China under Grant No.62072214, the Industry-University-Research Collaboration Project of Zhuhai under Grant No.ZH22017001210048PWC, the International Cooperation Project of Guangdong Province under Grant No.2020A0505100040, and the Open Project Program of Wuhan National Laboratory for Optoelectronics No.2020WNLOKF006. The corresponding author of this paper is Yuhui Deng.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Chen, P., Deng, Y., Zou, Q. et al. EAAE: A Generative Adversarial Mechanism Based Classfication Method for Small-scale Datasets. Neural Process Lett 55, 969–987 (2023). https://doi.org/10.1007/s11063-022-10921-7
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11063-022-10921-7