Skip to main content

Advertisement

Log in

Image classification with a MSF dropout

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

In recent years, as the main carrier of deep learning, Deep Neural Network has attracted the attention of experts in computer field. The application of deep neural network can effectively solve complex problems in life. However, in the process of training, the complex relationship caused by noisy data leads to an overfitting, which can impact the robustness of network model. Dropout, as one kind of random regularization techniques, carries a significant effect on restraining the overfitting of deep neural network. The traditional standard dropout can restrain the overfitting in a simple and quick way, but the accuracy is impacted because it cannot accurately locate the appropriate scale. This paper proposes a multi-scale fusion (MSF) dropout method on the basis of standard dropout. At first, several groups of network model with different combinations of dropout rates were trained; then the improved genetic algorithm was used to calculate the optimal scale of each network model; by reducing the corresponding network parameters through the optimal scale, the prediction sub-models were obtained; finally, these sub-models are fused into a final prediction model with certain weight. The present study applies MSF dropout to carry out the experiments in MNIST and CIFAR-10 standard datasets. The result of the study shows that the prediction accuracy is significantly improved compared with the other two kinds of dropout, which verifies the effectiveness of the multi-scale fusion method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

References

  1. Baldi P, Sadowski P (2013) Understanding dropout. NIPS 26:2814–2822

    Google Scholar 

  2. Baldi P, Sadowski P (2014) The dropout learning algorithm. Artif Intell 210:78–122

    Article  MathSciNet  Google Scholar 

  3. Bulo SR, Porzi L, Kontschieder P (2016) Dropout distillation. ICML 48:99–107

    Google Scholar 

  4. Cybenko G (1989) Approximation by superpositions of a sigmoidal function. MCSS 2(4):303–314

    MathSciNet  MATH  Google Scholar 

  5. Gal Y, Ghahramani Z (2015) Dropout as a Bayesian approximation: insights and applications, ICML, In Deep Learning Workshop

  6. Gal Y, Ghahramani Z (2015) On modern deep learning and variational inference, NIPS, In Advances in Approximate Bayesian Inference workshop

  7. Ghezaiel W, Slimane AB, Braiek EB (2017) Nonlinear multi-scale decomposition by EMD for Co-Channel speaker identification. Multimed Tools Appl 76(20):20973–20988

    Article  Google Scholar 

  8. Glorot X, Bordes A, Bengio Y (2011) Deep sparse rectifier neural networks, AISTATS, 15:315–323

  9. Greovnik I, Kodelja T, Vertnik R, Sencic B et al (2012) Application of artificial neural networks in design of steel production path. Comput Mater Continua 30(1):19–38

    Google Scholar 

  10. Hasheminejad M, Farsi H (2017) Frame level sparse representation classification for speaker verification. Multimed Tools Appl 76(20):21211–21224

    Article  Google Scholar 

  11. Hinton GE, Srivastava N, Krizhevsky A (2012) Improving neural networks by preventing co-adaptation of feature detectors. arXiv:1207.0580

  12. Kingma DP, Salimans T, Welling M (2015) Variational dropout and the local reparameterization trick. NIPS 28:2575–2583

    Google Scholar 

  13. Liu CS (2011) A highly accurate multi-scale full/half-order polynomial interpolation. Comput Mater Continua 25(25):239–263

    Google Scholar 

  14. Nemirovski A, Juditsky A, Lan G et al (2009) Robust stochastic approximation approach to stochastic programming. Siam J Optim 19(4):1574–1609

    Article  MathSciNet  Google Scholar 

  15. Nowlan SJ, Hinton GE (1992) Simplifying neural networks by soft weight-sharing. Neural Comput 4(4):473–493

    Article  Google Scholar 

  16. Pham V, Bluche T, Kermorvant C et al (2014) Dropout improves recurrent neural networks for handwriting recognition, ICFHR, 55:285–290

  17. Ricks TM (2014) A multiscale modeling methodology for composites that includes fiber strength stochastics. Comput Mater Continua 40:99–129

    Google Scholar 

  18. Srivastava N, Hinton GE, Krizhevsky A et al (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15:1929–1958

    MathSciNet  MATH  Google Scholar 

  19. Wager S, Wang S, Liang PS (2013) Dropout training as adaptive regularization. NIPS 26:351–359

    Google Scholar 

  20. Wei G, Zhihua Z (2015) Dropout rademacher complexity of deep neural networks. Sci China In Sc 59:072104

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xian Zhong.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Luo, R., Zhong, X. & Chen, E. Image classification with a MSF dropout. Multimed Tools Appl 79, 4365–4375 (2020). https://doi.org/10.1007/s11042-019-7172-9

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-019-7172-9

Keywords

Navigation