Skip to main content
Log in

Self-paced hybrid dilated convolutional neural networks

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Convolutional neural networks (CNNs) can learn the features of samples by supervised manner, and obtain outstanding achievements in many application fields. In order to improve the performance and generalization of CNNs, we propose a self-learning hybrid dilated convolution neural network (SPHDCNN), which can choose relatively reliable samples according to the current learning ability during training. In order to avoid the loss of useful feature map information caused by pooling, we introduce hybrid dilated convolution. In the proposed SPHDCNN, weight is applied to each sample to reflect the easiness of the sample. SPHDCNN employs easier samples for training first, and then adds more difficulty samples gradually according to the current learning ability. It gradually improves its performance by this learning mechanism. Experimental results show SPHDCNN has strong generalization ability, and it achieves more advanced performance compared to the baseline method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  1. Basu S, Christensen J (2013) Teaching classification boundaries to humans. In: Twenty-Seventh AAAI Conference on Artificial Intelligence

  2. Bengio Y, Louradour J, Collobert R, Weston J (2009) Curriculum learning. In: Proceedings of the 26th annual international conference on machine learning, pp 41–48

  3. Fang Y, Li Y, Tu X, Tan T, Wang X (2020) Face completion with hybrid dilated convolution. Signal Process Image Commun 80:115664

    Article  Google Scholar 

  4. Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2414–2423

  5. Jiao C, Wang X, Gou S, Chen W, Li D, Chen C, Li X (2019) Self-paced convolutional neural network for polsar images classification. Remote Sens 11(4):424

    Article  Google Scholar 

  6. Khan F, Mutlu B, Zhu J (2011) How do humans teach: On curriculum learning and teaching dimension. In: Advances in neural information processing systems, pp 1449–1457

  7. Kingma DP, Ba J (2014) Adam: A method for stochastic optimization. arXiv:1412.6980

  8. Kumar MP, Packer B, Koller D (2010) Self-paced learning for latent variable models. In: Advances in Neural Information Processing Systems, pp 1189–1197

  9. Larochelle H, Erhan D, Courville A, Bergstra J, Bengio Y (2007) An empirical evaluation of deep architectures on problems with many factors of variation. In: Proceedings of the 24th international conference on Machine learning, pp 473–480

  10. LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324

    Article  Google Scholar 

  11. Li H, Gong M (2017) Self-paced convolutional neural networks.. In: IJCAI, pp 2110–2116

  12. Li H, Gong M, Wang C, Miao Q (2018) Self-paced stacked denoising autoencoders based on differential evolution for change detection. Appl Soft Comput 71:698–714

    Article  Google Scholar 

  13. Liu C, Shang Z, Qin A (2019) A multiscale image denoising algorithm based on dilated residual convolution network. In: Chinese Conference on Image and Graphics Technologies, Springer, pp 193–203

  14. Lu Z, Yu Z, Yali P, Shigang L, Xiaojun W, Gang L, Yuan R (2018) Fast single image super-resolution via dilated residual networks. IEEE Access 7:109729–109738

    Article  Google Scholar 

  15. Meng D, Zhao Q, Jiang L (2015) What objective does self-paced learning indeed optimize?. arXiv:1511.06049

  16. Oord A, Dieleman S, Zen H, Simonyan K, Vinyals O, Graves A, Kalchbrenner N, Senior A, Kavukcuoglu K (2016) Wavenet: A generative model for raw audio. arXiv:1609.03499

  17. Paoletti ME, Haut JM, Plaza J, Plaza A (2018) A new deep convolutional neural network for fast hyperspectral image classification. ISPRS journal of photogrammetry and remote sensing 145:120–147

    Article  Google Scholar 

  18. Radovic M, Adarkwa O, Wang Q (2017) Object recognition in aerial images using convolutional neural networks. Journal of Imaging 3(2):21

    Article  Google Scholar 

  19. Rifai S, Vincent P, Muller X, Glorot X, Bengio Y (2011) Contractive auto-encoders: Explicit invariance during feature extraction

  20. Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556

  21. Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15(1):1929–1958

    MathSciNet  MATH  Google Scholar 

  22. Tang K, Ramanathan V, Fei-Fei L, Koller D (2012) Shifting weights: Adapting object detectors from image to video. In: Advances in Neural Information Processing Systems, pp 638–646

  23. Wang P, Chen P, Yuan Y, Liu D, Huang Z, Hou X, Cottrell G (2018) Understanding convolution for semantic segmentation. In: 2018 IEEE winter conference on applications of computer vision (WACV), IEEE, pp 1451–1460

  24. Yu F, Koltun V (2015) Multi-scale context aggregation by dilated convolutions. arXiv:1511.07122

  25. Zhao Q, Meng D, Jiang L, Xie Q, Xu Z, Hauptmann AG (2015) Self-paced learning for matrix factorization. In: Twenty-ninth AAAI conference on artificial intelligence

  26. Zhu P, Hao C, Hu Q, Wang Q, Zhang C (2018) Towards generalized and efficient metric learning on riemannian manifold. In: Twenty-Seventh International Joint Conference on Artificial Intelligence IJCAI-18

  27. Zhu P, Ren Q, Hu Q, Wang Q, Liu Y (2018) Beyond similar and dissimilar relations : A kernel regression formulation for metric learning. In: Twenty-Seventh International Joint Conference on Artificial Intelligence IJCAI-18

  28. Zhu X, Zhu Y, Zheng W (2019) Spectral rotation for deep one-step clustering, Pattern Recogn, https://doi.org/10.1016/j.patcog.2019.107175

  29. Zhu X, Gan J, Lu G, Li J, Zhang S (2020) Spectral clustering via half-quadratic optimization. World Wide Web 23:1969–1988

    Article  Google Scholar 

  30. Zöhrer M, Pernkopf F (2014) General stochastic networks for classification. In: Advances in Neural Information Processing Systems, pp 2015–2023

Download references

Acknowledgments

This work was supported by the National Natural Science Foundation of China (Grant No: 61836016, 61876046, 81701780 and 61672177); the Guangxi Natural Science Foundation (Grant No: 2017GXNSFBA198221); the Project of Guangxi Science and Technology (GuiKeAD19110133 and GuiKeAD17195062); Innovation Project of Guangxi Graduate Education (Grant No: JXXYYJSCXXM-012, JXXYYJSCXXM-011, JGY20200026); Research and Education Project of Guangxi Normal University (Grant No: 2019YR006); Research Fund of Guangxi Key Lab of Multi-source Information Mining & Security (Grant No: 20-A-01-01); and Basic Competence Promotion Project for Young and Middle-aged Teachers in Guangxi Education Department (Grant No: 2019KY0062).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shichao Zhang.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, W., Lu, G., Zhang, S. et al. Self-paced hybrid dilated convolutional neural networks. Multimed Tools Appl 81, 34169–34181 (2022). https://doi.org/10.1007/s11042-020-09868-5

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-020-09868-5

Keywords

Navigation