Skip to main content

Augmentations: An Insight into Their Effectiveness on Convolution Neural Networks

  • Conference paper
  • First Online:
Advances in Computing and Data Sciences (ICACDS 2022)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1613))

Included in the following conference series:

  • 654 Accesses

Abstract

Augmentations are the key factor in determining the performance of any neural network as they provide a model with a critical edge in boosting its performance. Their ability to boost a model’s robustness depends on two factors, viz-a-viz, the model architecture, and the type of augmentations. Augmentations are very specific to a dataset, and it is not imperative that all kinds of augmentation would necessarily produce a positive effect on a model’s performance. Hence there is a need to identify augmentations that perform consistently well across a variety of datasets and also remain invariant to the type of architecture, convolutions, and the number of parameters used. This paper evaluates the effect of parameters using 3 × 3 and depth-wise separable convolutions on different augmentation techniques on MNIST, FMNIST, and CIFAR10 datasets. Statistical Evidence shows that techniques such as Cutouts and Random horizontal flip were consistent on both parametrically low and high architectures. Depth-wise separable convolutions outperformed 3 × 3 convolutions at higher parameters due to their ability to create deeper networks. Augmentations resulted in bridging the accuracy gap between the 3 × 3 and depth-wise separable convolutions, thus establishing their role in model generalization. At higher number augmentations did not produce a significant change in performance. The synergistic effect of multiple augmentations at higher parameters, with antagonistic effect at lower parameters, was also evaluated. The work proves that a delicate balance between architectural supremacy and augmentations needs to be achieved to enhance a model’s performance in any given deep learning task.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Mikołajczyk, A., Grochowski, M.: Data augmentation for improving deep learning in image classification problem. In: 2018 International Interdisciplinary PhD Workshop, IIPhDW 2018, Jun 2018, pp. 117–122. https://doi.org/10.1109/IIPHDW.2018.8388338

  2. Falconi, L.G., Perez, M., Aguilar, W.G.: Transfer learning in breast mammogram abnormalities classification with Mobilenet and Nasnet. In: International Conference on Systems, Signals, and Image Processing, vol. 2019 June, pp. 109–114, Jun 2019. https://doi.org/10.1109/IWSSIP.2019.8787295

  3. Chollet, F.: Xception: Deep Learning with Depthwise Separable Convolutions (2016)

    Google Scholar 

  4. Lin, M., Chen, Q., Yan, S.: Network in network. arXiv:1312.4400 [cs], Mar 2014, Accessed 17 Jan 2022 [Online]. Available http://arxiv.org/abs/1312.4400

  5. Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2921–2929 (2016). https://doi.org/10.1109/CVPR.2016.319

  6. Ethiraj, S., Bolla, B.K.: Classification of astronomical bodies by efficient layer fine-tuning of deep neural networks. In: 2021 5th Conference on Information and Communication Technology (CICT), Kurnool, India, pp. 1–6, Dec 2021. https://doi.org/10.1109/CICT53865.2020.9672430

  7. Li, Y., Wang, K.: Modified convolutional neural network with global average pooling for intelligent fault diagnosis of industrial gearbox. EiN 22(1), 63–72 (2019). https://doi.org/10.17531/ein.2020.1.8

    Article  Google Scholar 

  8. Li, Z., Wang, S., Fan, R., Cao, G., Zhang, Y., Guo, T.: Teeth category classification via seven-layer deep convolutional neural network with max pooling and global average pooling. Int. J. Imaging Syst. Technol. 29(4), 577–583 (2019). https://doi.org/10.1002/ima.22337

    Article  Google Scholar 

  9. Girosi, F., Jones, M., Poggio, T.: Regularization theory and neural networks architectures. Neural Comput. 7(2), 219–269 (1995). https://doi.org/10.1162/neco.1995.7.2.219

    Article  Google Scholar 

  10. Burden, F., Winkler, D.: Bayesian regularization of neural networks. In: Livingstone, D.J. (ed) Artificial Neural Networks, vol. 458, pp. 23–42. Humana Press, Totowa, NJ (2008). https://doi.org/10.1007/978-1-60327-101-1_3

  11. DeVries, T., Taylor, G.W.: Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552 [cs], Nov 2017. [Online]. Available http://arxiv.org/abs/1708.04552. Accessed 17 Jan 2022

  12. Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: beyond empirical risk minimization. arXiv:1710.09412 [cs, stat], Apr 2018 [Online]. Available http://arxiv.org/abs/1710.09412. Accessed 17 Jan 2022

  13. He, X., et al.: Sample-efficient deep learning for COVID-19 diagnosis based on CT scans. Health Inform. preprint, Apr 2020. https://doi.org/10.1101/2020.04.13.20063941

  14. Shorten, C., Khoshgoftaar, T.M.: A survey on image data augmentation for deep learning. J. Big Data 6(1), 1–48 (2019). https://doi.org/10.1186/s40537-019-0197-0

    Article  Google Scholar 

  15. Wang, J., Perez, L.: The effectiveness of data augmentation in image classification using deep learning (2017)

    Google Scholar 

  16. T.M.M. Team: composer [Online] (2021). Available https://github.com/mosaicml/composer/

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bharath Kumar Bolla .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ethiraj, S., Bolla, B.K. (2022). Augmentations: An Insight into Their Effectiveness on Convolution Neural Networks. In: Singh, M., Tyagi, V., Gupta, P.K., Flusser, J., Ören, T. (eds) Advances in Computing and Data Sciences. ICACDS 2022. Communications in Computer and Information Science, vol 1613. Springer, Cham. https://doi.org/10.1007/978-3-031-12638-3_26

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-12638-3_26

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-12637-6

  • Online ISBN: 978-3-031-12638-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics