Abstract
Augmentations are the key factor in determining the performance of any neural network as they provide a model with a critical edge in boosting its performance. Their ability to boost a model’s robustness depends on two factors, viz-a-viz, the model architecture, and the type of augmentations. Augmentations are very specific to a dataset, and it is not imperative that all kinds of augmentation would necessarily produce a positive effect on a model’s performance. Hence there is a need to identify augmentations that perform consistently well across a variety of datasets and also remain invariant to the type of architecture, convolutions, and the number of parameters used. This paper evaluates the effect of parameters using 3 × 3 and depth-wise separable convolutions on different augmentation techniques on MNIST, FMNIST, and CIFAR10 datasets. Statistical Evidence shows that techniques such as Cutouts and Random horizontal flip were consistent on both parametrically low and high architectures. Depth-wise separable convolutions outperformed 3 × 3 convolutions at higher parameters due to their ability to create deeper networks. Augmentations resulted in bridging the accuracy gap between the 3 × 3 and depth-wise separable convolutions, thus establishing their role in model generalization. At higher number augmentations did not produce a significant change in performance. The synergistic effect of multiple augmentations at higher parameters, with antagonistic effect at lower parameters, was also evaluated. The work proves that a delicate balance between architectural supremacy and augmentations needs to be achieved to enhance a model’s performance in any given deep learning task.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Mikołajczyk, A., Grochowski, M.: Data augmentation for improving deep learning in image classification problem. In: 2018 International Interdisciplinary PhD Workshop, IIPhDW 2018, Jun 2018, pp. 117–122. https://doi.org/10.1109/IIPHDW.2018.8388338
Falconi, L.G., Perez, M., Aguilar, W.G.: Transfer learning in breast mammogram abnormalities classification with Mobilenet and Nasnet. In: International Conference on Systems, Signals, and Image Processing, vol. 2019 June, pp. 109–114, Jun 2019. https://doi.org/10.1109/IWSSIP.2019.8787295
Chollet, F.: Xception: Deep Learning with Depthwise Separable Convolutions (2016)
Lin, M., Chen, Q., Yan, S.: Network in network. arXiv:1312.4400 [cs], Mar 2014, Accessed 17 Jan 2022 [Online]. Available http://arxiv.org/abs/1312.4400
Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2921–2929 (2016). https://doi.org/10.1109/CVPR.2016.319
Ethiraj, S., Bolla, B.K.: Classification of astronomical bodies by efficient layer fine-tuning of deep neural networks. In: 2021 5th Conference on Information and Communication Technology (CICT), Kurnool, India, pp. 1–6, Dec 2021. https://doi.org/10.1109/CICT53865.2020.9672430
Li, Y., Wang, K.: Modified convolutional neural network with global average pooling for intelligent fault diagnosis of industrial gearbox. EiN 22(1), 63–72 (2019). https://doi.org/10.17531/ein.2020.1.8
Li, Z., Wang, S., Fan, R., Cao, G., Zhang, Y., Guo, T.: Teeth category classification via seven-layer deep convolutional neural network with max pooling and global average pooling. Int. J. Imaging Syst. Technol. 29(4), 577–583 (2019). https://doi.org/10.1002/ima.22337
Girosi, F., Jones, M., Poggio, T.: Regularization theory and neural networks architectures. Neural Comput. 7(2), 219–269 (1995). https://doi.org/10.1162/neco.1995.7.2.219
Burden, F., Winkler, D.: Bayesian regularization of neural networks. In: Livingstone, D.J. (ed) Artificial Neural Networks, vol. 458, pp. 23–42. Humana Press, Totowa, NJ (2008). https://doi.org/10.1007/978-1-60327-101-1_3
DeVries, T., Taylor, G.W.: Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552 [cs], Nov 2017. [Online]. Available http://arxiv.org/abs/1708.04552. Accessed 17 Jan 2022
Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: beyond empirical risk minimization. arXiv:1710.09412 [cs, stat], Apr 2018 [Online]. Available http://arxiv.org/abs/1710.09412. Accessed 17 Jan 2022
He, X., et al.: Sample-efficient deep learning for COVID-19 diagnosis based on CT scans. Health Inform. preprint, Apr 2020. https://doi.org/10.1101/2020.04.13.20063941
Shorten, C., Khoshgoftaar, T.M.: A survey on image data augmentation for deep learning. J. Big Data 6(1), 1–48 (2019). https://doi.org/10.1186/s40537-019-0197-0
Wang, J., Perez, L.: The effectiveness of data augmentation in image classification using deep learning (2017)
T.M.M. Team: composer [Online] (2021). Available https://github.com/mosaicml/composer/
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Ethiraj, S., Bolla, B.K. (2022). Augmentations: An Insight into Their Effectiveness on Convolution Neural Networks. In: Singh, M., Tyagi, V., Gupta, P.K., Flusser, J., Ören, T. (eds) Advances in Computing and Data Sciences. ICACDS 2022. Communications in Computer and Information Science, vol 1613. Springer, Cham. https://doi.org/10.1007/978-3-031-12638-3_26
Download citation
DOI: https://doi.org/10.1007/978-3-031-12638-3_26
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-12637-6
Online ISBN: 978-3-031-12638-3
eBook Packages: Computer ScienceComputer Science (R0)