Skip to main content

Advertisement

Log in

An analysis of generative adversarial networks and variants for image synthesis on MNIST dataset

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Generative Adversarial Networks (GANs) are most popular generative frameworks that have achieved compelling performance. They follow an adversarial approach where two deep models generator and discriminator compete with each other. They have been used for many applications especially for image synthesis because of their capability to generate high quality images. In past few years, different variants of GAN have proposed and they produced high quality results for image generation. This paper conducts an analysis of working and architecture of GAN and its popular variants for image generation in detail. In addition, we summarize and compare these models according to different parameters such as architecture, training method, learning type, benefits and performance metrics. Finally, we apply all these methods on a benchmark MNIST dataset, which contains handwritten digits and compare qualitative and quantitative results. The evaluation is based on quality of generated images, classification accuracy, discriminator loss, generator loss and computational time of these models. The aim of this study is to provide a comprehensive information about GAN and its various models in the field of image synthesis. Our main contribution in this work is critical comparison of popular GAN variants for image generation on MNIST dataset. Moreover, this paper gives insights regarding existing limitations and challenges faced by GAN and discusses associated future research work.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

References

  1. Antic J (2018) Introducing deoldify: a progressive, self-attention gan based image colorization/restoration project. https://forums.fast.ai/t/introducing-deoldify-a-progressive-self-attention-gan-based-image-colorization-restoration-project/28902/9 https://forums.fast.ai/t/introducing-deoldify-a-progressive-self-attention-gan-based-image-colorization-restoration-project/28902/9

  2. Arjovsky M, Chintala S, Bottou L (2017) Wasserstein gan. arXiv:1701.07875

  3. Barratt S, Sharma R (2018) A note on the inception score. arXiv:1801.01973

  4. Benhenda M (2017) Chemgan challenge for drug discovery: can ai reproduce natural chemical diversity? arXiv:1708.08227

  5. Berthelot D, Schumm T, Metz L (2017) Began: boundary equilibrium generative adversarial networks. arXiv:1703.10717

  6. Bora A, Price E, Dimakis AG (2018) Ambientgan: Generative models from lossy measurements. In: International conference on learning representations (ICLR)

  7. Borji A (2018) Pros and cons of gan evaluation measures. arXiv:1802.03446

  8. Cao YJ, Jia LL, Chen YX, Lin N, Yang C, Zhang B, Liu Z, Li XX, Dai HH (2018) Recent advances of generative adversarial networks in computer vision. IEEE Access 7:14985–15006

    Article  Google Scholar 

  9. Chen TH, Liao YH, Chuang CY, Hsu WT, Fu J, Sun M (2017) Show, adapt and tell: Adversarial training of cross-domain image captioner. In: ICCV, pp 521–530

  10. Chen X, Duan Y, Houthooft R, Schulman J, Sutskever I, Abbeel P (2016) Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In: NIPS

  11. Cheong B, Teo H (2018) Can we train dogs and humans at the same time? gans and hidden distributions. Tech. rep., Technical report, Stanford

  12. Chhetri SR, Lopez AB, Wan J, Al Faruque MA (2019) Gan-sec: Generative adversarial network modeling for the security analysis of cyber-physical production systems. In: 2019 Design, automation & test in europe conference & exhibition (DATE). IEEE, pp 770–775

  13. Creswell A, White T, Dumoulin V, Arulkumaran K, Sengupta B, Bharath AA (2018) Generative adversarial networks: an overview. IEEE Signal Proc Mag 35(1):53–65

    Article  Google Scholar 

  14. Denton EL, Chintala S, Szlam A, Fergus R (2015) Deep generative image models using a laplacian pyramid of adversarial networks. In: NIPS

  15. Desjardins G, Courville A, Bengio Y (2012) Disentangling factors of variation via generative entangling. arXiv:1210.5474

  16. Dickson B (2018) What is gan, the ai technique that makes computers creative? https://www.experfy.com/blog/what-is-gna-the-ai-technique-that-makes-computers-creative

  17. Donahue J, Krähenbühl P, Darrell T (2016) Adversarial feature learning. arXiv:1605.09782

  18. Duarte A, Roldan F, Tubau M, Escur J, Pascual S, Salvador A, Mohedano E, McGuinness K, Torres J, Giro-i-Nieto X (2019) Wav2pix: Speech-conditioned face generation using generative adversarial networks. In: 2019 IEEE International conference on acoustics, speech and signal processing (ICASSP). IEEE

  19. Engel J (2019) Gansynth: Making music with gans. https://magenta.tensorflow.org/gansynth

  20. Engel J, Agrawal KK, Chen S, Gulrajani I, Donahue C, Roberts A (2019) Gansynth: Adversarial neural audio synthesis. arXiv:1902.08710

  21. Evtimova K, Drozdov A (2016) Understanding mutual information and its use in infogan

  22. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In: Advances in neural information processing systems, pp 2672–2680

  23. Gorijala M, Dukkipati A (2017) Image generation and editing with variational info generative adversarial networks. arXiv:1701.04568

  24. Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: NIPS

  25. Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313(5786):504–507

    Article  MathSciNet  Google Scholar 

  26. Hitawala S (2018) Comparative study on generative adversarial networks. arXiv:1801.04271

  27. Hong Y, Hwang U, Yoo J, Yoon S (2017) How generative adversarial networks and their variants work: an overview of gan

  28. Huang H, Yu PS, Wang C (2018) An introduction to image synthesis with generative adversarial nets. arXiv:1803.04469

  29. Huang Q, Jackson PJB, Plumbley MD, Wang W (2018) Synthesis of images by two-stage generative adversarial networks. In: 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp 1593–1597

  30. Hukkelås H, Mester R, Lindseth F (2019) Deepprivacy: a generative adversarial network for face anonymization. arXiv:1909.04538

  31. Isola P, Zhu JY, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 5967– 5976

  32. Jiang C, Zhang Q, Ge Y, Liang D, Yang Y, Liu X, Zheng H, Hu Z (2019) Wasserstein generative adversarial networks for motion artifact removal in dental ct imaging. In: Medical imaging 2019: Physics of medical imaging. International society for optics and photonics, vol 10948, p 1094836

  33. Kaneko T, Hiramatsu K, Kashino K (2017) Generative attribute controller with conditional filtered generative adversarial networks. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 7006–7015

  34. Karras T, Aila T, Laine S, Lehtinen J (2017) Progressive growing of gans for improved quality, stability, and variation. arXiv:1710.10196

  35. Kim J, Kim M, Kang H, Lee K (2019) U-gat-it: Unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation

  36. Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv:1412.6980

  37. Kingma DP, Welling M (2013) Auto-encoding variational bayes. arXiv:1312.6114

  38. Li J, Monroe W, Shi T, Jean S, Ritter A, Jurafsky D (2017) Adversarial learning for neural dialogue generation. arXiv:1701.06547

  39. Liang X, Hu Z, Zhang H, Gan C, Xing EP (2017) Recurrent topic-transition gan for visual paragraph generation. arXiv:1703.07022

  40. Liu MY, Tuzel O (2016) Coupled generative adversarial networks. In: NIPS

  41. Luc P, Couprie C, Chintala S, Verbeek J (2016) Semantic segmentation using adversarial networks. arXiv:1611.08408

  42. Lucic M, Kurach K, Michalski M, Gelly S, Bousquet O (2018) Are gans created equal? a large-scale study. In: Advances in neural information processing systems, pp 700–709

  43. Makhzani A, Shlens J, Jaitly N, Goodfellow I, Frey B (2015) Adversarial autoencoders. arXiv:1511.05644

  44. Mao X, Li Q, Xie H, Lau RY, Wang Z, Paul Smolley S (2017) Least squares generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2794–2802

  45. Mariani G, Scheidegger F, Istrate R, Bekas C, Malossi C (2018) Bagan: Data augmentation with balancing gan. arXiv:1803.09655

  46. Mirza M, Osindero S (2014) Conditional generative adversarial nets. arXiv:1411.1784

  47. Mogren O (2016) C-rnn-gan: Continuous recurrent neural networks with adversarial training. arXiv:1611.09904

  48. Nimavat K, Champaneria T (2018) Generative adversarial networks (gans): Applications and future scope

  49. Odena A, Olah C, Shlens J (2017) Conditional image synthesis with auxiliary classifier gans. In: ICML

  50. Park T, Liu MY, Wang TC, Zhu JY (2019) Semantic image synthesis with spatially-adaptive normalization. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2337–2346

  51. Patel MJ, Pandya MS, Shah V (2018) Review on generative adversarial networks. In other words 4(7)

  52. Perarnau G, van de Weijer J, Raducanu B, Álvarez JM (2016) Invertible conditional gans for image editing. arXiv:1611.06355

  53. Pieters M, Wiering M (2018) Comparing generative adversarial network techniques for image creation and modification. arXiv:1803.09093

  54. Radford A, Metz L, Chintala S (2015) Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv:1511.06434

  55. Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 234–241

  56. Salakhutdinov R, Hinton G (2009) Deep boltzmann machines. In: Artificial intelligence and statistics, pp 448–455

  57. Salimans T, Goodfellow IJ, Zaremba W, Cheung V, Radford A, Chen X (2016) Improved techniques for training gans. In: NIPS

  58. Samangouei P, Kabkab M, Chellappa R (2018) Defense-gan: Protecting classifiers against adversarial attacks using generative models. arXiv:1805.06605

  59. Sanjeevi M (2019) Generative adversarial networks (gans) with math. https://medium.com/deep-math-machine-learning-ai/ch-14-general-adversarial-networks-gans-with-math-1318faf46b43

  60. Saqur R, Vivona S (2018) Capsgan: Using dynamic routing for generative adversarial networks. arXiv:1806.03968

  61. Schlegl T, Seeböck P., Waldstein SM, Schmidt-Erfurth U, Langs G (2017) Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In: International conference on information processing in medical imaging. Springer, pp 146–157

  62. Smolensky P (1986) Information processing in dynamical systems: Foundations of harmony theory. Tech. rep., Colorado Univ at Boulder Dept of Computer Science

  63. Theis L, van den Oord A, Bethge M (2015) A note on the evaluation of generative models. arXiv:1511.01844

  64. Vey BL, Gichoya JW, Prater A, Hawkins CM (2019) The role of generative adversarial networks in radiation reduction and artifact correction in medical imaging. J Am Coll Radiol 16(9):1273–1278

    Article  Google Scholar 

  65. Voulodimos A, Doulamis N, Doulamis A, Protopapadakis E (2018) Deep learning for computer vision: a brief review. Computational Intelligence and Neuroscience 2018

  66. Vuppuluri M, Dash A (2017) Survey on generative adversarial networks

  67. Wang K, Gou C, Duan Y, Lin Y, Zheng X, Wang FY (2017) Generative adversarial networks: introduction and outlook. IEEE/CAA Journal of Automatica Sinica 4:588–598

    Article  MathSciNet  Google Scholar 

  68. Wang X, Shrivastava A, Gupta A (2017) A-fast-rcnn: Hard positive generation via adversary for object detection. In: IEEE conference on computer vision and pattern recognition

  69. Wu X, Xu K, Hall P (2017) A survey of image synthesis and editing with generative adversarial networks. Tsinghua Sci Technol 22(6):660–674

    Article  Google Scholar 

  70. Yang J, Kannan A, Batra D, Parikh D (2017) Lr-gan: Layered recursive generative adversarial networks for image generation. arXiv:1703.01560

  71. Yang LC, Chou SY, Yang YH (2017) Midinet: A convolutional generative adversarial network for symbolic-domain music generation. arXiv:1703.10847

  72. Yeh R, Chen C, Lim TY, Hasegawa-Johnson M, Do MN (2016) Semantic image inpainting with perceptual and contextual losses. arXiv:1607.07539 2

  73. Yi X, Walia E, Babyn P (2019) Generative adversarial network in medical imaging: a review. Medical image analysis, p 101552

  74. Yi Z, Zhang H, Tan P, Gong M (2017) Dualgan: Unsupervised dual learning for image-to-image translation. 2017 IEEE International Conference on Computer Vision (ICCV), pp 2868–2876

  75. Yu L, Zhang W, Wang J, Yu Y (2017) Seqgan: Sequence generative adversarial nets with policy gradient. In: AAAI, pp 2852–2858

  76. Yu Z, Xiang Q, Meng J, Kou C, Ren Q, Lu Y (2019) Retinal image synthesis from multiple-landmarks input with generative adversarial networks. Biomedical Engineering Online 18(1):62

    Article  Google Scholar 

  77. Zhao W, Xu W, Yang M, Ye J, Zhao Z, Feng Y, Qiao Y (2017) Dual learning for cross-domain image captioning. In: Proceedings of the 2017 ACM on conference on information and knowledge management. ACM, pp 29–38

Download references

Acknowledgements

This research is supported by Natural Science Foundation of China (No.61602215, No.61672268), and the science foundation of Jiangsu province (No.BK20150527).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Rabia Tahir.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Cheng, K., Tahir, R., Eric, L.K. et al. An analysis of generative adversarial networks and variants for image synthesis on MNIST dataset. Multimed Tools Appl 79, 13725–13752 (2020). https://doi.org/10.1007/s11042-019-08600-2

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-019-08600-2

Keywords

Navigation