ABSTRACT
The Generative Adversarial Networks (GANs) are deep generative models that can generate realistic samples, but they are difficult to train in practice due to the problem of mode collapse, where the generator only repeatedly generates one mode in samples during the learning process, or only generates a small number of modes after reaching the Nash equilibrium during the adversarial training. In order to solve this issue while making the generator contains promising generation ability, we propose a mixture density hyperspherical generative model namely MDH-GAN that combines variational autoencoder (VAE) and generative adversarial network. Unlike most of the GAN-based generative models that consider a Gaussian prior, MDH-GAN adopts the von Mises-Fisher (vMF) prior defined on a unit hypersphere. Our model combines VAE with GAN by integrating the encoder of VAE with GAN to form a jointly training framework. Therefore, the generator of our model can learn data distribution with a hyperspherical latent structure, leading to an improved generative ability of the generator. Moreover, a vMF mixture model is deployed in the discriminator to form a hypersphere space to avoid mode collapse of the model. In our experiments, by calculating the Fréchet Inception distance (FID) between the generated images and real ones, we prove that MDH-GAN has a better ability to generate high-quality images with high diversity.
- Goodfellow, I. (2016). Nips 2016 tutorial: Generative adversarial networks. In the Conference on Neural Information Processing Systems.Google Scholar
- Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., & Bengio, Y. (2014). Generative adversarial nets. Advances in neural information processing systems, 27.Google Scholar
- Arjovsky, M., & Bottou, L. (2017). Towards principled methods for training generative adversarial networks. In International Conference on Learning Representations.Google Scholar
- Mescheder, L., Nowozin, S., & Geiger, A. (2017). The numerics of gans. In Proceedings of the 31st International Conference on Neural Information Processing Systems (pp. 1823–1833).Google Scholar
- Arjovsky, M., Chintala, S., & Bottou, L. (2017, July). Wasserstein generative adversarial networks. In International conference on machine learning (pp. 214-223). PMLR.Google Scholar
- Dumoulin, V., Belghazi, I., Poole, B., Mastropietro, O., Lamb, A., Arjovsky, M., & Courville, A. (2016). Adversarially learned inference. In the 29th Conference on Neural Information Processing Systems.Google Scholar
- Kingma, D. P., & Welling, M. (2014). Auto-encoding variational bayes. In the 2nd International Conference on Learning Representations.Google Scholar
- Rezende, D. J., Mohamed, S., & Wierstra, D. (2014, June). Stochastic backpropagation and approximate inference in deep generative models. In International conference on machine learning (pp. 1278-1286). PMLR.Google Scholar
- Metz, L., Poole, B., Pfau, D., & Sohl-Dickstein, J. (2017). Unrolled generative adversarial networks. In the 5th International Conference on Learning Representations.Google Scholar
- Diggle, P. J., & Gratton, R. J. (1984). Monte Carlo methods of inference for implicit statistical models. Journal of the Royal Statistical Society: Series B (Methodological), 46(2), 193-212.Google Scholar
- Srivastava, A., Valkov, L., Russell, C., Gutmann, M. U., & Sutton, C. (2017, December). Veegan: Reducing mode collapse in gans using implicit variational learning. In Proceedings of the 31st International Conference on Neural Information Processing Systems (pp. 3310-3320).Google Scholar
- Eghbal-zadeh, H., Zellinger, W., & Widmer, G. (2019). Mixture density generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 5820-5829).Google Scholar
- Xu, J., & Durrett, G. (2018). Spherical Latent Spaces for Stable Variational Autoencoders. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (pp. 4503-4513).Google Scholar
- Banerjee, A., Dhillon, I. S., Ghosh, J., Sra, S., & Ridgeway, G. (2005). Clustering on the Unit Hypersphere using von Mises-Fisher Distributions. Journal of Machine Learning Research, 6(9).Google Scholar
- Fisher, N. I., Lewis, T., & Embleton, B. J. J. (1987). Statistical analysis of spherical data. Cambridge: University Press.Google ScholarCross Ref
- Davidson, T. R., Falorsi, L., De Cao, N., Kipf, T., & Tomczak, J. M. (2018, January). Hyperspherical variational auto-encoders. In 34th Conference on Uncertainty in Artificial Intelligence 2018, UAI 2018 (pp. 856-865). Association For Uncertainty in Artificial Intelligence (AUAI).Google Scholar
- Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., & Hochreiter, S. (2017). Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30.Google Scholar
- Mardia, K. V. and Jupp, P. E. (2000). Directional Statistics, New York: Wiley.Google Scholar
- Mardia, K. V. (1975). Statistics of directional data. Journal of the Royal Statistical Society. Series B (Methodological), pages 349–393.Google ScholarCross Ref
- Taghia, J., Ma, Z., & Leijon, A. (2014). Bayesian estimation of the von-mises fisher mixture model with variational inference. IEEE Transactions on Pattern Analysis & Machine Intelligence, 36(9), 1701-15.Google ScholarCross Ref
- Fan, W., & Bouguila, N. (2020). Spherical data clustering and feature selection through nonparametric Bayesian mixture models with von Mises distributions. Engineering Applications of Artificial Intelligence, 94(4), 103781.Google ScholarCross Ref
- Fan, W., Yang, L., Bouguila, N., & Chen, Y. (2020). Sequentially spherical data modeling with hidden Markov models and its application to fMRI data analysis. Knowledge-Based Systems, 206, 106341.Google ScholarCross Ref
- Kim, M. (2021). On PyTorch Implementation of Density Estimators for von Mises-Fisher and Its Mixture. arXiv preprint arXiv:2102.05340.Google Scholar
- Monatsh, K. G. (1907). Mehrdimensionale geometrie. Monatshefte für Mathematik und Physik.Google Scholar
- Larsen, A. B. L., Sønderby, S. K., Larochelle, H., & Winther, O. (2016, June). Autoencoding beyond pixels using a learned similarity metric. In International conference on machine learning (pp. 1558-1566). PMLR.Google Scholar
- Lamb, A., Dumoulin, V., & Courville, A. (2016). Discriminative regularization for generative models. arXiv preprint arXiv:1602.03220.Google Scholar
- Dosovitskiy, A., & Brox, T. (2016). Generating images with perceptual similarity metrics based on deep networks. Advances in neural information processing systems, 29, 658-666.Google Scholar
- Nalisnick, E., & Smyth, P. (2017). Stick-breaking variational autoencoders. In the 5th International Conference on Learning Representations.Google Scholar
- Ulrich, G. (1984). Computer Generation of Distributions on the M‐Sphere. Journal of the Royal Statistical Society: Series C (Applied Statistics), 33(2), 158-163.Google Scholar
- LeCun, Y. (1998). The MNIST database of handwritten digits. http://yann.lecun.com/exdb/mnist/Google Scholar
- Xiao, H., Rasul, K., & Vollgraf, R. (2017). Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747.Google Scholar
- Krizhevsky, A., & Hinton, G. (2009). Learning multiple layers of features from tiny images. https://www.cs.toronto.edu/∼kriz/cifar.htmlGoogle Scholar
- Liu, Z., Luo, P., Wang, X., & Tang, X. (2015). Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision (pp. 3730-3738).Google ScholarDigital Library
- Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., & Courville, A. (2017, December). Improved training of wasserstein GANs. In Proceedings of the 31st International Conference on Neural Information Processing Systems (pp. 5769-5779).Google Scholar
- Kodali, N., Abernethy, J., Hays, J., & Kira, Z. (2018). On convergence and stability of gans. In the 6th International Conference on Learning Representations.Google Scholar
- Berthelot, D., Schumm, T., & Metz, L. (2017). Began: Boundary equilibrium generative adversarial networks. arXiv preprint arXiv:1703.10717.Google Scholar
- Lucic, M., Kurach, K., Michalski, M., Gelly, S., & Bousquet, O. (2018). Are gans created equal? a large-scale study. In Advances in neural information processing systems (pp. 700-709).Google Scholar
- Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., & Chen, X. (2016). Improved techniques for training gans. Advances in neural information processing systems, 29, 2234-2242.Google Scholar
Recommendations
Generative adversarial networks
Generative adversarial networks are a kind of artificial intelligence algorithm designed to solve the generative modeling problem. The goal of a generative model is to study a collection of training examples and learn the probability distribution that ...
Unregularized Auto-Encoder with Generative Adversarial Networks for Image Generation
MM '18: Proceedings of the 26th ACM international conference on MultimediaWith the development of deep neural networks, recent years have witnessed the increasing research interest on generative models. Specificly, Variational Auto-Encoders (VAE) and Generative Adversarial Networks (GAN) have achieved impressive results in ...
Graphical generative adversarial networks
NIPS'18: Proceedings of the 32nd International Conference on Neural Information Processing SystemsWe propose Graphical Generative Adversarial Networks (Graphical-GAN) to model structured data. Graphical-GAN conjoins the power of Bayesian networks on compactly representing the dependency structures among random variables and that of generative ...
Comments