skip to main content
10.1145/3483207.3483218acmotherconferencesArticle/Chapter ViewAbstractPublication PagesspmlConference Proceedingsconference-collections
research-article

Facial Emotion Generation using StarGAN with Differentiable Augmentation

Published:28 October 2021Publication History

ABSTRACT

Emotion generation has remained a challenging task due to the high similarity between each emotion class. In addition, the model had to learn images with various lighting conditions and diverse facial structures. To address this limitation, we propose a modification of StarGAN by applying differentiable augmentation for generating realistic facial emotions. Furthermore, our approach allows both the generator and discriminator to generalize the data better. Finally, we evaluate the performance of the model through an emotion classifier and conduct a quantitative analysis by calculating the accuracy of the generated emotion.

References

  1. J. M. Saragih, S. Lucey, and J. F. Cohn, “Real-time avatar animation from a single image,” 2011 IEEE Int. Conf. Autom. Face Gesture Recognit. Work. FG 2011, pp. 117–124, 2011, doi: 10.1109/FG.2011.5771383.Google ScholarGoogle Scholar
  2. J. Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks,” Proc. IEEE Int. Conf. Comput. Vis., vol. 2017-Octob, pp. 2242–2251, 2017, doi: 10.1109/ICCV.2017.244.Google ScholarGoogle ScholarCross RefCross Ref
  3. X. Xia, F. Yu, N. Li, Y. Qu, J. Zhang, and C. Zhu, “Self-attention-masking semantic decomposition and segmentation for facial attribute manipulation,” IEEE Access, vol. 8, pp. 36154–36165, 2020, doi: 10.1109/ACCESS.2020.2974239.Google ScholarGoogle ScholarCross RefCross Ref
  4. H. Ding, K. Sricharan, and R. Chellappa, “ExprGAN: Facial expression editing with controllable expression intensity,” 32nd AAAI Conf. Artif. Intell. AAAI 2018, pp. 6781–6788, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  5. A. Pumarola, A. Agudo, A. M. Martinez, A. Sanfeliu, and F. Moreno-Noguer, “GANimation: One-Shot Anatomically Consistent Facial Animation,” Int. J. Comput. Vis., vol. 128, no. 3, pp. 698–713, 2020, doi: 10.1007/s11263-019-01210-3.Google ScholarGoogle ScholarCross RefCross Ref
  6. X. Wang, J. Gong, M. Hu, Y. Gu, and F. Ren, “LAUN Improved StarGAN for Facial Emotion Recognition,” IEEE Access, vol. 8, pp. 161509–161518, 2020, doi: 10.1109/access.2020.3021531.Google ScholarGoogle ScholarCross RefCross Ref
  7. Y. Choi, M. Choi, M. Kim, J. W. Ha, S. Kim, and J. Choo, “StarGAN: Unified Generative Adversarial Networks for Multi-domain Image-to-Image Translation,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., pp. 8789–8797, 2018, doi: 10.1109/CVPR.2018.00916.Google ScholarGoogle ScholarCross RefCross Ref
  8. S. Zhao, Z. Liu, J. Lin, J. Y. Zhu, and S. Han, “Differentiable Augmentation for Data-Efficient GAN Training,” arXiv, no. NeurIPS, 2020.Google ScholarGoogle Scholar
  9. P. Isola, J. Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” Proc. - 30th IEEE Conf. Comput. Vis. Pattern Recognition, CVPR 2017, vol. 2017-Janua, pp. 5967–5976, 2017, doi: 10.1109/CVPR.2017.632.Google ScholarGoogle ScholarCross RefCross Ref
  10. M. Mobini and F. Ghaderi, “StarGAN Based Facial Expression Transfer for Anime Characters,” 2020 25th Int. Comput. Conf. Comput. Soc. Iran, CSICC 2020, pp. 0–4, 2020, doi: 10.1109/CSICC49403.2020.9050061.Google ScholarGoogle ScholarCross RefCross Ref
  11. Z. He, W. Zuo, M. Kan, S. Shan, and X. Chen, “AttGAN: Facial Attribute Editing by only Changing What You Want,” IEEE Trans. Image Process., vol. 28, no. 11, pp. 5464–5478, 2019, doi: 10.1109/TIP.2019.2916751.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. I. J. Goodfellow , “Generative Adversarial Nets,” 2014, doi: 10.1109/ICCVW.2019.00369.Google ScholarGoogle Scholar
  13. H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena, “Self-attention generative adversarial networks,” 36th Int. Conf. Mach. Learn. ICML 2019, vol. 2019-June, pp. 12744–12753, 2019.Google ScholarGoogle Scholar
  14. A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” 4th Int. Conf. Learn. Represent. ICLR 2016 - Conf. Track Proc., pp. 1–16, 2016.Google ScholarGoogle Scholar
  15. P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-To-Image Translation With Conditional Adversarial Networks,” 2017, doi: 10.1007/978-3-030-11009-3_37.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. C. Ledig , “Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network,” 2017, [Online]. Available: http://openaccess.thecvf.com/content_cvpr_2017/papers/Ledig_Photo-Realistic_Single_Image_CVPR_2017_paper.pdf.Google ScholarGoogle ScholarCross RefCross Ref
  17. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Commun. ACM, vol. 60, no. 6, pp. 84–90, 2017, doi: 10.1145/3065386.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. L. Perez and J. Wang, “The Effectiveness of Data Augmentation in Image Classification using Deep Learning,” 2017, [Online]. Available: http://arxiv.org/abs/1712.04621.Google ScholarGoogle Scholar
  19. H. Zhang, Z. Zhang, A. Odena, and H. Lee, “Consistency Regularization for Generative Adversarial Networks,” in International Conference on Learning Representations, 2020, pp. 1–19.Google ScholarGoogle Scholar
  20. Z. Zhao, Z. Zhang, T. Chen, S. Singh, and H. Zhang, “Image Augmentations for GAN Training,” 2020, [Online]. Available: http://arxiv.org/abs/2006.02595.Google ScholarGoogle Scholar
  21. P. Viola, O. M. Way, and M. J. Jones, “Robust Real-Time Face Detection,” Int. J. Comput. Vis., vol. 57, no. 2, pp. 137–154, 2004.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., vol. 2016-Decem, pp. 770–778, 2016, doi: 10.1109/CVPR.2016.90.Google ScholarGoogle ScholarCross RefCross Ref
  23. D. Y. Liliana, T. Basaruddin, and I. I. D. Oriza, “The Indonesian Mixed Emotion Dataset (IMED): A facial expression dataset for mixed emotion recognition,” ACM Int. Conf. Proceeding Ser., pp. 56–60, 2018, doi: 10.1145/3293663.3293671.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. N. Aifanti, C. Papachristou, and A. Delopoulos, “The MUG Facial Expression Database,” 11th Int. Work. Image Anal. Multimed. Interact. Serv. WIAMIS 10, pp. 1–4, 2010.Google ScholarGoogle Scholar
  25. X. Zhang , “BP4D-Spontaneous: A high-resolution spontaneous 3D dynamic facial expression database,” Image Vis. Comput., vol. 32, no. 10, pp. 692–706, 2014, doi: 10.1016/j.imavis.2014.06.002.Google ScholarGoogle ScholarCross RefCross Ref
  26. M. Mavadati, P. Sanger, and M. H. Mahoor, “Extended DISFA Dataset: Investigating Posed and Spontaneous Facial Expressions,” IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. Work., pp. 1452–1459, 2016, doi: 10.1109/CVPRW.2016.182.Google ScholarGoogle ScholarCross RefCross Ref
  27. D. P. Kingma and J. L. Ba, “Adam: A method for stochastic optimization,” 3rd Int. Conf. Learn. Represent. ICLR 2015 - Conf. Track Proc., pp. 1–15, 2015.Google ScholarGoogle Scholar

Index Terms

  1. Facial Emotion Generation using StarGAN with Differentiable Augmentation
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          SPML '21: Proceedings of the 2021 4th International Conference on Signal Processing and Machine Learning
          August 2021
          183 pages
          ISBN:9781450390170
          DOI:10.1145/3483207

          Copyright © 2021 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 28 October 2021

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed limited
        • Article Metrics

          • Downloads (Last 12 months)17
          • Downloads (Last 6 weeks)2

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format .

        View HTML Format