Skip to main content

AE-GAN: Attention Embedded GAN for Irregular and Large-Area Mask Face Image Inpainting

  • Conference paper
  • First Online:
Advances in Computer Graphics (CGI 2022)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13443))

Included in the following conference series:

  • 1156 Accesses

Abstract

Existing image inpainting methods have shown promising results for regular and small-area breaks. However, restoration of irregular and large-area damage is still tricky and achieves mediocre results due to the lack of restrictions on the center of the hole. In contrast, face inpainting is also problematic due to facial structure and texture complexity, which always results in structural confusion and texture blurring. We propose an attention embedded adversarial generative network (AE-GAN) in the paper to solve this problem. Overall our framework is a U-shape GAN model. To enable the network to capture the practical features faster to reconstruct the content of the masked region in the face image, we also embed the attention mechanism that simplifies the Squeeze-and-Excitation channel attention mechanism and then set it reasonably in our generator. Our generator is chosen the U-net structure as a backbone. Because the structure can encode information from low-level pixels context features to high-level semantic features. And it can decode the features back into an image. Experiments on CelebA-HQ datasets demonstrate that our proposed method generates higher quality inpainting, results in consistent and harmonious facial structures and appearance than existing methods, and achieves state-of-the-art performance.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Aleem, S., Yang, P., Masood, S., Li, P., Sheng, B.: An accurate multi-modal biometric identification system for person identification via fusion of face and finger print. World Wide Web 23(2), 1299–1317 (2019). https://doi.org/10.1007/s11280-019-00698-6

    Article  Google Scholar 

  2. Badatia, P., Tasgaonkar, P.P.: Crowd counting and density estimation using multicolumn discriminator in GAN. In: 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), pp. 1179–1183 (2018). https://doi.org/10.1109/ICACCI.2018.8554372

  3. Chen, Z., et al.: Structure-aware image inpainting using patch scale optimization. J. Vis. Commun. Image Represent. 40, 312–323 (2016). https://doi.org/10.1016/j.jvcir.2016.06.029, https://www.sciencedirect.com/science/article/pii/S1047320316301262

  4. Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2414–2423 (2016). https://doi.org/10.1109/CVPR.2016.265

  5. Goodfellow, I., et al.: Generative adversarial networks. Commun. ACM 63(11), 139–144 (2020). https://doi.org/10.1145/3422622, https://doi.org/10.1145/3422622

  6. Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANS trained by a two time-scale update rule converge to a local nash equilibrium. In: 30th Proceedings Conference on Advances in Neural Information Processing System (2017)

    Google Scholar 

  7. Iizuka, S., Simo-Serra, E., Ishikawa, H.: Globally and locally consistent image completion. ACM Trans. Graph. 36(4) (2017). https://doi.org/10.1145/3072959.3073659, https://doi.org/10.1145/3072959.3073659

  8. Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694–711. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46475-6_43

    Chapter  Google Scholar 

  9. Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8107–8116 (2020). https://doi.org/10.1109/CVPR42600.2020.00813

  10. Ke, Q., Ming, L.D., Daxing, Z.: Image steganalysis via multi-column convolutional neural network. In: 2018 14th IEEE International Conference on Signal Processing (ICSP), pp. 550–553 (2018). https://doi.org/10.1109/ICSP.2018.8652324

  11. Li, J., Wang, N., Zhang, L., Du, B., Tao, D.: Recurrent feature reasoning for image inpainting. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7757–7765 (2020). https://doi.org/10.1109/CVPR42600.2020.00778

  12. Li, L., Tang, J., Ye, Z., Sheng, B., Mao, L., Ma, L.: Unsupervised face super-resolution via gradient enhancement and semantic guidance. The Vis. Comput. 37(9), 2855–2867 (2021). https://doi.org/10.1007/s00371-021-02236-w

  13. Li, Y., Liu, S., Yang, J., Yang, M.H.: Generative face completion. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5892–5900 (2017). https://doi.org/10.1109/CVPR.2017.624

  14. Liu, B., Li, P., Sheng, B., Nie, Y., Wu, E.: Structure-preserving image completion with multi-level dynamic patches. The Vis. Comput. 35(1), 85–98 (2019). https://doi.org/10.1007/s00371-017-1454-x

  15. Liu, G., Reda, F.A., Shih, K.J., Wang, T.-C., Tao, A., Catanzaro, B.: Image inpainting for irregular holes using partial convolutions. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11215, pp. 89–105. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01252-6_6

    Chapter  Google Scholar 

  16. Liu, H., Jiang, B., Xiao, Y., Yang, C.: Coherent semantic attention for image inpainting. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 4169–4178 (2019). https://doi.org/10.1109/ICCV.2019.00427

  17. Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 3730–3738 (2015). https://doi.org/10.1109/ICCV.2015.425

  18. Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral normalization for generative adversarial networks. arXiv preprint arXiv:1802.05957 (2018)

  19. Pathak, D., Krähenbühl, P., Donahue, J., Darrell, T., Efros, A.A.: Context encoders: Feature learning by inpainting. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2536–2544 (2016). https://doi.org/10.1109/CVPR.2016.278

  20. Peng, J., Liu, D., Xu, S., Li, H.: Generating diverse structure for image inpainting with hierarchical vq-vae. In: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10770–10779 (2021). https://doi.org/10.1109/CVPR46437.2021.01063

  21. Sagong, M.c., Shin, Y.g., Kim, S.w., Park, S., Ko, S.j.: Pepsi : fast image inpainting with parallel decoding network. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11352–11360 (2019). https://doi.org/10.1109/CVPR.2019.01162

  22. Sheng, B., Li, P., Gao, C., Ma, K.L.: Deep neural representation guided face sketch synthesis. IEEE Trans. Visual Comput. Graphics 25(12), 3216–3230 (2019). https://doi.org/10.1109/TVCG.2018.2866090

    Article  Google Scholar 

  23. Song, L., Cao, J., Song, L., Hu, Y., He, R.: Geometry-aware face completion and editing. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 2506–2513 (2019). https://doi.org/10.1609/aaai.v33i01.33012506

  24. Wang, Z., Bovik, A., Sheikh, H., Simoncelli, E.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004). https://doi.org/10.1109/TIP.2003.819861

    Article  Google Scholar 

  25. Wu, X., Xu, K., Hall, P.: A survey of image synthesis and editing with generative adversarial networks. Tsinghua Sci. Technol. 22(6), 660–674 (2017). https://doi.org/10.23919/TST.2017.8195348

  26. Yu, J., Lin, Z., Yang, J., Shen, X., Lu, X., Huang, T.S.: Generative image inpainting with contextual attention. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5505–5514 (2018). https://doi.org/10.1109/CVPR.2018.00577

  27. Zeng, Y., Fu, J., Chao, H., Guo, B.: Learning pyramid-context encoder network for high-quality image inpainting. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1486–1494 (2019). https://doi.org/10.1109/CVPR.2019.00158

  28. Zhou, T., Ding, C., Lin, S., Wang, X., Tao, D.: Learning oracle attention for high-fidelity face completion. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7677–7686 (2020). https://doi.org/10.1109/CVPR42600.2020.00770

Download references

Acknowledgments

We would like to thank the anonymous reviewers for their valuable suggestions. This paper is supported by the Shandong Provincial Natural Science Foundation (ZR2020MF132), the National Natural Science Foundation of China (No. 62072020), the Leading Talents in Innovation and Entrepreneurship of Qingdao (19-3-2-21-zhc), and the Key-Area Research and Development Program of Guangdong Province (No. 209B010150001).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yue Qi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Bao, Y., Xiao, X., Qi, Y. (2022). AE-GAN: Attention Embedded GAN for Irregular and Large-Area Mask Face Image Inpainting. In: Magnenat-Thalmann, N., et al. Advances in Computer Graphics. CGI 2022. Lecture Notes in Computer Science, vol 13443. Springer, Cham. https://doi.org/10.1007/978-3-031-23473-6_26

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-23473-6_26

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-23472-9

  • Online ISBN: 978-3-031-23473-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics