Abstract
The computer vision community considers the pose-invariant face recognition (PIFR) as one of the most challenging applications. Many works were devoted to enhancing face recognition performance when facing profile samples. They mainly focused on 2D- and 3D-based frontalization techniques trying to synthesize frontal views from profile ones. In the same context, we propose in this paper a new 2D PIFR technique based on Generative Adversarial Network image translation. The used GAN is Pix2Pix paired architecture covering many generator and discriminator models that will be comprehensively evaluated on a new benchmark proposed in this paper referred to as Combined-PIFR database, which is composed of four datasets that provide profiles images and their corresponding frontal ones. The paired architecture we are using is based on computing the L1 distance between the generated image and the ground truth one (pairs). Therefore, both generator and discriminator architectures are paired ones. The Combined-PIFR database is partitioned respecting person-independent constraints to evaluate our proposed framework’s frontalization and classification sub-systems fairly. Thanks to the GAN-based frontalization, the recorded results demonstrate an important improvement of 33.57% compared to the baseline.
Similar content being viewed by others
References
Ahonen T, Hadid A, Pietikainen M (2006) Face description with local binary patterns: application to face recognition. IEEE Trans Pattern Anal Mach Intell 28(12):2037–2041
Calvo MG, Lundqvist D (2008) Facial expressions of emotion (kdef): identification under different display-duration conditions. Behav Res Methods 40(1):109–115
Fischer M, Ekenel HK, Stiefelhagen R (2012) Analysis of partial least squares for pose-invariant face recognition. In: 2012 IEEE Fifth international conference on biometrics: theory, applications and systems (BTAS), pp. 331–338
Gang SM, Lee JJ (2019) Depth map extraction from the single image using pix2pix model. J Korea Multim Soc 22(5):547–557
Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. Adv Neural Infor Process Syst 2:2672–2680
Gross R, Matthews I, Cohn J, Kanade T, Baker S (2010) Multi-pie. Image Vis Comput 28(5):807–813
Hassner T, Harel S, Paz E, Enbar R (2015) Effective face frontalization in unconstrained images. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4295–4304
He K, Zhang X, Ren S, Sun J.(2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778
Huang G, Liu Z, Van Der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4700–4708
Huang R, Zhang S, Li T, He R (2017) Beyond face rotation: global and local perception gan for photorealistic and identity preserving frontal view synthesis. In: Proceedings of the IEEE international conference on computer vision, pp. 2439–2448
Kan M, Shan S, Chang H, Chen X (2014) Stacked progressive auto-encoders (SPAE) for face recognition across poses. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1883–1890
Karras T, Laine S, Aila T (2019) A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4401–4410
Kim TK, Kittler J (2006) Design and fusion of pose-invariant face-identification experts. IEEE Trans Circuits Syst Video Technol 16(9):1096–1106
King DE (2009) Dlib-ml: A machine learning toolkit. J Mach Learn Res 10:1755–1758
Langner O, Dotsch R, Bijlstra G, Wigboldus DH, Hawk ST, Van Knippenberg A (2010) Presentation and validation of the Radboud faces database. Cogn Emot 24(8):1377–1388
Li A, Shan S, Gao W (2011) Coupled bias-variance tradeoff for cross-pose face recognition. IEEE Trans Image Process 21(1):305–315
Li D, Zhou H, Lam KM (2015) High-resolution face verification using pore-scale facial features. IEEE Trans Image Process 24(8):2317–2327
Li P, Wu X, Hu Y, He R, Sun Z.(2019) M2fpa: A multi-yaw multi-pitch high-quality dataset and benchmark for facial pose analysis. In: Proceedings of the IEEE International conference on computer vision, pp. 10, 043–10,051
Luo P, Wang X, Tang X (2012) Hierarchical face parsing via deep learning. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, IEEE, pp. 2480–2487
Park U, Jain AK (2010) Face matching and retrieval using soft biometrics. IEEE Trans Inf Forensics Secur 5(3):406–415
Phillips PJ, Moon H, Rizvi SA, Rauss PJ (2000) The Feret evaluation methodology for face-recognition algorithms. IEEE Trans Pattern Anal Mach Intell 22(10):1090–1104
Prince SJ, Elder JH, Warrell J, Felisberti FM (2008) Tied factor analysis for face recognition across large pose differences. IEEE Trans Pattern Anal Mach Intell 30(6):970–984
Ronneberger O, Fischer P, Brox, T (2015) U-net: Convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention, Springer, pp. 234–241
Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, Berg AC, Fei-Fei L (2015) ImageNet large scale visual recognition challenge. Int J Comput Vision (IJCV) 115(3):211–252. https://doi.org/10.1007/s11263-015-0816-y
Sato M, Hotta K, Imanishi A, Matsuda M, Terai K (2018) Segmentation of cell membrane and nucleus by improving pix2pix. In: BIOSIGNALS, pp. 216–220
Tang H, Xu D, Sebe N, Yan Y (2019) Attention-guided generative adversarial networks for unsupervised image-to-image translation. In: 2019 International Joint Conference on Neural Networks (IJCNN), IEEE, pp. 1–8
Thomaz CE, Giraldi GA (2010) A new ranking method for principal components analysis and its application to face image analysis. Image Vis Comput 28(6):902–913
Wu X, He R, Sun Z, Tan T (2018) A light CNN for deep face representation with noisy labels. IEEE Trans Inf Forensics Secur 13(11):2884–2896
Xu R, Zhou Z, Zhang W, Yu Y (2017) Face transfer with generative adversarial network. arXiv preprint arXiv:1710.06090
Yin Y, Jiang S, Robinson JP, Fu Y (2020) Dual-attention gan for large-pose face frontalization. arXiv preprint arXiv:2002.07227
Zhu JY, Park T, Isola P, Efros AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp. 2223–2232
Zhu X, Lei Z, Yan J, Yi D, Li SZ (2015) High-fidelity pose and expression normalization for face recognition in the wild. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 787–796
Zhu Z, Luo P, Wang X, Tang X (2014) Recover canonical-view faces in the wild with deep neural networks. arXiv preprint arXiv:1404.3543
Zou H, Zhang H, Li X, Liu J, He Z (2018) Generation textured contact lenses iris images based on 4dcycle-gan. In: 2018 24th international conference on pattern recognition (ICPR), IEEE, pp. 3561–3566
Acknowledgements
The authors gratefully acknowledge the funding received from CNSRT-Maroc (Centre National de la Recherche Scientifique et Technique) and the French government (Eiffel scholarship).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Kas, M., El-merabet, Y., Ruichek, Y. et al. Generative adversarial networks for 2D-based CNN pose-invariant face recognition. Int J Multimed Info Retr 11, 639–651 (2022). https://doi.org/10.1007/s13735-022-00249-2
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13735-022-00249-2