Abstract
This paper proposes an effective framework for improving angle transformation and face replacement. This ATFS framework aims to solve the image distortion caused by angled face swapping in face de-identification. In the ATFS framework, TransNet and SWGAN were proposed based on generative adversarial networks. When TransNet uses neural networks to reconstruct images, it combines the predicted facial points. As a result, Transnet can locate the target’s facial features and generate multi-angle transformed images to achieve angle transformation. The SWGAN applies a complicated neural network that combines residual operations and self-attention modules to extract the input image features and replace the face region of the input image with the target image to achieve face de-identification. Both TransNet and SWGAN adopt discriminators and various loss functions to train the neural network and accurately accelerate the training process. The experimental results show that the proposed method can maintain high-quality images and avoid image distortion during face swapping for image angle transformation compared to previously proposed methods.



















Similar content being viewed by others
References
Amelio, A., Bonifazi, G., Cauteruccio, F., Corradini, E., Marchetti, M., Ursino, D., Virgili, L.: Representation and compression of residual neural networks through a multilayer network based approach. Expert Systems with Applications 215, 119391 (2023). doi: https://doi.org/10.1016/j.eswa.2022.119391.https://www.sciencedirect.com/science/article/pii/S0957417422024095
Bakkouri, I., Afdel, K.: Computer-aided diagnosis (cad) system based on multi-layer feature fusion network for skin lesion recognition in dermoscopy images. Multimedia Tools and Applications 79(29–30), 20483–20518 (2020)
Bakkouri, I., Afdel, K.: Mlca2f: Multi-level context attentional feature fusion for covid-19 lesion segmentation from ct scans. Signal, Image and Video Processing pp. 1–8 (2022)
Bulat, A., Tzimiropoulos, G.: How far are we from solving the 2d & 3d face alignment problem?(and a dataset of 230,000 3d facial landmarks). In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1021–1030 (2017)
Chen, D., Chang, Y., Yan, R., Yang, J.: Protecting personal identification in video. In: Protecting Privacy in Video Surveillance, pp. 115–128. Springer (2009)
Chen, D., Chen, Q., Wu, J., Yu, X., Jia, T.: Face swapping: realistic image synthesis based on facial landmarks alignment. Mathematical Problems in Engineering 2019 (2019)
Cho, D., Lee, J.H., Suh, I.H.: Cleanir: Controllable attribute-preserving natural identity remover. Applied Sciences 10(3), 1120 (2020)
Cootes, T.F., Edwards, G.J., Taylor, C.J.: Active appearance models. IEEE Transactions on pattern analysis and machine intelligence 23(6), 681–685 (2001)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778 (2016)
Iglovikov, V., Shvets, A.: Ternausnet: U-net with vgg11 encoder pre-trained on imagenet for image segmentation. arXiv preprint arXiv:1801.05746 (2018)
Kazemi, V., Sullivan, J.: One millisecond face alignment with an ensemble of regression trees. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1867–1874 (2014)
Lin, C.H., Wang, Z.H., Jong, G.J.: A de-identification face recognition using extracted thermal features based on deep learning. IEEE Sensors Journal 20(16), 9510–9517 (2020)
Meden, B., Mallı, R.C., Fabijan, S., Ekenel, H.K., Štruc, V., Peer, P.: Face deidentification with generative deep neural networks. IET Signal Processing 11(9), 1046–1054 (2017)
Naruniec, J., Helminger, L., Schroers, C., Weber, R.M.: High-resolution neural face swapping for visual effects. In: Computer Graphics Forum, vol. 39, pp. 173–184. Wiley Online Library (2020)
Nirkin, Y., Keller, Y., Hassner, T.: Fsgan: Subject agnostic face swapping and reenactment. In: Proceedings of the IEEE/CVF international conference on computer vision, pp. 7184–7193 (2019)
Nirkin, Y., Keller, Y., Hassner, T.: Fsganv 2: Improved subject agnostic face swapping and reenactment. IEEE Transactions on Pattern Analysis and Machine Intelligence 45(1), 560–575 (2022)
Nirkin, Y., Masi, I., Tuan, A.T., Hassner, T., Medioni, G.: On face segmentation, face swapping, and face perception. In: 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), pp. 98–105. IEEE (2018)
Perov, I., Gao, D., Chervoniy, N., Liu, K., Marangonda, S., Umé, C., Dpfks, M., Facenheim, C.S., RP, L., Jiang, J., et al.: Deepfacelab: Integrated, flexible and extensible face-swapping framework. arXiv preprint arXiv:2005.05535 (2020)
Ribaric, S., Pavesic, N.: An overview of face de-identification in still images and videos. In: 2015 11th IEEE International conference and workshops on automatic face and gesture recognition (FG), vol. 4, pp. 1–6. IEEE (2015)
Rowley, H.A., Baluja, S., Kanade, T.: Neural network-based face detection. IEEE Transactions on pattern analysis and machine intelligence 20(1), 23–38 (1998)
Samarzija, B., Ribaric, S.: An approach to the de-identification of faces in different poses. In: 2014 37th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), pp. 1246–1251. IEEE (2014)
Shu, C., Wu, H., Zhou, H., Liu, J., Hong, Z., Ding, C., Han, J., Liu, J., Ding, E., Wang, J.: Few-shot head swapping in the wild. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10789–10798 (2022)
Souček, T., Moravec, J., Lokoč, J.: Transnet: A deep network for fast detection of common shot transitions. arXiv preprint arXiv:1906.03363 (2019)
Sun, K., Xiao, B., Liu, D., Wang, J.: Deep high-resolution representation learning for human pose estimation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 5693–5703 (2019)
Tuan Tran, A., Hassner, T., Masi, I., Medioni, G.: Regressing robust and discriminative 3d morphable models with a very deep neural network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5163–5172 (2017)
Umeyama, S.: Least-squares estimation of transformation parameters between two point patterns. IEEE Transactions on Pattern Analysis & Machine Intelligence 13(04), 376–380 (1991)
Wu, J., Huang, Z., Acharya, D., Li, W., Thoma, J., Paudel, D.P., Gool, L.V.: Sliced wasserstein generative models. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3713–3722 (2019)
Wu, Y., Yang, F., Xu, Y., Ling, H.: Privacy-protective-gan for privacy preserving face de-identification. Journal of Computer Science and Technology 34(1), 47–60 (2019)
Xu, Y., Deng, B., Wang, J., Jing, Y., Pan, J., He, S.: High-resolution face swapping via latent semantics disentanglement. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7642–7651 (2022)
Funding
This research was funded by Ministry of Science and Technology grant number MOST 109-2221-E-005-057-MY2 and .
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interests/Competing interests
The authors declare no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Tsai, CS., Wu, HC., Chen, WT. et al. ATFS: A deep learning framework for angle transformation and face swapping of face de-identification. Multimed Tools Appl 83, 36797–36822 (2024). https://doi.org/10.1007/s11042-023-16123-0
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-023-16123-0