Skip to main content

Capturing the Persistence of Facial Expression Features for Deepfake Video Detection

  • Conference paper
  • First Online:
Information and Communications Security (ICICS 2019)

Part of the book series: Lecture Notes in Computer Science ((LNSC,volume 11999))

Included in the following conference series:

Abstract

The security of the Deepfake video has become the focus of social concern. This kind of fake video not only infringes copyright and privacy but also poses potential risks to politics, journalism, social trust, and other aspects. Unfortunately, fighting against Deepfake video is still in its early stage and practical solutions are required. Currently, biological signal based and learning-based are two major ways in detecting Deepfake video. We explore that facial expression between two adjacent frames appears significant differences in generative adversarial network (GAN)-synthesized fake video, while in a real video the facial expression looks naturally and transforms in a smooth way across frames. In this paper, we employ optical flow to capture the obvious differences of facial expressions between adjacent frames in a video and incorporate the temporal characteristics of consecutive frames into a convolutional neural network (CNN) model to distinguish the Deepfake video. In our experiments, we evaluate the effectiveness of our approach on a publicly fake video dataset, FaceForensics++. Experimental results show that our proposed approach achieves an accuracy higher than 98.1% and the AUC score reaches more than 0.9981.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Deepfacelab. http://deepfakes.com.cn/. Accessed 22 Apr 2019

  2. Deepfake github. http://github.com/deepfakes/faceswap. Accessed 20 Apr 2019

  3. Deepfake wikipedia. https://en.wikipedia.org/wiki/Deepfake. Accessed 12 Sept 2019

  4. Dlib. http://dlib.net/. Accessed 20 Mar 2019

  5. Fakeapp. http://www.fakeapp.com/. Accessed 22 Apr 2019

  6. Ffmpeg. http://ffmpeg.org/. Accessed 20 Mar 2019

  7. Keras. http://keras.io/. Accessed 10 July 2018

  8. Tensorflow. http://tensorflow.google.cn/. Accessed 10 July 2018

  9. Zao. https://apps.apple.com/cn/app/zao/id1465199127. Accessed 12 Sept 2019

  10. Afchar, D., Nozick, V., Yamagishi, J., Echizen, I.: Mesonet: a compact facial video forgery detection network. In: 2018 IEEE International Workshop on Information Forensics and Security, WIFS 2018, Hong Kong, China, 11–13 December 2018, pp. 1–7 (2018). https://doi.org/10.1109/WIFS.2018.8630761

  11. Agarwal, S., Farid, H., Gu, Y., He, M., Nagano, K., Li, H.: Protecting world leaders against deep fakes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 38–45 (2019)

    Google Scholar 

  12. Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017, pp. 1800–1807 (2017). https://doi.org/10.1109/CVPR.2017.195

  13. Goodfellow, I.J., et al.: Generative adversarial networks. CoRR abs/1406.2661 (2014). http://arxiv.org/abs/1406.2661

  14. Güera, D., Delp, E.J.: Deepfake video detection using recurrent neural networks. In: 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 1–6 (2018). https://doi.org/10.1109/AVSS.2018.8639163

  15. Li, H., Li, B., Tan, S., Huang, J.: Detection of deep network generated images using disparities in color components. arXiv preprint arXiv:1808.07276 (2018)

  16. Li, Y., Chang, M., Lyu, S.: In ictu oculi: exposing AI created fake videos by detecting eye blinking. In: 2018 IEEE International Workshop on Information Forensics and Security, WIFS 2018, Hong Kong, China, 11–13 December 2018, pp. 1–7 (2018). https://doi.org/10.1109/WIFS.2018.8630787

  17. Lyu, Y.L.S.: Exposing deepfake videos by detecting face warping artifacts. CoRR abs/1811.00656 (2018). http://arxiv.org/abs/1811.00656

  18. McCloskey, S., Albright, M.: Detecting GAN-generated imagery using color cues. arXiv preprint arXiv:1812.08247 (2018)

  19. Nataraj, L., et al.: Detecting GAN generated fake images using co-occurrence matrices. arXiv preprint arXiv:1903.06836 (2019)

  20. Rössler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nießner, M.: Faceforensics: a large-scale video dataset for forgery detection in human faces. CoRR abs/1803.09179 (2018). http://arxiv.org/abs/1803.09179

  21. Sabir, E., Cheng, J., Jaiswal, A., AbdAlmageed, W., Masi, I., Natarajan, P.: Recurrent convolutional strategies for face manipulation detection in videos. Interfaces (GUI) 3, 1 (2019)

    Google Scholar 

  22. Taheri, S., Veidenbaum, A.V., Nicolau, A., Hu, N., Haghighat, M.R.: Opencv.js: computer vision processing for the open web platform. In: Proceedings of the 9th ACM Multimedia Systems Conference, MMSys 2018, Amsterdam, The Netherlands, 12–15 June 2018, pp. 478–483 (2018). https://doi.org/10.1145/3204949.3208126

  23. Wang, L., et al.: Temporal segment networks: towards good practices for deep action recognition. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9912, pp. 20–36. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46484-8_2

    Chapter  Google Scholar 

  24. Yang, X., Li, Y., Lyu, S.: Exposing deep fakes using inconsistent head poses. In: IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2019, Brighton, United Kingdom, 12–17 May 2019, pp. 8261–8265 (2019). https://doi.org/10.1109/ICASSP.2019.8683164

Download references

Acknowledgment

This work is partly supported by National Natural Science Foundation of China under Grant No.61672394 and 61872273. Any opinions, findings, and conclusions or recommendations expressed in this paper are those of the authors and do not necessarily reflect the views of the funding agencies.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lei Zhao .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhao, Y., Ge, W., Li, W., Wang, R., Zhao, L., Ming, J. (2020). Capturing the Persistence of Facial Expression Features for Deepfake Video Detection. In: Zhou, J., Luo, X., Shen, Q., Xu, Z. (eds) Information and Communications Security. ICICS 2019. Lecture Notes in Computer Science(), vol 11999. Springer, Cham. https://doi.org/10.1007/978-3-030-41579-2_37

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-41579-2_37

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-41578-5

  • Online ISBN: 978-3-030-41579-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics