skip to main content
10.1145/3577163.3595103acmconferencesArticle/Chapter ViewAbstractPublication Pagesih-n-mmsecConference Proceedingsconference-collections
short-paper

Exposing Deepfakes using Dual-Channel Network with Multi-Axis Attention and Frequency Analysis

Published:28 June 2023Publication History

ABSTRACT

This paper proposes a dual-channel network for DeepFake detection. The network comprises two channels: one using a stacked Maxvit block to process the downsampled original images, and the other using a stacked ResNet basic block to capture features from the discrete cosine transform of the image spectrums. The components extracted from the two channels are concatenated using a linear layer to train the entire model for exposing DeepFakes. Experimental results demonstrate that the proposed method could achieve satisfactory forensics performance. Besides, the experiments of cross-dataset evaluations prove it is also high in generalizability.

References

  1. 1] Darius Afchar, Vincent Nozick, Junichi Yamagishi, and Isao Echizen. 2018. Mesonet: a compact facial video forgery detection network. In 2018 IEEE in- ternational workshop on information forensics and security (WIFS). IEEE, 1--7.Google ScholarGoogle Scholar
  2. Nicolo Bonettini, Edoardo Daniele Cannas, Sara Mandelli, Luca Bondi, Paolo Bestagini, and Stefano Tubaro. 2021. Video face manipulation detection through ensemble of cnns. In 2020 25th international conference on pattern recognition (ICPR). IEEE, 5012--5019.Google ScholarGoogle ScholarCross RefCross Ref
  3. François Chollet. 2017. Xception: Deep learning with depthwise separable con- volutions. In Proceedings of the IEEE conference on computer vision and pattern recognition. 1251--1258.Google ScholarGoogle ScholarCross RefCross Ref
  4. Xiangxiang Chu, Zhi Tian, Bo Zhang, Xinlong Wang, Xiaolin Wei, Huaxia Xia, and Chunhua Shen. 2021. Conditional positional encodings for vision transformers. arXiv preprint arXiv:2102.10882 (2021).Google ScholarGoogle Scholar
  5. Feng Ding, Bing Fan, Zhangyi Shen, Keping Yu, Gautam Srivastava, Kapal Dev, and Shaohua Wan. 2022. Securing Facial Bioinformation by Eliminating Adver- sarial Perturbations. IEEE Transactions on Industrial Informatics (2022).Google ScholarGoogle Scholar
  6. Feng Ding, Zhangyi Shen, Guopu Zhu, Sam Kwong, Yicong Zhou, and Siwei Lyu. 2022. ExS-GAN: Synthesizing Anti-Forensics Images via Extra Supervised GAN. IEEE Transactions on Cybernetics (2022).Google ScholarGoogle Scholar
  7. Feng Ding, Guopu Zhu, Yingcan Li, Xinpeng Zhang, Pradeep K. Atrey, and Siwei Lyu. 2022. Anti-Forensics for Face Swapping Videos via Adversarial Training. IEEE Transactions on Multimedia 24 (2022), 3429--3441.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xi- aohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020).Google ScholarGoogle Scholar
  9. Ricard Durall, Margret Keuper, Franz-Josef Pfreundt, and Janis Keuper. 2019. Unmasking deepfakes with simple features. arXiv preprint arXiv:1911.00686 (2019).Google ScholarGoogle Scholar
  10. Patrick Esser, Robin Rombach, and Bjorn Ommer. 2021. Taming transformers for high-resolution image synthesis. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 12873--12883.Google ScholarGoogle ScholarCross RefCross Ref
  11. Joel Frank, Thorsten Eisenhofer, Lea Schönherr, Asja Fischer, Dorothea Kolossa, and Thorsten Holz. 2020. Leveraging frequency analysis for deep fake image recognition. In International conference on machine learning. PMLR, 3247--3258.Google ScholarGoogle Scholar
  12. Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017).Google ScholarGoogle Scholar
  13. Yonghyun Jeong, Doyeon Kim, Youngmin Ro, and Jongwon Choi. 2022. FrePGAN: robust deepfake detection using frequency-level perturbations. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 36. 1060--1068.Google ScholarGoogle ScholarCross RefCross Ref
  14. Sohail Ahmed Khan and Duc-Tien Dang-Nguyen. 2022. Hybrid Transformer Net- work for Deepfake Detection. In Proceedings of the 19th International Conference on Content-based Multimedia Indexing. 8--14.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Lingzhi Li, Jianmin Bao, Ting Zhang, Hao Yang, Dong Chen, Fang Wen, and Bain- ing Guo. 2020. Face x-ray for more general face forgery detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 5001--5010.Google ScholarGoogle ScholarCross RefCross Ref
  16. Yuezun Li and Siwei Lyu. 2018. Exposing deepfake videos by detecting face warping artifacts. arXiv preprint arXiv:1811.00656 (2018).Google ScholarGoogle Scholar
  17. Yuezun Li, Xin Yang, Pu Sun, Honggang Qi, and Siwei Lyu. 2020. Celeb-df: A large-scale challenging dataset for deepfake forensics. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 3207--3216.Google ScholarGoogle ScholarCross RefCross Ref
  18. Honggu Liu, Xiaodan Li, Wenbo Zhou, Yuefeng Chen, Yuan He, Hui Xue, Weiming Zhang, and Nenghai Yu. 2021. Spatial-phase shallow learning: rethinking face forgery detection in frequency domain. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 772--781.Google ScholarGoogle ScholarCross RefCross Ref
  19. Zhuang Liu, Hanzi Mao, Chao-Yuan Wu, Christoph Feichtenhofer, Trevor Darrell, and Saining Xie. 2022. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 11976--11986.Google ScholarGoogle ScholarCross RefCross Ref
  20. Yuchen Luo, Yong Zhang, Junchi Yan, and Wei Liu. 2021. Generalizing face forgery detection with high-frequency features. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 16317--16326.Google ScholarGoogle ScholarCross RefCross Ref
  21. Falko Matern, Christian Riess, and Marc Stamminger. 2019. Exploiting visual artifacts to expose deepfakes and face manipulations. In 2019 IEEE Winter Appli- cations of Computer Vision Workshops (WACVW). IEEE, 83--92.Google ScholarGoogle Scholar
  22. Huy H Nguyen, Fuming Fang, Junichi Yamagishi, and Isao Echizen. 2019. Multi- task learning for detecting and segmenting manipulated facial images and videos. In 2019 IEEE 10th International Conference on Biometrics Theory, Applications and Systems (BTAS). IEEE, 1--8.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Huy H Nguyen, Junichi Yamagishi, and Isao Echizen. 2019. Capsule-forensics: Using capsule networks to detect forged images and videos. In ICASSP 2019--2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2307--2311.Google ScholarGoogle ScholarCross RefCross Ref
  24. Hua Qi, Qing Guo, Felix Juefei-Xu, Xiaofei Xie, Lei Ma, Wei Feng, Yang Liu, and Jianjun Zhao. 2020. Deeprhythm: Exposing deepfakes with attentional visual heartbeat rhythms. In Proceedings of the 28th ACM international conference on multimedia. 4318--4327.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Yuyang Qian, Guojun Yin, Lu Sheng, Zixuan Chen, and Jing Shao. 2020. Thinking in frequency: Face forgery detection by mining frequency-aware clues. In Com- puter Vision--ECCV 2020: 16th European Conference, Glasgow, UK, August 23--28, 2020, Proceedings, Part XII. Springer, 86--103.Google ScholarGoogle Scholar
  26. Jonas Ricker, Simon Damm, Thorsten Holz, and Asja Fischer. 2022. Towards the Detection of Diffusion Model Deepfakes. arXiv preprint arXiv:2210.14571 (2022).Google ScholarGoogle Scholar
  27. Andreas Rossler, Davide Cozzolino, Luisa Verdoliva, Christian Riess, Justus Thies, and Matthias Nießner. 2019. Faceforensics: Learning to detect manipulated facial images. In Proceedings of the IEEE/CVF international conference on computer vision. 1--11.Google ScholarGoogle ScholarCross RefCross Ref
  28. Kaede Shiohara and Toshihiko Yamasaki. 2022. Detecting deepfakes with self- blended images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 18720--18729.Google ScholarGoogle ScholarCross RefCross Ref
  29. Mingxing Tan and Quoc Le. 2019. Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning. PMLR, 6105--6114.Google ScholarGoogle Scholar
  30. Zhengzhong Tu, Hossein Talebi, Han Zhang, Feng Yang, Peyman Milanfar, Alan Bovik, and Yinxiao Li. 2022. Maxvit: Multi-axis vision transformer. In Computer Vision--ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23--27, 2022, Proceedings, Part XXIV. Springer, 459--479.Google ScholarGoogle Scholar
  31. Tete Xiao, Mannat Singh, Eric Mintun, Trevor Darrell, Piotr Dollár, and Ross Girshick. 2021. Early convolutions help transformers see better. Advances in Neural Information Processing Systems 34 (2021), 30392--30400.Google ScholarGoogle Scholar
  32. Hanqing Zhao, Wenbo Zhou, Dongdong Chen, Tianyi Wei, Weiming Zhang, and Nenghai Yu. 2021. Multi-attentional deepfake detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2185--2194.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Exposing Deepfakes using Dual-Channel Network with Multi-Axis Attention and Frequency Analysis

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        IH&MMSec '23: Proceedings of the 2023 ACM Workshop on Information Hiding and Multimedia Security
        June 2023
        190 pages
        ISBN:9798400700545
        DOI:10.1145/3577163

        Copyright © 2023 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 28 June 2023

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • short-paper

        Acceptance Rates

        Overall Acceptance Rate128of318submissions,40%
      • Article Metrics

        • Downloads (Last 12 months)153
        • Downloads (Last 6 weeks)23

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader