Detecting Deepfake Videos with Temporal Dropout 3DCNN

56Citations
Citations of this article
31Readers
Mendeley users who have this article in their library.

Abstract

While the abuse of deepfake technology has brought about a serious impact on human society, the detection of deepfake videos is still very challenging due to their highly photorealistic synthesis on each frame. To address that, this paper aims to leverage the possible inconsistent cues among video frames and proposes a Temporal Dropout 3-Dimensional Convolutional Neural Network (TD-3DCNN) to detect deepfake videos. In the approach, the fixed-length frame volumes sampled from a video are fed into a 3-Dimensional Convolutional Neural Network (3DCNN) to extract features across different scales and identify whether they are real or fake. Especially, a temporal dropout operation is introduced to randomly sample frames in each batch. It serves as a simple yet effective data augmentation and can enhance the representation and generalization ability, avoiding model overfitting and improving detecting accuracy. In this way, the video-level classifier is trained to identify deepfake videos accurately and effectively. Extensive experiments on popular benchmarks clearly demonstrate the effectiveness and generalization capacity of our approach.

Cite

CITATION STYLE

APA

Zhang, D., Li, C., Lin, F., Zeng, D., & Ge, S. (2021). Detecting Deepfake Videos with Temporal Dropout 3DCNN. In IJCAI International Joint Conference on Artificial Intelligence (pp. 1288–1294). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2021/178

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free