计算机科学
人工智能
块(置换群论)
RGB颜色模型
光学(聚焦)
帧(网络)
卷积神经网络
计算机视觉
模式识别(心理学)
数据挖掘
电信
物理
几何学
数学
光学
作者
Kaihan Lin,Weihong Han,Shudong Li,Zhaoquan Gu,Huimin Zhao,Yangyang Mei
摘要
The widespread dissemination of Deepfake in social networks has posed serious security risks, thus necessitating the development of an effective Deepfake detection technique. Currently, video-based detectors have not been explored as extensively as image-based detectors. Most existing video-based methods only consider temporal features without combining spatial features, and do not mine deeper-level subtle forgeries, resulting in limited detection performance. In this paper, a novel spatiotemporal trident network (STN) is proposed to detect both spatial and temporal inconsistencies of Deepfake videos. Since the large amount of redundant information in Deepfake video frames, we introduce convolutional block attention module (CBAM) on the basis of the I3D network and optimize the structure to make the network better focus on the meaningful information of the input video. Aiming at the defects in the deeper-level subtle forgeries, we designed three feature extraction modules (FEMs) of RGB, optical flow, and noise to further extract deeper video frame information. Extensive experiments on several well-known datasets demonstrate that our method has promising performance, surpassing several state-of-the-art Deepfake video detection methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI