How Generalizable are Deepfake Image Detectors? An Empirical Study

Authors: Boquan Li, Jun Sun, Christopher M. Poskitt, Xingmei Wang

Published: 2023-08-08 10:30:34+00:00

AI Summary

This paper presents the first empirical study on the generalizability of deepfake image detectors. The study reveals that existing detectors struggle to generalize to unseen datasets, primarily due to learning unwanted synthesis-method-specific properties rather than extracting discriminative features.

Abstract

Deepfakes are becoming increasingly credible, posing a significant threat given their potential to facilitate fraud or bypass access control systems. This has motivated the development of deepfake detection methods, in which deep learning models are trained to distinguish between real and synthesized footage. Unfortunately, existing detectors struggle to generalize to deepfakes from datasets they were not trained on, but little work has been done to examine why or how this limitation can be addressed. Especially, those single-modality deepfake images reveal little available forgery evidence, posing greater challenges than detecting deepfake videos. In this work, we present the first empirical study on the generalizability of deepfake detectors, an essential goal for detectors to stay one step ahead of attackers. Our study utilizes six deepfake datasets, five deepfake image detection methods, and two model augmentation approaches, confirming that detectors do not generalize in zero-shot settings. Additionally, we find that detectors are learning unwanted properties specific to synthesis methods and struggling to extract discriminative features, limiting their ability to generalize. Finally, we find that there are neurons universally contributing to detection across seen and unseen datasets, suggesting a possible path towards zero-shot generalizability.


Key findings
Deepfake detectors generalize poorly in zero-shot settings but show some improvement in few-shot scenarios. The detectors primarily learn unwanted synthesis method-specific properties, hindering generalization. Analysis reveals the presence of neurons universally contributing to detection across seen and unseen datasets, suggesting a potential path towards zero-shot generalizability.
Approach
The researchers conducted an empirical study using six deepfake image datasets and five deepfake image detection methods. They evaluated the generalizability of these methods in zero-shot and few-shot settings, employing transfer and merge learning augmentation techniques to analyze their performance and identify contributing factors.
Datasets
FaceForensics++, DeepFakeDetection (DFD), CELEB-DF-V2, CELEB-DF-V1 (used to create CELEB-M), specialized datasets FS-I and NT-I.
Model(s)
MesoNet, MesoInception, ShallowNet, XceptionNet, EfficientNet
Author countries
Singapore, China