Video Deepfake Abuse: How Company Choices Predictably Shape Misuse Patterns

Authors: Max Kamachee, Stephen Casper, Michelle L. Ding, Rui-Jie Yew, Anka Reuel, Stella Biderman, Dylan Hadfield-Menell

Published: 2025-11-26 18:59:43+00:00

AI Summary

This paper analyzes how the widespread misuse patterns observed with open-weight AI image generators are emerging within AI video generation models, leading to low-effort creation of videorealistic AIG-NCII and AIG-CSAM. The authors argue that company choices—specifically releasing open-weight models trained without sufficient data curation or post-training safeguards—foreseeably contribute to mitigatable downstream harm. They conclude that robust risk management by developers and distributors is crucial for limiting the future ease of creating non-consensual deepfake video content.

Abstract

In 2022, AI image generators crossed a key threshold, enabling much more efficient and dynamic production of photorealistic deepfake images than before. This enabled opportunities for creative and positive uses of these models. However, it also enabled unprecedented opportunities for the low-effort creation of AI-generated non-consensual intimate imagery (AIG-NCII), including AI-generated child sexual abuse material (AIG-CSAM). Empirically, these harms were principally enabled by a small number of models that were trained on web data with pornographic content, released with open weights, and insufficiently safeguarded. In this paper, we observe ways in which the same patterns are emerging with video generation models in 2025. Specifically, we analyze how a small number of open-weight AI video generation models have become the dominant tools for videorealistic AIG-NCII video generation. We then analyze the literature on model safeguards and conclude that (1) developers who openly release the weights of capable video generation models without appropriate data curation and/or post-training safeguards foreseeably contribute to mitigatable downstream harm, and (2) model distribution platforms that do not proactively moderate individual misuse or models designed for AIG-NCII foreseeably amplify this harm. While there are no perfect defenses against AIG-NCII and AIG-CSAM from open-weight AI models, we argue that risk management by model developers and distributors, informed by emerging safeguard techniques, will substantially affect the future ease of creating AIG-NCII and AIG-CSAM with generative AI video tools.


Key findings
Analysis showed that a small number of open-weight models, particularly Wan2.x, Stable Video Diffusion, HunyuanVideo, and LTX-Video, are disproportionately used to generate NSFW content online. Empirically effective mitigations like strict training data filtering (as seen with Stable Diffusion 2.0) significantly increase barriers to misuse, demonstrating that AIG-NCII harms are mitigatable even in an open-weight ecosystem. Developers generally fail to report on or implement sufficient safeguards, and model distribution platforms amplify harm by hosting models explicitly designed for AIG-NCII.
Approach
The researchers analyzed usage patterns across online platforms (CivitAI, Reddit) to quantify the SFW vs. NSFW market share of 10 popular open-weight video generation model families. Based on these empirical observations, they assessed the effectiveness of developer and distribution platform safeguards using existing academic literature on data filtering, unlearning, and risk mitigation.
Datasets
Online usage data from CivitAI, Reddit forums (SFW and NSFW subreddits), and a CivitAI model archive site.
Model(s)
Wan2.x, Stable Video Diffusion, HunyuanVideo, LTX-Video, CogVideoX, AnimateDiff-Lightning, Stable Virtual Camera, Cosmos, Mochi 1 (Analyzed generation models)
Author countries
USA