We introduce Celeb-DF++ for addressing a pratical and urgent challenge for DeepFake detection, referred to as Generalizable forensics, i.e., detecting a wide range of unseen DeepFake types using a single model.
Celeb-DF++ is extended from our earlier Celeb-DF (CVPR'20) dataset with significantly more diversity, including 22 various DeepFake methods that span across three commonly witnessed scenarios:
- Face-swap (FS) represents that the face of a source individual is replaced with a synthesized face of a target individual, while retaining consistent facial attributions.
- Face-reenactment (FR) involves generating new videos of a target individual driven by the behaviors of a source individual, ensuring behavioral consistency.
- Talking-face (TF) leverages audio input to generate synchronized lip movements in synthesized videos of the target individuals.
To comprehensively measure the generalizability of DeepFake detection methods, we establish three challenging evaluation protocols:
- Generalized Forgery evaluation (GF-eval) assesses detection performance across DeepFake methods in all three forgery scenarios, reflecting real-world conditions where forgeries are diverse and potentially unseen
- Generalized Forgery across Quality evaluation (GFQ-eval) increases the challenge by evaluating the detection performance under varying video compression levels. This aligns with the reality that online videos are often subject to compression, and this compression can obscure forgery traces, raising the challenge for generalizable forensics.
- Generalized Forgery across Datasets evaluation (GFD-eval) measures the generalizability of detection methods across different datasets, simulating the practical case where training and testing data originate from different sources.
With these evaluation protocols, we conduct comprehensive up-to-date evaluations using 24 recent detectors (including 5 recent ones released after 2024).
If you would like to access the Celeb-DF++ dataset, please fill out this form. Once your request is approved, a download link will be provided. For any questions, please contact us at liyuezun@ouc.edu.cn.
If the above form link is not working for you, please try using this form instead (not preferred).
Celeb-DF++
├──Celeb-real
├──YouTube-real
├──Celeb-synthesis
│ ├── FaceSwap
│ │ ├── Celeb-DF (original DeepFake videos in Celeb-DF)
│ │ ├── BlendFace
│ │ ├── GHOST
│ │ ├── HifiFace
│ │ ├── InSwapper
│ │ ├── MobileFaceSwap
│ │ ├── SimSwap
│ │ └── UniFace
│ ├── FaceReenact
│ │ ├── DaGAN
│ │ ├── FSRT
│ │ ├── HyperReenact
│ │ ├── LIA
│ │ ├── LivePortrait
│ │ ├── MCNET
│ │ └── TPSMM
│ └── TalkingFace
│ ├── AniTalker
│ ├── EchoMimic
│ ├── EDTalk
│ ├── FLOAT
│ ├── IP_LAP
│ ├── Real3DPortrait
│ └── SadTalker
└──List_of_testing_videos.txt
For real testing videos, we follow the split of Celeb-DF, where 178 videos are selected. For DeepFake testing videos, we randomly select 200 videos per method in the Face-swap scenario, 200 videos per method in the Face-reenactment scenario, and 300 videos per method in the Talking-face scenario, respectively. The split details can be found in List_of_testing_videos.txt.
If you find this dataset useful for your research, welcome to star this repo and cite our work using the following BibTeX:
@article{li2025celeb,
title={Celeb-DF++: A Large-scale Challenging Video DeepFake Benchmark for Generalizable Forensics},
author={Li, Yuezun and Zhu, Delong and Cui, Xinjie and Lyu, Siwei},
journal={arXiv preprint arXiv:2507.18015},
year={2025}
}
@inproceedings{li2020celeb,
title={Celeb-DF: A Large-scale Challenging Dataset for Deepfake Forensics},
author={Li, Yuezun and Yang, Xin and Sun, Pu and Qi, Honggang and Lyu, Siwei},
booktitle={IEEE Conference on Computer Vision and Pattern Recognition},
year={2020}
}This dataset is released under the Terms to Use Celeb-DF++, which is provided "as it is" and we are not responsible for any subsequence from using this dataset. All original videos of the Celeb-DF dataset are obtained from the Internet which are not property of the authors or the authors’ affiliated institutions. Neither the authors or the authors’ affiliated institution are responsible for the content nor the meaning of these videos. If you feel uncomfortable about your identity shown in this dataset, please contact us and we will remove corresponding information from the dataset.
We would like to thank the contributors of DeepfakeBench, ForensicsAdapter, Effort, ProDet and all related repositories, for their open research and contributions.
Yuezun Li (liyuezun@ouc.edu.cn);
Delong Zhu (zhudelong@stu.ouc.edu.cn)


