1st place solution for The Global Multimedia Deepfake Detection (Image Track) by "JTGroup" team.
The visualization of features in the official split of training and validation sets reveals two key insights: 1) the similarity in distribution between the training and validation sets makes it challenging to evaluate the model's performance on unseen test sets; 2) the feature types of fake data are richer compared to real data. These insights inspired us to initially use unsupervised clustering to partition the challenging validation set, followed by augmenting the types of real and fake data based on adversarial learning principles. These insights led to the development of our proposed generalizable deepfake detection method.
The framework of our proposed method consists of two primary stages: Data Preparing and Training. In the Data Preparing stage, we focus on augmenting the existing dataset by generating new data through image editing and Stable Diffusion (SD) techniques. We then perform clustering to reassemble the dataset, aiming to enhance both the detection performance and the robustness of our method. In the Training stage, we introduce a series of expert models and optimize them using three types of losses:
- torch 1.9.0
- timm 1.0.8
- scikit-learn 1.2.1
- torch-ema 0.3
- torch_kmeans 0.2.0
- albumentations 1.3.1
>> python infer.py
Enter the image path and expect the following result:
>> demo/img_1_should_be_0.0016829653177410.jpg
>> Prediction of [demo/img_1_should_be_0.0016829653177410.jpg] being Deepfake: 0.001683078
Note: The pretrained weights can be downloaded from Google Drive or Baidu Pan.
The training code (e.g., unsupervised clustering and joint optimization loss) and detailed technical report will be available later.
- THIS WORK WAS PERFORMED IN PART AT SICC WHICH IS SUPPORTED BY SKL-IOTSC, UNIVERSITY OF MACAU.
- INCLUSION·CONFERENCE ON THE BUND, THE ORGANIZER OF THE COMPETITION.
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
For inquiries or to obtain permission for commercial use, please contact Yiming Chen ([email protected]).