Entire pipeline of the proposed method. (a) shows the process of dividing the
training dataset into three unbalanced subsets, each with respect to a specific class
(DM, GAN, real) used for training a specific Base Model. (b) illustrates the architecture
of the final model, which takes the three Base Models ϕc trained in the previous phase
with frozen weights, and uses them to extract the features from a digital image $$\phi_{c(I)}$$,
where $$c \in C = {DM, GAN, REAL}$$. These are then concatenated in channel dimension
$$ \phi_(I) = \phi_{DM}(I) \oplus \phi_{GAN}(I) \oplus \phi_{REAL}(I)$$ and processed to solve the classification task.
ABSTRACT
Deepfakes, synthetic images generated by deep learning algorithms, represent one of the biggest challenges in the field of Digital Forensics. The scientific community is working to develop approaches that can discriminate the origin of digital images (real or AI-generated). However, these methodologies face the challenge of generalization, that is, the ability to discern the nature of an image even if it is generated by an architecture not seen during training. This usually leads to a drop in performance. In this context, we propose a novel approach based on three blocks called Base Models, each of which is responsible for extracting the discriminative features of a specific image class (Diffusion Model-generated, GAN-generated, or real) as it is trained by exploiting deliberately unbalanced datasets. The features extracted from each block are then concatenated and processed to discriminate the origin of the input image. Experimental results showed that this approach not only demonstrates good robust capabilities to JPEG compression and other various attacks but also outperforms state-of-the-art methods in several generalization tests. Code, models and dataset are available at https://github.com/opontorno/block-based_deepfake-detection.
![]() |
Download Paper |