International Joint Conference on Neural Networks (IJCNN 2025) – VERIMEDIA Workshop

The paper WILD: a new in-the-Wild Image Linkage Dataset for synthetic image attribution has been accepted for presentation at 2025 International Joint Conference on Neural Networks (IJCNN 2025) – VERIMEDIA Workshop!(https://sites.unica.it/ff4all/verimediaijcnn/)
The paper introduces WILD, a new dataset designed for synthetic image source attribution, addressing the challenge of identifying which generative model created a given synthetic image. The dataset is divided into two parts:
  • Closed set: Contains 10,000 images from 10 popular commercial and open-source text-to-image generators (e.g., Adobe Firefly, DALL-E 3, Midjourney), with 1,000 images per generator. Images are generated using fixed, randomized prompts to ensure diversity and minimize bias.
  • Open set: Includes 10,000 images from 10 additional generators (e.g., StyleGAN variants, DALL-E 1) to simulate real-world scenarios where unknown models may be encountered.
Half of the images in both sets are post-processed (e.g., compressed, cropped, resized) to test model robustness. We evaluated seven baseline methods (e.g., CLIP-based classifiers, CNNs, Vision Transformers) on WILD. Key findings:
  • CNNs excel on plain images but degrade with post-processing.
  • Vision Transformers (ViT) show greater robustness to post-processing.
  • Open-set attribution remains challenging, with performance dropping significantly for post-processed images.
Thanks to all the forensic researchers from various Italian institutions who contributed to this study! 🙌
IPLAB Authors: Mirko Casu, Orazio Pontorno, Claudio RagagliaLuca GuarneraSebastiano Battiato