Mastering GenAI Attribution: A Comprehensive Benchmark for Image Generator Identification

Presented by Deep Media AI (See our other research in Deepfake Detection here) Published: 2024 - 08 - 07

Updated: 2024 - 08 - 14 Dataset ID: 240807

https://www.loom.com/share/3ad66853ca084babb72eafc929a271ce?sid=9367d4da-0d7a-429f-a376-a477911a9ac3

Download This Dataset

Fill out this form with your information and someone from the Deep Media research term will grant you access: https://docs.google.com/forms/d/e/1FAIpQLSeTupVdPJDJM6hAEYDCD5ndIu1lvqZhc_xR_woC5hike532hw/viewform?usp=sf_link

Abstract

As AI-generated images flood our digital landscape, the ability to attribute these creations to specific generators has become crucial for maintaining digital integrity and trust. Deep Media AI, a trailblazer in Deepfake Detection, now turns its expertise to the challenge of GenAI Image Generator Attribution.

We are proud to introduce our GenAI Image Generator Attribution Lab Benchmark Dataset, a powerful new tool in the fight against digital misinformation and fraud.

multi_class_generator_distribution.png

Key Features:

Firefly

Firefly

DallE

DallE

MidJourney

MidJourney

Stable Diffusion

Stable Diffusion

Results

Our image classification model demonstrates strong performance in distinguishing between images generated by different AI models (DALLE, MidJourney, StableDiffusion, and Adobe Firefly) and potentially real images. The overall accuracy of 90.54% indicates that the model correctly classifies more than 9 out of 10 images, which is a robust result in the complex domain of AI-generated image detection.

Classification Report:
                 precision    recall  f1-score   support

          DALLE       0.92      0.91      0.92      2000
     MidJourney       0.96      0.95      0.95      2000
StableDiffusion       0.90      0.92      0.91      2000
  Adobe_Firefly       0.84      0.84      0.84      2000
      accuracy                           0.91      8000
      macro avg       0.91      0.91      0.91      8000
   weighted avg       0.91      0.91      0.91      8000

multi_class_correlation_grids.png

multi_class_correlation_grids.png

Examining the confusion matrix and classification report reveals several key insights:

  1. Overall Performance: The model achieves high precision and recall across all classes, with a macro average F1-score of 0.91. This balanced performance suggests that the model is equally capable of identifying images from different generators without significant bias towards any particular class.
  2. Generator-Specific Performance:
  3. Misclassifications: The confusion matrix shows that the most common misclassifications occur between StableDiffusion and Adobe Firefly, with 167 StableDiffusion images misclassified as Adobe Firefly and 133 Adobe Firefly images misclassified as StableDiffusion. This indicates a higher degree of similarity between the outputs of these two generators.

These results are particularly impressive given the challenging nature of distinguishing between different AI-generated images. The high accuracy and balanced performance across multiple generators demonstrate the model's robustness and potential for real-world applications in detecting and classifying AI-generated content.

Areas for Improvement:

  1. Adobe Firefly Detection: Given the lower performance on Adobe Firefly images, future work should focus on identifying unique characteristics of these images to improve classification accuracy. This may involve fine-tuning the model with a larger dataset of Adobe Firefly images or exploring additional features that could help distinguish them from other generators.
  2. Reducing StableDiffusion and Adobe Firefly Confusion: The model shows higher confusion between these two generators. Further analysis of the misclassified images could reveal common patterns or features that lead to this confusion, allowing for targeted improvements in the model's ability to distinguish between these two generators.
  3. Expanding Generator Coverage: While the current model performs well on four major image generators, expanding the model to recognize images from additional AI generators would increase its utility and robustness in real-world scenarios.
  4. Real vs. AI-Generated Distinction: Although not explicitly mentioned in the provided metrics, improving the model's ability to distinguish between real and AI-generated images across all generators remains a crucial area for ongoing research and development.
  5. Adversarial Testing: To further validate the model's robustness, it would be beneficial to test it against adversarial examples or images intentionally designed to fool the classifier. This could uncover potential vulnerabilities and guide further improvements in the model's resilience.

REAL: Firefly - PREDICTED: DallE

REAL: Firefly - PREDICTED: DallE

REAL: Firefly - PREDICTED: Stable Diffusion

REAL: Firefly - PREDICTED: Stable Diffusion

REAL: DallE - PREDICTED: MidJourney

REAL: DallE - PREDICTED: MidJourney

REAL: DallE - PREDICTED: Stable Diffusion

REAL: DallE - PREDICTED: Stable Diffusion

REAL: MidJourney - PREDICTED: Firefly

REAL: MidJourney - PREDICTED: Firefly

REAL: MidJourney - PREDICTED: Stable Diffusion

REAL: MidJourney - PREDICTED: Stable Diffusion

REAL: Stable Diffusion - PREDICTED: DallE

REAL: Stable Diffusion - PREDICTED: DallE

REAL: Stable Diffusion - PREDICTED: MidJourney

REAL: Stable Diffusion - PREDICTED: MidJourney

REAL: Firefly - PREDICTED: Firefly

REAL: Firefly - PREDICTED: Firefly

REAL: MidJourney - PREDICTED: MidJourney

REAL: MidJourney - PREDICTED: MidJourney

REAL: Firefly - PREDICTED: MidJourney

REAL: Firefly - PREDICTED: MidJourney

REAL: DallE - PREDICTED: Firefly

REAL: DallE - PREDICTED: Firefly

REAL: MidJourney - PREDICTED: Firefly

REAL: MidJourney - PREDICTED: Firefly

REAL: Stable Diffusion - PREDICTED: Firefly

REAL: Stable Diffusion - PREDICTED: Firefly

REAL: Stable Diffusion - PREDICTED: Stable Diffusion

REAL: Stable Diffusion - PREDICTED: Stable Diffusion

REAL: DallE - PREDICTED: DallE

REAL: DallE - PREDICTED: DallE