Skip to content

MI-BioLab/AI-GenBench

Repository files navigation

AI-GenBench: A New Ongoing Benchmark for AI-Generated Image Detection

This is the official repository of the AI-GenBench benchmark, a new benchmark for the detection of AI-generated images.

New paper out: Generalized Design Choices for Deepfake Detectors on arXiv!

Resources:

AI-GenBench

Unlike existing solutions that evaluate models on static datasets, AI-GenBench introduces a temporal evaluation framework where detection methods are incrementally trained on synthetic images, historically ordered by their generative models, to test their ability to generalize to new generative models, such as the transition from GANs to diffusion models.

The goal of AI-GenBench is to provide:

  • a benchmark protocol used to evaluate the ability of detection models to detect images generated using both past and future generation techniques.
  • the related training and evaluation datasets encompassing a wide variety of image generators released in the last 7 years (2017-2024), from older GANs to the most recent diffusion approaches.
  • a framework to train and evaluate detection models on the benchmark, which is based on PyTorch Lightning.

AI-GenBench is an on-going benchmark, which means that we will receive submissions and update the leaderboard over time. Please check the leaderboard section for more information.

In the future, we will also release new versions of the benchmark with the goal to cover the latest datasets, conditioning methods, generation techniques, and so on.

Content

This repository contains the code for:

Getting started

Leaderboard

We maintain a public leaderboard to track the performance of different detection methods on the AI-GenBench benchmark. You can find the leaderboard on the official website.

You may always submit a new entry to the leaderboard by contacting the authors of the papers!

License

Code license

This code is released under the BSD-3-Clause license.

Dataset license

The images are obtained from multiple sources. Please check the dataset_creation/README.md file for more information on the sources. You'll find the list of the datasets websites / repositories and, from there, you will be able to find the license terms for each dataset.

Citing our work

If you use this benchmark and/or code in your research, please cite our paper(s):

AI-GenBench: A New Ongoing Benchmark for AI-Generated Image Detection

Where we firstly introduce AI-GenBench.
Please cite the proceedings version:

@INPROCEEDINGS{pellegrini2025aigenbench,
  author={Pellegrini, Lorenzo and Cozzolino, Davide and Pandolfini, Serafino and Maltoni, Davide and Ferrara, Matteo and Verdoliva, Luisa and Prati, Marco and Ramilli, Marco},
  booktitle={2025 International Joint Conference on Neural Networks (IJCNN)}, 
  title={{AI-GenBench: A New Ongoing Benchmark for AI-Generated Image Detection}}, 
  year={2025},
  volume={},
  number={},
  pages={1-9},
  doi={10.1109/IJCNN64981.2025.11228377}
}

Generalized Design Choices for Deepfake Detectors

Where we analyze in depth the design choices for deepfake detectors using the AI-GenBench benchmark.
Currently under review. For the moment, please cite the arXiv version:

@ARTICLE{pellegrini2024generalized,
  author={Pellegrini, Lorenzo and Pandolfini, Serafino and Maltoni, Davide and Ferrara, Matteo and Prati, Marco and Ramilli, Marco},
  journal={arXiv preprint arXiv:2511.21507}, 
  title={{Generalized Design Choices for Deepfake Detectors}}, 
  year={2024},
}

Credits

We would like to thank identifAI for their valuable contribution to this project.

About

AI-GenBench: A New Ongoing Benchmark for AI-Generated Image Detection

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published