A Sanity Check for AI-generated Image Detection
Abstract
With the rapid development of generative models, discerning AI-generated content has evoked increasing attention from both industry and academia. In this paper, we conduct a sanity check on “whether the task of AI-generated image detection has been solved”. To start with, we present Chameleon dataset, consisting of AI-generated images that are genuinely challenging for human perception. To quantify the generalization of existing methods, we evaluate 9 off-the-shelf AI-generated image detectors on Chameleon dataset. Upon analysis, almost all models misclassify AI-generated images as real ones. Later, we propose AIDE (AI-generated Image DEtector with Hybrid Features), which leverages multiple experts to simultaneously extract visual artifacts and noise patterns. Specifically, to capture the high-level semantics, we utilize CLIP to compute the visual embedding. This effectively enables the model to discern AI-generated images based on semantics and contextual information; Secondly, we select the highest and lowest frequency patches in the image, and compute the low-level patchwise features, aiming to detect AI-generated images by low-level artifacts, for example, noise patterns, anti-aliasing, etc. While evaluating on existing benchmarks, for example, AIGCDetectBenchmark and GenImage, AIDE achieves +3.5% and +4.6% improvements to state-of-the-art methods, and on our proposed challenging Chameleon benchmarks, it also achieves the promising results, despite this problem for detecting AI-generated images is far from being solved.
1 Introduction
Recently, the vision community has witnessed remarkable advancements in generative models. These methods, ranging from generative adversarial networks (GANs) (Goodfellow et al., 2014; Zhu et al., 2017; Brock et al., 2018; Karras et al., 2019) to diffusion models (DMs) (Ho et al., 2020; Nichol & Dhariwal, 2021; Rombach et al., 2022; Song et al., 2020; Liu et al., 2022b; Lu et al., 2022; Hertz et al., 2022; Nichol et al., 2021) have demonstrated unprecedented capabilities in synthesizing high-quality images that closely resemble real-world scenes. On the positive side, such generative models have enabled various valuable tools for artists and designers, democratizing access to advanced graphic design capabilities. However, it also raises concerns about the authenticity of visual content, posing significant challenges for image forensics (Ferreira et al., 2020), misinformation combating (Xu et al., 2023a), and copyright protection (Ren et al., 2024). In this paper, we consider the problem of distinguishing between images generated by AI models and those originating from real-world sources.
In the literature, although there are numerous AI-generated image detectors (Wang et al., 2020; Frank et al., 2020; Ojha et al., 2023; Wang et al., 2023; Zhong et al., 2023; Ricker et al., 2024) and benchmarks (Wang et al., 2020; 2023; Zhu et al., 2024; Hong & Zhang, 2024), the prevailing problem formulation typically involves training models on images generated solely by GANs (e.g., ProGAN (Karras et al., 2017)) and evaluating their performance on datasets including images from various generative models, including GANs and DMs. However, such formulation poses two fundamental issues in practice. Firstly, evaluation benchmarks are simple, as they often feature test sets composed of random images from generative models, rather than images that present genuine challenges for human perception. Secondly, confining models to train exclusively on images from certain type of generative models (GANs or DMs) imposes an unrealistic constraint, hindering the model’s ability to learn from the diverse properties exhibited by more advanced generative models.
To address the aforementioned issues, we propose two pivotal strategies. Firstly, we introduce a novel testset for AI-generated image detection, named Chameleon, manually annotated to include images that genuinely challenge human perception. This dataset has three key features: (i) Deceptively real: all AI-generated images in the dataset have passed a human perception ”Turing Test”, i.e., human annotators have misclassified them as real images. (ii) Diverse categories: comprising images of human, animal, object, and scene categories, the dataset depicts real-world scenarios across various contexts. (iii) High resolution: with most images having resolutions exceeding 720P and going up to 4K, all images in the dataset exhibit exceptional clarity. Consequently, this test set offers a more realistic evaluation of model performance. After evaluating 9 off-the-shelf AI-generated image detectors on Chameleon, unfortunately, all detectors suffer from significant performance drops, misclassifying the AI-generated images as real ones. Secondly, we reformulate the AI-generated image detection problem setup, which enables models to train across a broader spectrum of generative models, enhancing their adaptability and robustness in real-world scenarios.
Based on the above observation, it is clear that detecting AI-generated images remains challenging, and is far from being solved. Therefore, a fundamental question arises: what distinguishes AI-generated images from real ones? Intuitively, such cues may appear from various aspects, including low-level textures or pixel statistics (e.g., the presence of white noise during image capturing), and high-level semantics (e.g., penguins are unlikely to be appearing on the grassland in Africa), geometry principle (e.g., perspective), physics (e.g., lighting condition). To reflect such intuition, we propose a simple AI-generated image detector, termed as AIDE (AI-generated Image DEtector with Hybrid Features). Specifically, AIDE incorporates a DCT (Ahmed et al., 1974) scoring module to capture low-level pixel statistics by extracting both high and low-frequency patches from the image, which are then processed through SRM (Spatial Rich Model) filters (Fridrich & Kodovsky, 2012) to characterize the noise pattern. Additionally, to capture global semantics, we utilize the pre-trained OpenCLIP (Ilharco et al., 2021) to encode the entire image. The features from various levels are fused in the channel dimension for the final prediction. To evaluate the effectiveness of our model, we conduct extensive experiments on two popular benchmarks, including AIGCDetectBenchmark (Wang et al., 2020) and GenImage (Zhu et al., 2024), for AI-generated image detection. On AIGCDetectBenchmark and GenImage benchmarks, AIDE surpasses state-of-the-art (SOTA) methods by +3.5% and +4.6% in accuracy scores, respectively. Moreover, AIDE also achieves competitive performance on our Chameleon benchmark.
Overall, our contributions are summarized as follows: (i) We present the Chameleon dataset, a meticulously curated test set designed to challenge human perception by including images that deceptively resemble real-world scenes. With thorough evaluation of 9 different off-the-shelf detectors, this dataset exposes the limitations of existing approaches. (ii) We present a simple mixture-of-expert model, termed as AIDE, that enables to discern AI-generated images based on both low-level pixel statistics and high-level semantics. (iii) Experimentally, our model achieves state-of-the-art results on public benchmarks for AIGCDetectBenchmark (Wang et al., 2020) and GenImage (Zhu et al., 2024). While on Chameleon, it acts as a competitive baseline on a realistic evaluation benchmark, to foster future research in this community.
2 Related Works
AI-generated image detection. The demand for detecting AI-generated images has long been present. Early studies primarily focus on spatial domain cues, such as color (McCloskey & Albright, 2018), saturation (McCloskey & Albright, 2019), co-occurrence (Nataraj et al., 2019), and reflections (O’brien & Farid, 2012). However, these methods often suffer from limited generalization capabilities as generators progress. To address this limitation, CNNSpot (Wang et al., 2020) discovers that an image classifier trained exclusively on ProGAN (Karras et al., 2017) generator could generalize effectively to other unseen GAN architectures, with careful pre- and post-processing and data augmentation. FreDect (Frank et al., 2020) observes significant artifacts in the frequency domain of GAN-generated images, attributed to the upsampling operation in GAN architectures. More recent approaches have explored novel perspectives for superior generalization ability. UnivFD (Ojha et al., 2023) proposes to train a universal liner classifier with pretrained CLIP-ViT (Dosovitskiy et al., 2020; Radford et al., 2021) features. DIRE (Wang et al., 2023) introduces DIRE features, which computes the difference between images and their reconstructions from pretrained ADM (Dhariwal & Nichol, 2021), to train a deep classifier. PatchCraft (Zhong et al., 2023) compares rich-texture and poor-texture patches from images, extracting the inter-pixel correlation discrepancy as a universal fingerprint, which is reported to achieve the state-of-the-art (SOTA) generalization performance. AEROBLADE (Ricker et al., 2024) proposes a training-free detection method for latent diffusion models using autoencoder reconstruction errors. However, these methods only discriminate real or fake images from a single perspective, often failing to generalize across images from different generators.
AI-generated image datasets. To facilitate AI-generated image detection, many datasets containing both real and fake images have been organized for training and evaluation. Early dataset from CNNSpot (Wang et al., 2020) collects fake images from GAN series generators (Goodfellow et al., 2014; Zhu et al., 2017; Brock et al., 2018; Karras et al., 2019). Particularly, this dataset generates fake images exclusively using ProGAN (Karras et al., 2017) as training data and evaluates the generalization ability on a set of GAN-based testing data. However, with recent emergence of more advanced generators, such as diffusion model (DM) (Ho et al., 2020) and its variants (Dhariwal & Nichol, 2021; Nichol & Dhariwal, 2021; Rombach et al., 2022; Song et al., 2020; Liu et al., 2022b; Lu et al., 2022; Hertz et al., 2022; Nichol et al., 2021), their realistic generations make visual differences between real and fake images progressively harder to detect. Subsequently, more datasets including DM-generated images have been proposed one after another, including DE-FAKE (Xu et al., 2023b), CiFAKE (Bird & Lotfi, 2024), DiffusionDB (Wang et al., 2022), ArtiFact (Rahman et al., 2023). One representative dataset is GenImage (Zhu et al., 2024), which comprises ImageNet’s 1,000 classes generated using 8 SOTA generators in both academia (e.g., Stable Diffusion (Sta, 2022)) and industry (e.g., Midjourney (mid, )). More recently, Hong et al. introduce a more comprehensive dataset, WildFake (Hong & Zhang, 2024), which includes AI-generated images sourced from multiple generators, architectures, weights, and versions. However, existing benchmarks only evaluate AI-generated images using current foundational models with simple prompts and few modifications, whereas deceptively real images from online communities usually necessitate hundreds to thousands of manual parameter adjustments.
3 Chameleon Dataset
3.1 Problem Formulation
In this paper, our goal is to train a computational model that can distinguish the AI-generated images from the ones captured by the camera, i.e., , where denotes an input RGB image, refers to the learnable parameters. For training and testing, we consider the following two settings:
Train-Test Setting-I. In the literature, existing works on detecting AI-generated images (Wang et al., 2020; Frank et al., 2020; Ojha et al., 2023; Wang et al., 2023; Zhong et al., 2023) have exclusively considered the scenario of training on images from single generative model, for example, ProGAN (Karras et al., 2017), or Stable Diffusion (Sta, 2022), and then evaluated on images from various generative models. That is,
(1) |
Generally speaking, such problem formulation poses two critical issues: (i) evaluation benchmarks are simple, as these randomly sampled images from generative models, can be far from being photo-realistic, as shown in Figure 1; (ii) confining models to train exclusively on GAN-generated images imposes an unrealistic constraint, hindering the model’s ability to learn from the diverse properties exhibited by more advanced generative models.
Train-Test Setting-II. Herein, we propose an alternative problem formulation, where the models are allowed to train on images generated from a wide spectrum of generative models, and then tested on images that are genuinely challenging for human perception.
(2) |
refers to our proposed benchmark, as detailed below. We believe this setting resembles more practical scenario for future model development in this community.
3.2 Chameleon Dataset
The primary objective of the Chameleon dataset is to evaluate the generalization and robustness of existing AI-generated image detectors, for a sanity check on the progress of AI-generated image detection. In this section, we outline the progression of the proposed dataset in three critical phases: (i) dataset collection, (ii) dataset curation, (iii) dataset annotation, and (iv) dataset comparison. The statistical results of our dataset are illustrated in Table 1 and we compare our dataset with existing benchmarks in Fig. 1.
Real Images | Fake Images | |
Scene | 3,574 | 2,976 |
Object | 3,578 | 2,016 |
Animal | 3,998 | 313 |
Human | 3,713 | 5,865 |
Total | 14,863 | 11,170 |
3.2.1 Dataset Collection
To simulate real-world cases on detecting AI-generated images, we structure our Chameleon dataset based on three main principles: (i) images must be deceptively real, and (ii) they should cover a diverse range of categories, and (iii) they should also have very high image quality. Importantly, each image must have (a) Creative Commons (CC BY 4.0) license, or (b) explicit permissions obtained from the owners to use in our research. Herein, we present the details of image collection.
Fake Image Collection: To collect images that are deceptively real, and cover sufficiently diverse categories, we source user-created AI-generated images from popular AI-painting communities (i.e., ArtStation (art, ), Civitai (civ, ), and Liblib (lib, )), many of which utilize commercial APIs (e.g., Midjourney (mid, ) and DALLE-3 (Ramesh et al., 2022)) or various LoRA modules (Hu et al., 2021) with Stable Diffusion (SD) (Sta, 2022) that fine-tuned on their in-house data. Specifically, we initiate the process by utilizing GPT-4 (cha, 2022) to generate diverse query words to retrieve AI-generated images. Throughout the collection process, we enforce stringent NSFW (Not Safe For Work) restrictions. Ultimately, our collection comprises over 150K fake images.
Real Image Collection: To ensure that real and fake images fall into the same distribution, we employ identical query words to retrieve real images, mirroring the approach used for gathering AI-generated images. Eventually, we collect over 20K images from platforms like Unsplash (uns, ), which is an online community providing high-quality, free-to-use images contributed by photographers worldwide.
3.2.2 Dataset Curation
To ensure the collection of high-quality images, we implement a comprehensive pipeline for image cleaning: (i) we discard images with resolution lower than , as higher-resolution images generally provide better assessments of the robustness of existing models; (ii) due to the potential presence of violent and inappropriate content, we utilize SD’s safety checker model (saf, 2022) to filter out NSFW images; (iii) to avoid duplicated images, we compare their hash values to filter out duplicated images. In addition to this general cleaning pipeline, we introduce CLIP (Radford et al., 2021) to further filter out images with low image-text similarity. Specifically, for fake images, the online website provides prompts used to generate these images, and we calculate similarity using their corresponding prompts. For real images, we used the mean of the 80 prompt templates (e.g., a photo of {category} and a photo of the {category}) evaluated in CLIP’s ImageNet zero-shot as the text embedding.
3.2.3 Dataset Annotation
At this stage, we establish an annotation platform and recruit 20 human workers to manually label each of the AI-generated images for their category and realism. For categorization, annotators are instructed to assign each image to one of four major categories: human, animal, object, and scene. Regarding realism assessment, workers are tasked with labeling the images as either Real or AI-generated, based on the criterion of “whether this image could be taken with a camera”. It’s important to note that as the annotators are not informed whether the images are generated by AI algorithms beforehand. Each image was assessed independently by two annotators, and those have been misclassified as real by both annotators can thus be considered to pass the “perception turing test” and labeled as “highly realistic”. Subsequently, we retain only those images judged as “highly realistic”. Similarly, for real images, we follow the same procedure, retaining only those belonging to the four predefined categories, as we have done for AI-generated images.
3.2.4 Dataset Comparison
Our objective is to construct a sophisticated and exhaustive test dataset for AI-generated image detection. In Table 2, we conduct a comparative analysis between our Chameleon dataset and existing datasets. Our dataset is characterized by three pivotal features: Magnitude. Comprising about 26,000 test images, the Chameleon dataset represents the most extensive collection available, enhancing its robustness. Variety. Our dataset incorporates images from a vast array of real-world scenarios, surpassing the limited categorical focus of other datasets. Resolution. With resolutions spanning from 720P to 4K, With image resolutions ranging from 720P to 4K, artifacts demand more nuanced analysis, thus presenting additional challenges to the model due to the need for fine-grained discernment. In summary, our dataset offers a more demanding and pragmatically relevant benchmark for the advancement of AI-generated image detection methodologies.
ProGAN |
StyleGAN |
BigGAN |
CycleGAN |
StarGAN |
GauGAN |
StyleGAN2 |
WFIR |
ADM |
Glide |
Midjourney |
SD v1.4 |
SD v1.5 |
VQDM |
Wukong |
DALLE2 |
Chameleon |
|
Magnitude | 8.0 | 12.0 | 4.0 | 2.6k | 4.0 | 10.0 | 15.9k | 2.0 | 12.0 | 12.0 | 12.0 | 12.0 | 16.0 | 12.0 | 12.0 | 2.0 | 26.0 |
Resolution | 256 | 256 | 256 | 256 | 256 | 256 | 256 | 1024 | 256 | 256 | 1024 | 512 | 512 | 256 | 512 | 256 | 720P-4K |
Variety | LSUN | LUSN | ImageNet | ImageNet | CelebA | COCO | LSUN | FFHQ | ImageNet | ImageNet | ImageNet | ImageNet | ImageNet | ImageNet | ImageNet | ImageNet | Real-life |
4 Methodology
In this section, we present AIDE (AI-generated Image DEtector with Hybrid Features), consisting of a module to compute patchwise low-level statistics of texture or smooth patches, a high-level semantic embedding module, and a discriminator to classify the image as being generated or photographed. The overview of our AIDE model is illustrated in Fig. 2.
4.1 Patchwise Feature Extraction
We leverage insights from the disparities in low-level patch statistics between AI-generated images and real-world scenes. Models like generative adversarial networks or diffusion models often yield images with certain artifacts, such as excessive smoothness or anti-aliasing effects. To capture such discrepancy, we adopt a Discrete Cosine Transform (DCT) score module to identify patches with the highest and lowest frequency. By focusing on these extreme patches, we aim to highlight the distinctive characteristics of AI-generated images, thus facilitating the discriminative power of our detection system.
Patch Selection via DCT Scoring. For an RGB image, we first divide this image into multiple patches with a fixed window size, , . In our case, the patch size is set to be pixels. We apply the discrete cosine transform to each of the image patches, obtaining the corresponding results in the frequency domain, , .
To acquire the highest and lowest image patches, we use the complexity of the frequency components as an indicator. From this, we design a simple yet effective scoring mechanism. Specifically, we design different band-pass filters:
(3) |
where is the weight at the position of the -th filter. Next, for -th patch , we apply the filters to multiply the logarithm of the absolute DCT coefficients and sum all the positions to obtain the grade of the patch . We formulated it as
(4) |
where is the number of patch channels. In this way, we acquire the grades of all patches . We then sort them to identify the highest and lowest frequency patches.
Through the scoring module, we can obtain the top patches with the highest frequency and the top patches with the lowest frequency, where , .
Patchwise Feature Encoder. Next, firstly, these patches are resized to a size of . Secondly, they are input into the SRM (Fridrich & Kodovsky, 2012) to extract their noise pattern. Lastly, these features are input into two ResNet-50 (He et al., 2016) networks ( and ) to obtain the final feature map , . We acquire the highest frequency embedding and lowest frequency embedding on the mean-pooled feature:
(5) |
4.2 Semantic Feature Embedding
To capture the rich semantic features within images, such as object co-occurrence and contextual relationships, we compute the visual embedding for input image with an off-the-shelf visual-language foundation model. Specifically, we adopt the ConvNeXt-based OpenCLIP model (Ilharco et al., 2021) to get the final feature map (). To capture the global contexts, we append a linear projection layer followed by mean spatial pooling:
(6) |
4.3 Discriminator
To distinguish between AI-generated images and real images, we utilize a mixture-expert-model for the final discrimination. At low-level, we take the average of the highest frequency featured:
(7) |
Then, we channel-wisely concatenate the representations between it and high-level embedding . At last, the features are encoded into MLP to acquire the score,
(8) |
where denotes the MLP consisting of a linear layer, GELU (Hendrycks & Gimpel, 2016) and classifier, refers to the operation of channel-wise concatenation.
5 Experiments
5.1 Experimental Details
Detectors. We evaluate 9 off-the-shelf detectors including CNNSpot (Wang et al., 2020), FreDect (Frank et al., 2020), Fusing (Ju et al., 2022), LNP (Liu et al., 2022a), LGrad (Tan et al., 2023), UnivFD (Ojha et al., 2023), DIRE (Wang et al., 2023), PatchCraft (Zhong et al., 2023) and NPR (Tan et al., 2024) for comparison.
Datasets. To comprehensively evaluate the generalization ability of existing approaches, we conduct experiments across two kinds of settings: Setting-I and Setting-II, which are summarized in Sec. 3.1. For the Setting-I setting, we evaluate the detectors on two general and comprehensive benchmarks of AIGCDetectBenchmark () (Zhong et al., 2023) and GenImage (Zhu et al., 2024) (). For the Setting-II setting, we evaluate the detectors on our Chameleon () benchmark. More details can be found in Appendix.
Implementation Details. AIDE includes two key modules: Patchwise Feature Extraction (PFE) and Semantic Feature Embedding (SFE). For PFE channel, we first patchify each image into patches and the patch size is set to be pixels. Then these patches are sorted using our DCT Scouring module with different band-pass filters in the frequency domain. Subsequently, we select two highest-frequency and two lowest-frequency patches using the calculated DCT scores. These selected patches are then resized to and extracted their noise pattern using SRM (Fridrich & Kodovsky, 2012). For SFE channel, we use the pre-trained OpenCLIP (Ilharco et al., 2021) to extract semantic features. We adopt data augmentations including random JPEG compression (QF ) and random Gaussian blur () to improve the robustness of detectors. Each augmentation is conducted with 10% probability. During the training phase, we use AdamW optimizer with the learning rate of in and in , respectively. The batch size is set to and the model is trained on 8 NVIDIA A100 GPUs for only 5 epochs. Our method trains very quickly, only 2 hours are sufficient.
Metrics. In accordance with existing AI-generated detection arpproaches (Wang et al., 2020; 2019; Zhou et al., 2018), we report both classification accuracy (Acc) and average precision (AP) in our experiments. All results are averaged over both real and AI-generated images unless otherwise specified. We primarily report Acc for evaluation and comparison in the main paper, and AP results are presented in the Appendix.
5.2 Comparison to State-of-the-art Models
On Benchmark AIGCDetectBenchmark: The quantitative results in Table 3 present the classification accuracies of various methods and generators within . In this evaluation, all methods, except for DIRE-D, were exclusively trained on ProGAN-generated data.
AIDE demonstrates a significant advancement over the current state-of-the-art (SOTA) approach, PatchCraft, achieving an average accuracy increase of 3.5%. UnivFD utilizes CLIP semantic features for detecting AI-generated images, proving effective for GAN-generated images. However, it shows pronounced performance degradation with diffusion model (DM)-generated images. This suggests that as generation quality improves, diffusion models produce images with fewer semantic artifacts, as depicted in Fig. 1 (a). Our approach, which integrates semantic, low-frequency, and high-frequency information at the feature level, enhances detection performance, yielding a 5.2% increase for GAN-based images and a 1.7% increase for DM-based images compared to the SOTA method.
Method |
ProGAN |
StyleGAN |
BigGAN |
CycleGAN |
StarGAN |
GauGAN |
StyleGAN2 |
WFIR |
ADM |
Glide |
Midjourney |
SD v1.4 |
SD v1.5 |
VQDM |
Wukong |
DALLE2 |
Mean |
CNNSpot | 100.00 | 90.17 | 71.17 | 87.62 | 94.60 | 81.42 | 86.91 | 91.65 | 60.39 | 58.07 | 51.39 | 50.57 | 50.53 | 56.46 | 51.03 | 50.45 | 70.78 |
FreDect | 99.36 | 78.02 | 81.97 | 78.77 | 94.62 | 80.57 | 66.19 | 50.75 | 63.42 | 54.13 | 45.87 | 38.79 | 39.21 | 77.80 | 40.30 | 34.70 | 64.03 |
Fusing | 100.00 | 85.20 | 77.40 | 87.00 | 97.00 | 77.00 | 83.30 | 66.80 | 49.00 | 57.20 | 52.20 | 51.00 | 51.40 | 55.10 | 51.70 | 52.80 | 68.38 |
LNP | 99.67 | 91.75 | 77.75 | 84.10 | 99.92 | 75.39 | 94.64 | 70.85 | 84.73 | 80.52 | 65.55 | 85.55 | 85.67 | 74.46 | 82.06 | 88.75 | 83.84 |
LGrad | 99.83 | 91.08 | 85.62 | 86.94 | 99.27 | 78.46 | 85.32 | 55.70 | 67.15 | 66.11 | 65.35 | 63.02 | 63.67 | 72.99 | 59.55 | 65.45 | 75.34 |
UnivFD | 99.81 | 84.93 | 95.08 | 98.33 | 95.75 | 99.47 | 74.96 | 86.90 | 66.87 | 62.46 | 56.13 | 63.66 | 63.49 | 85.31 | 70.93 | 50.75 | 78.43 |
DIRE-G | 95.19 | 83.03 | 70.12 | 74.19 | 95.47 | 67.79 | 75.31 | 58.05 | 75.78 | 71.75 | 58.01 | 49.74 | 49.83 | 53.68 | 54.46 | 66.48 | 68.68 |
DIRE-D | 52.75 | 51.31 | 49.70 | 49.58 | 46.72 | 51.23 | 51.72 | 53.30 | 98.25 | 92.42 | 89.45 | 91.24 | 91.63 | 91.90 | 90.90 | 92.45 | 71.53 |
PatchCraft | 100.00 | 92.77 | 95.80 | 70.17 | 99.97 | 71.58 | 89.55 | 85.80 | 82.17 | 83.79 | 90.12 | 95.38 | 95.30 | 88.91 | 91.07 | 96.60 | 89.31 |
NPR | 99.79 | 97.70 | 84.35 | 96.10 | 99.35 | 82.50 | 98.38 | 65.80 | 69.69 | 78.36 | 77.85 | 78.63 | 78.89 | 78.13 | 76.11 | 64.90 | 82.91 |
AIDE | 99.99 | 99.64 | 83.95 | 98.48 | 99.91 | 73.25 | 98.00 | 94.20 | 93.43 | 95.09 | 77.20 | 93.00 | 92.85 | 95.16 | 93.55 | 96.60 | 92.77 |
Method | Midjourney | SD v1.4 | SD v1.5 | ADM | GLIDE | Wukong | VQDM | BigGAN | Mean |
ResNet-50 (He et al., 2016) | 54.90 | 99.90 | 99.70 | 53.50 | 61.90 | 98.20 | 56.60 | 52.00 | 72.09 |
DeiT-S (Touvron et al., 2021) | 55.60 | 99.90 | 99.80 | 49.80 | 58.10 | 98.90 | 56.90 | 53.50 | 71.56 |
Swin-T (Liu et al., 2021) | 62.10 | 99.90 | 99.80 | 49.80 | 67.60 | 99.10 | 62.30 | 57.60 | 74.78 |
CNNSpot (Wang et al., 2020) | 52.80 | 96.30 | 95.90 | 50.10 | 39.80 | 78.60 | 53.40 | 46.80 | 64.21 |
Spec (Zhang et al., 2019) | 52.00 | 99.40 | 99.20 | 49.70 | 49.80 | 94.80 | 55.60 | 49.80 | 68.79 |
F3Net (Qian et al., 2020) | 50.10 | 99.90 | 99.90 | 49.90 | 50.00 | 99.90 | 49.90 | 49.90 | 68.69 |
GramNet (Liu et al., 2020) | 54.20 | 99.20 | 99.10 | 50.30 | 54.60 | 98.90 | 50.80 | 51.70 | 69.85 |
DIRE (Wang et al., 2023) | 60.20 | 99.90 | 99.80 | 50.90 | 55.00 | 99.20 | 50.10 | 50.20 | 70.66 |
UnivFD (Ojha et al., 2023) | 73.20 | 84.20 | 84.00 | 55.20 | 76.90 | 75.60 | 56.90 | 80.30 | 73.29 |
GenDet (Zhu et al., 2023) | 89.60 | 96.10 | 96.10 | 58.00 | 78.40 | 92.80 | 66.50 | 75.00 | 81.56 |
PatchCraft (Zhong et al., 2023) | 79.00 | 89.50 | 89.30 | 77.30 | 78.40 | 89.30 | 83.70 | 72.40 | 82.30 |
AIDE | 79.38 | 99.74 | 99.76 | 78.54 | 91.82 | 98.65 | 80.26 | 66.89 | 86.88 |
On Benchmark GenImage: In the experiments conducted on , all models were trained on SD v1.4 and evaluated across eight contemporary generators. Table 4 presents the results, illustrating our method’s superior performance over the current state-of-the-art, PatchCraft, with a 4.6% improvement in average accuracy. The architectural similarities between SD v1.5, Wukong, and SD v1.4, as noted by GenImage (Zhu et al., 2024), enable models to achieve near-perfect accuracy, approaching 100% on such datasets. Thus, evaluating generalization performance across other generators, such as Midjourney, ADM, and Glide, becomes essential. Our model demonstrates either the best or second-best performance in these cases, achieving an average accuracy of 86.88%.
Training Dataset | CNNSpot | FreDect | Fusing | GramNet | LNP | UnivFD | DIRE | PatchCraft | NPR | AIDE |
ProGAN | 56.94 | 55.62 | 56.98 | 58.94 | 57.11 | 57.22 | 58.19 | 53.76 | 57.29 | 56.45 |
0.08/99.67 | 13.72/87.12 | 0.01/99.79 | 4.76/99.66 | 0.09/99.97 | 3.18/97.83 | 3.25/99.48 | 1.78/92.82 | 2.20/98.70 | 0.63/98.46 | |
SD v1.4 | 60.11 | 56.86 | 57.07 | 60.95 | 55.63 | 55.62 | 59.71 | 56.32 | 58.13 | 61.10 |
8.86/98.63 | 1.37/98.57 | 0.00/99.96 | 17.65/93.50 | 0.57/97.01 | 74.97/41.09 | 11.86/95.67 | 3.07/96.35 | 2.43/100.00 | 16.82/94.38 | |
All GenImage | 60.89 | 57.22 | 57.09 | 59.81 | 58.52 | 60.42 | 57.83 | 55.70 | 57.81 | 63.89 |
9.86/99.25 | 0.89/99.55 | 0.02/99.98 | 8.23/98.58 | 7.72/96.70 | 85.52/41.56 | 2.09/99.73 | 1.39/96.52 | 1.68/100.00 | 22.40/95.06 |
On Benchmark Chameleon: As highlighted in Sec. 1, we contend that success on existing public benchmarks may not accurately reflect real-world scenarios or the advancement in AI-generated image detection, given that test sets are typically randomly sampled from generative models without ”Turing Test”. To address potential biases related to training setups—such as generator types and image categories—we evaluate the performance of existing detectors under diverse training conditions. Despite their high performance on existing benchmarks, as depicted in Fig. 3, the state-of-the-art detector, PatchCraft, experiences substantial performance declines. Additionally, Table 5 reveals significant performance decreases across all methods, with most struggling to surpass an average accuracy close to random guessing (about 50%), indicating a failure in these contexts.
While our method achieves state-of-the-art results on available datasets, its performance on Chameleon remains lacking. This underscores that our dataset, Chameleon, which challenges human perception, represents a critical issue requiring attention in this field.
Method | JPEG Compression | Gaussian Blur | ||
QF=95 | QF=90 | |||
CNNSpot | 64.03 | 62.26 | 68.39 | 67.26 |
FreDect | 66.95 | 67.45 | 65.75 | 66.48 |
Fusing | 62.43 | 61.39 | 68.09 | 66.69 |
GramNet | 65.47 | 64.94 | 68.63 | 68.09 |
LNP | 53.58 | 54.09 | 67.91 | 66.42 |
LGrad | 51.55 | 51.39 | 71.73 | 69.12 |
DIRE-G | 66.49 | 66.12 | 64.00 | 63.09 |
UnivFD | 74.10 | 74.02 | 70.31 | 68.29 |
PatchCraft | 72.48 | 71.41 | 75.99 | 74.90 |
AIDE | 75.54 | 74.21 | 81.88 | 80.35 |
5.3 Robustness to Unseen Perturbations
In real-world scenarios, images often encounter unseen perturbations during transmission and interaction, complicating the detection of AI-generated images. Here, we assess the performance of various methods in handling potential perturbations, such as JJPEG compression (Quality Factor (QF) = 95, 90) and Gaussian blur ( = 1.0, 2.0). As illustrated in Table 6, all methods exhibit a decline in performance due to disruptions in the pixel distribution. These disruptions diminish the discriminative artifacts left by generative models, complicating the differentiation between real and AI-generated images. Consequently, the robustness of these detectors in identifying AI-generated images is significantly compromised. Despite these challenging conditions, our method consistently outperforms others, maintaining a relatively higher accuracy in detecting AI-generated images. This superior performance is attributed to our model’s ability to effectively capture and leverage multi-perspective features, semantics and noise, even when the pixel distribution is distorted.
5.4 Ablation Studies
Our method focuses on detecting AI-generated images with mixture of experts, namely patchwise feature extraction (PFE-H and PFE-L for high-frequency and low-frequency patches, respectively) and semantic feature extraction (SFE). These modules collectively contribute to comprehensively identifying AI-generated images from different perspectives. Herein, we conduct extensive experiments to investigate the roles of each module on .
Module | Mean | ||
PFE-H | PFE-L | SFE | |
✓ | ✗ | ✗ | 76.09 |
✗ | ✓ | ✗ | 75.24 |
✗ | ✗ | ✓ | 75.26 |
✓ | ✓ | ✗ | 76.70 |
✓ | ✗ | ✓ | 80.69 |
✗ | ✓ | ✓ | 84.20 |
✓ | ✓ | ✓ | 92.77 |
Patchwise Feature Extraction. As shown in Table 7, removing either the high-frequency or the low-frequency patches results in obvious performance degradation in terms of accuracy. Without the high-frequency patches, the proposed method is unable to discern that the high-frequency regions of AI-generated images are smoother than those of real images, resulting in performance degradation. Similarly, without the low-frequency patches, the method cannot extract the underlying noise information, which is crucial for identifying AI-generated images with higher fidelity, leading to incorrect predictions.
Semantic Feature Extraction. As shown in Table 7, the performance degrades significantly (76.70% vs 92.77%) when we remove the semantic branch. Intuitively, if the branch for semantic information extraction is absent, our method struggles to effectively capture images with semantic artifacts, resulting in significant drops.
Visualization. To vividly demonstrate the effectiveness of our modules patchwise feature extraction (PFE) and semantic feature extraction (SFE), we conducted a visualization, as depicted in Fig. 4. In first row, the absence of semantic feature extraction results in many images with evident semantic errors going undetected. Similarly, the second row shows that, without patchwise feature extraction, numerous images lacking semantic errors still contain differing underlying information that remains unrecognized. Overall, our method, AIDE, achieves the best performance.
6 Conclusion
In this paper, we have conducted a sanity check on detecting AI-generated images. Specifically, we re-examined the unreasonable assumption in existing training and testing settings and suggested new ones. In terms of benchmarks, we propose a novel, challenging benchmark, termed as Chameleon, which is manually annotated to challenge human perception. We evaluate 9 off-the-shelf models and demonstrate that all detectors suffered from significant performance declines. In terms of architecture, we propose a simple yet effective detector, AIDE, that simultaneously incorporates low-level patch statistics and high-level semantics for AI-generated image detection. Despite our approach demonstrates state-of-the-art performance on existing (AIGCDetectBenchmark (Zhong et al., 2023) and GenImage (Zhu et al., 2024)) and our proposed benchmark (Chameleon) compared to previous detectors, it leaves significant room for future improvement.
Potential societal impacts. Given that Chameleon demonstrates the capability to surpass the ”Turing Test”, there exists a significant risk of exploitation by malicious entities who may utilize AI-generated imagery to engineer fictitious social media profiles and propagate misinformation. To mitigate it, we will require all users of Chameleon to enter into an End-User License Agreement (EULA). Access to the dataset will be contingent upon a thorough review and subsequent approval of the signed agreement, thereby ensuring compliance with established ethical usage protocols.
References
- (1) Artstation. https://www.artstation.com.
- (2) Civitai. https://civitai.com.
- (3) Liblib. https://www.liblib.art.
- (4) Midjourney. https://www.midjourney.com/home.
- (5) Unsplash. https://unsplash.com.
- Sta (2022) Stable diffusion. https://github.com/Stability-AI/StableDiffusion, 2022.
- cha (2022) Chatgpt. https://chatgpt.com, 2022.
- saf (2022) Stable diffusion safety checker. https://huggingface.co/CompVis/stable-diffusion-safety-checker, 2022.
- Ahmed et al. (1974) Nasir Ahmed, T_ Natarajan, and Kamisetty R Rao. Discrete cosine transform. IEEE transactions on Computers, 100(1):90–93, 1974.
- Bird & Lotfi (2024) Jordan J Bird and Ahmad Lotfi. Cifake: Image classification and explainable identification of ai-generated synthetic images. IEEE Access, 2024.
- Brock et al. (2018) Andrew Brock, Jeff Donahue, and Karen Simonyan. Large scale gan training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096, 2018.
- Dhariwal & Nichol (2021) Prafulla Dhariwal and Alexander Nichol. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780–8794, 2021.
- Dosovitskiy et al. (2020) Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- Ferreira et al. (2020) William D Ferreira, Cristiane BR Ferreira, Gelson da Cruz Júnior, and Fabrizzio Soares. A review of digital image forensics. Computers & Electrical Engineering, 85:106685, 2020.
- Frank et al. (2020) Joel Frank, Thorsten Eisenhofer, Lea Schönherr, Asja Fischer, Dorothea Kolossa, and Thorsten Holz. Leveraging frequency analysis for deep fake image recognition. In International conference on machine learning, pp. 3247–3258. PMLR, 2020.
- Fridrich & Kodovsky (2012) Jessica Fridrich and Jan Kodovsky. Rich models for steganalysis of digital images. IEEE Transactions on information Forensics and Security, 7(3):868–882, 2012.
- Goodfellow et al. (2014) Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. Advances in neural information processing systems, 27, 2014.
- He et al. (2016) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.
- Hendrycks & Gimpel (2016) Dan Hendrycks and Kevin Gimpel. Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415, 2016.
- Hertz et al. (2022) Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. Prompt-to-prompt image editing with cross attention control. arXiv preprint arXiv:2208.01626, 2022.
- Ho et al. (2020) Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840–6851, 2020.
- Hong & Zhang (2024) Yan Hong and Jianfu Zhang. Wildfake: A large-scale challenging dataset for ai-generated images detection. arXiv preprint arXiv:2402.11843, 2024.
- Hu et al. (2021) Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685, 2021.
- Ilharco et al. (2021) Gabriel Ilharco, Mitchell Wortsman, Ross Wightman, Cade Gordon, Nicholas Carlini, Rohan Taori, Achal Dave, Vaishaal Shankar, Hongseok Namkoong, John Miller, Hannaneh Hajishirzi, Ali Farhadi, and Ludwig Schmidt. Openclip, July 2021. URL https://doi.org/10.5281/zenodo.5143773. If you use this software, please cite it as below.
- Ju et al. (2022) Yan Ju, Shan Jia, Lipeng Ke, Hongfei Xue, Koki Nagano, and Siwei Lyu. Fusing global and local features for generalized ai-synthesized image detection. In 2022 IEEE International Conference on Image Processing (ICIP), pp. 3465–3469. IEEE, 2022.
- Karras et al. (2017) Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of gans for improved quality, stability, and variation. arXiv preprint arXiv:1710.10196, 2017.
- Karras et al. (2019) Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 4401–4410, 2019.
- Liu et al. (2022a) Bo Liu, Fan Yang, Xiuli Bi, Bin Xiao, Weisheng Li, and Xinbo Gao. Detecting generated images by real images. In European Conference on Computer Vision, pp. 95–110. Springer, 2022a.
- Liu et al. (2022b) Luping Liu, Yi Ren, Zhijie Lin, and Zhou Zhao. Pseudo numerical methods for diffusion models on manifolds. arXiv preprint arXiv:2202.09778, 2022b.
- Liu et al. (2021) Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 10012–10022, 2021.
- Liu et al. (2020) Zhengzhe Liu, Xiaojuan Qi, and Philip HS Torr. Global texture enhancement for fake face detection in the wild. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8060–8069, 2020.
- Lu et al. (2022) Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. Dpm-solver: A fast ode solver for diffusion probabilistic model sampling in around 10 steps. Advances in Neural Information Processing Systems, 35:5775–5787, 2022.
- McCloskey & Albright (2018) Scott McCloskey and Michael Albright. Detecting gan-generated imagery using color cues. arXiv preprint arXiv:1812.08247, 2018.
- McCloskey & Albright (2019) Scott McCloskey and Michael Albright. Detecting gan-generated imagery using saturation cues. In 2019 IEEE international conference on image processing (ICIP), pp. 4584–4588. IEEE, 2019.
- Nataraj et al. (2019) Lakshmanan Nataraj, Tajuddin Manhar Mohammed, Shivkumar Chandrasekaran, Arjuna Flenner, Jawadul H Bappy, Amit K Roy-Chowdhury, and BS Manjunath. Detecting gan generated fake images using co-occurrence matrices. arXiv preprint arXiv:1903.06836, 2019.
- Nichol et al. (2021) Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob McGrew, Ilya Sutskever, and Mark Chen. Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741, 2021.
- Nichol & Dhariwal (2021) Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. In International conference on machine learning, pp. 8162–8171. PMLR, 2021.
- O’brien & Farid (2012) James F O’brien and Hany Farid. Exposing photo manipulation with inconsistent reflections. ACM Trans. Graph., 31(1):4–1, 2012.
- Ojha et al. (2023) Utkarsh Ojha, Yuheng Li, and Yong Jae Lee. Towards universal fake image detectors that generalize across generative models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 24480–24489, 2023.
- Qian et al. (2020) Yuyang Qian, Guojun Yin, Lu Sheng, Zixuan Chen, and Jing Shao. Thinking in frequency: Face forgery detection by mining frequency-aware clues. In European conference on computer vision, pp. 86–103. Springer, 2020.
- Radford et al. (2021) Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748–8763. PMLR, 2021.
- Rahman et al. (2023) Md Awsafur Rahman, Bishmoy Paul, Najibul Haque Sarker, Zaber Ibn Abdul Hakim, and Shaikh Anowarul Fattah. Artifact: A large-scale dataset with artificial and factual images for generalizable and robust synthetic image detection. In 2023 IEEE International Conference on Image Processing (ICIP), pp. 2200–2204. IEEE, 2023.
- Ramesh et al. (2022) Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125, 1(2):3, 2022.
- Ren et al. (2024) Jie Ren, Han Xu, Pengfei He, Yingqian Cui, Shenglai Zeng, Jiankun Zhang, Hongzhi Wen, Jiayuan Ding, Hui Liu, Yi Chang, et al. Copyright protection in generative ai: A technical perspective. arXiv preprint arXiv:2402.02333, 2024.
- Ricker et al. (2024) Jonas Ricker, Denis Lukovnikov, and Asja Fischer. Aeroblade: Training-free detection of latent diffusion images using autoencoder reconstruction error. arXiv preprint arXiv:2401.17879, 2024.
- Rombach et al. (2022) Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10684–10695, 2022.
- Song et al. (2020) Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502, 2020.
- Tan et al. (2023) Chuangchuang Tan, Yao Zhao, Shikui Wei, Guanghua Gu, and Yunchao Wei. Learning on gradients: Generalized artifacts representation for gan-generated images detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12105–12114, 2023.
- Tan et al. (2024) Chuangchuang Tan, Yao Zhao, Shikui Wei, Guanghua Gu, Ping Liu, and Yunchao Wei. Rethinking the up-sampling operations in cnn-based generative network for generalizable deepfake detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 28130–28139, 2024.
- Touvron et al. (2021) Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Hervé Jégou. Training data-efficient image transformers & distillation through attention. In International conference on machine learning, pp. 10347–10357. PMLR, 2021.
- Wang et al. (2019) Sheng-Yu Wang, Oliver Wang, Andrew Owens, Richard Zhang, and Alexei A Efros. Detecting photoshopped faces by scripting photoshop. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10072–10081, 2019.
- Wang et al. (2020) Sheng-Yu Wang, Oliver Wang, Richard Zhang, Andrew Owens, and Alexei A Efros. Cnn-generated images are surprisingly easy to spot… for now. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8695–8704, 2020.
- Wang et al. (2023) Zhendong Wang, Jianmin Bao, Wengang Zhou, Weilun Wang, Hezhen Hu, Hong Chen, and Houqiang Li. Dire for diffusion-generated image detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 22445–22455, 2023.
- Wang et al. (2022) Zijie J Wang, Evan Montoya, David Munechika, Haoyang Yang, Benjamin Hoover, and Duen Horng Chau. Diffusiondb: A large-scale prompt gallery dataset for text-to-image generative models. arXiv preprint arXiv:2210.14896, 2022.
- Xu et al. (2023a) Danni Xu, Shaojing Fan, and Mohan Kankanhalli. Combating misinformation in the era of generative ai models. In Proceedings of the 31st ACM International Conference on Multimedia, pp. 9291–9298, 2023a.
- Xu et al. (2023b) Qiang Xu, Hao Wang, Laijin Meng, Zhongjie Mi, Jianye Yuan, and Hong Yan. Exposing fake images generated by text-to-image diffusion models. Pattern Recognition Letters, 176:76–82, 2023b.
- Zhang et al. (2019) Xu Zhang, Svebor Karaman, and Shih-Fu Chang. Detecting and simulating artifacts in gan fake images. In 2019 IEEE international workshop on information forensics and security (WIFS), pp. 1–6. IEEE, 2019.
- Zhong et al. (2023) Nan Zhong, Yiran Xu, Zhenxing Qian, and Xinpeng Zhang. Rich and poor texture contrast: A simple yet effective approach for ai-generated image detection. arXiv preprint arXiv:2311.12397, 2023.
- Zhou et al. (2018) Peng Zhou, Xintong Han, Vlad I Morariu, and Larry S Davis. Learning rich features for image manipulation detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1053–1061, 2018.
- Zhu et al. (2017) Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision, pp. 2223–2232, 2017.
- Zhu et al. (2023) Mingjian Zhu, Hanting Chen, Mouxiao Huang, Wei Li, Hailin Hu, Jie Hu, and Yunhe Wang. Gendet: Towards good generalizations for ai-generated image detection. arXiv preprint arXiv:2312.08880, 2023.
- Zhu et al. (2024) Mingjian Zhu, Hanting Chen, Qiangyu Yan, Xudong Huang, Guanyu Lin, Wei Li, Zhijun Tu, Hailin Hu, Jie Hu, and Yunhe Wang. Genimage: A million-scale benchmark for detecting ai-generated image. Advances in Neural Information Processing Systems, 36, 2024.