Assessing Open-world Forgetting in
Generative Image Model Customization
Abstract
Recent advances in diffusion models have significantly enhanced image generation capabilities. However, customizing these models with new classes often leads to unintended consequences that compromise their reliability. We introduce the concept of open-world forgetting to emphasize the vast scope of these unintended alterations, contrasting it with the well-studied closed-world forgetting, which is measurable by evaluating performance on a limited set of classes or skills. Our research presents the first comprehensive investigation into open-world forgetting in diffusion models, focusing on semantic and appearance drift of representations. We utilize zero-shot classification to analyze semantic drift, revealing that even minor model adaptations lead to unpredictable shifts affecting areas far beyond newly introduced concepts, with dramatic drops in zero-shot classification of up to 60%. Additionally, we observe significant changes in texture and color of generated content when analyzing appearance drift. To address these issues, we propose a mitigation strategy based on functional regularization, designed to preserve original capabilities while accommodating new concepts. Our study aims to raise awareness of unintended changes due to model customization and advocates for the analysis of open-world forgetting in future research on model customization and finetuning methods. Furthermore, we provide insights for developing more robust adaptation methodologies.

1 Introduction
Recent advancements in image generation have led to the development of remarkably powerful foundational models capable of synthesizing highly realistic and diverse visual content. Techniques such as Generative Adversarial Networks (GANs) (Goodfellow et al., 2014), and more recently autoregressive models (Yu et al., 2022), Rectified Flows (Liu et al., 2023), and Denoising Diffusion Probabilistic Models (DDPMs) (Ho et al., 2020), have each contributed to significant progress in the field. These methods offer unique strengths in sample quality, diversity, and controllability. Among them, diffusion models have gained particular prominence due to their recent successes and growing influence, especially in enabling text-based image generation (Shonenkov et al., 2023; Ramesh et al., 2022) and complementary multimodal conditioning (Zhang & Agrawala, 2023; Mou et al., 2023), making them a key focus in current research and applications.
Given the high-quality image generation capabilities of these models, a major focus of research has been on how to efficiently incorporate new content and adapt them to new tasks and domains. To address this challenge, various state-of-the-art transfer learning methods have been introduced. These include finetuning approaches such as DreamBooth (Ruiz et al., 2023b) and CustomDiffusion (Kumari et al., 2023), which allow models to learn new concepts effectively. Conditioning-based methods like ControlNet (Zhang & Agrawala, 2023) and IP-Adapters (Ye et al., 2023) enable precise control over generated images by incorporating additional guidance signals. Prompt methods like Prompt-to-Prompt (Hertz et al., 2023) and Textual Inversion (Gal et al., 2023) enable semantic image editing and learning new concepts without modifying the base model. Parameter-efficient techniques such as Low-Rank Adaptations (LoRA) (Hu et al., 2022) have shown great promise, allowing for rapid adaptation with minimal computational overhead (Blattmann et al., 2023; Shi et al., 2023b). These techniques enable models to learn new concepts effectively, even with few examples.
These methods for adapting diffusion models (Ruiz et al., 2023b; Kumari et al., 2023) mainly rely on transfer learning and primarily focus on finetuning model weights to accommodate newly introduced data. However, they lead to unforeseen changes in the model’s behavior, which can have significant implications, altering the model’s existing knowledge, and skills, or the alignment between language and visual content within the network. The field of continual learning has long studied the issue of catastrophic forgetting in neural networks when these aim to adapt to new data (often referred to as new tasks) (Kirkpatrick et al., 2017; De Lange et al., 2021). Traditionally, it has focused on what we term closed-world forgetting, where evaluation is limited to a fixed set of classes encountered in previously learned tasks or skills (similar to concept editing). In contrast, modern foundation models introduce a new challenge: measuring open-world forgetting. Here, the model’s prior knowledge is vast, making it difficult to assess what has been forgotten or altered during the adaptation process.
In this paper, we focus on two popular personalization methods, namely Dreambooth (Ruiz et al., 2023b) and CustomDiffusion (Kumari et al., 2023), for a case study of open-world forgetting. These techniques are especially relevant, as they only add very little new knowledge to the network: a single new concept represented by a small set of typically 3-5 images. Although one might expect that finetuning the model with such limited data would have minimal impact on the vast knowledge of the foundation model (e.g., Stable Diffusion), our analysis reveals that even these small updates can lead to highly detrimental consequences. As Figure 1 illustrates, finetuning can drastically alter the image representation of concepts seemingly unrelated to the training images. The complexity of the forgetting underscores the need for a better understanding of how and where it occurs. Without this understanding, finetuned models risk becoming less reliable, less robust, and ultimately less trustworthy, particularly in safety-critical applications where precision and predictability are paramount.
We propose to analyze open-world forgetting from several perspectives. First, we examine semantic drift using the recent observation that diffusion models can function as zero-shot classifiers; we propose to compare zero-shot capacity of models before and after adaptation on a set of image classification data sets. Second, we analyze appearance drift by evaluating changes in color and perceptual measurements before and after adaptation. Lastly, we assess the extent of forgetting in closely related concepts (local drift) versus unrelated concepts. To address these three aspects of drift, we explore a straightforward, yet effective mitigation strategy by introducing a regularization technique during the training of new concepts. In conclusion, the main contributions of this work are:
-
•
We are the first to systematically analyze open-world forgetting in diffusion models due to model adaptation. Results show that even when adapting to very small domains, the consequences can be highly detrimental.
-
•
We propose two approaches to analyze open-world forgetting, which are designed to assess semantic and appearance drift caused by the adaptation. We leverage the zero-shot classification capabilities of diffusion models to measure the semantic drift, and observe drastic performance drops (of over 60% for some classes). Appearance drift analysis confirms that customization leads to considerable changes in intra-class representation, color, and texture.
-
•
We introduce a method to mitigate open-world forgetting, addressing the challenges of observed drift in text-to-image (T2I) models. This method aims to preserve the original model’s capabilities while allowing for effective customization. Experiments confirm that it greatly reduces both the semantic and appearance drift caused by open-world forgetting.
2 Related work
Text-to-image diffusion model adaptation.
Text-to-image (T2I) diffusion model adaptation is also referred to T2I personalization or subject-driven image generation. This aims to adapt a given model to a new concept by providing a few images and binding the new concept to a unique token. As a result, the adaptation model can generate various renditions of the new concept guided by text prompts. Depending on whether the adaptation method is finetuning the T2I model, they are categorized into two main streams. One of the most representative methods focuses on learning new concept tokens while freezing the T2I generative backbones. Textual Inversion (TI) (Gal et al., 2023) is a pioneering work focusing on finding new pseudo-words by performing personalization in the text embedding space. The following works (Dong et al., 2022; Daras & Dimakis, 2022; Voynov et al., 2023; Han et al., 2023a) continue to improve this technique stream. Another stream is finetuning the T2I generative models while updating the modifier tokens. One of the most representative methods is DreamBooth (Ruiz et al., 2023a), where the pre-trained T2I model learns to bind a modified unique identifier to a specific subject given 35 images, while it also updates the T2I model parameters. Custom Diffusion (Kumari et al., 2023) and other approaches (Han et al., 2023b; Chen et al., 2023b; Shi et al., 2023a) follow this pipeline and further improve the quality of the generation.
Finetuning methods often achieve state-of-the-art performance but introduce forgetting in large T2I models. While research focuses on improving new concept generation, it overlooks continuous model updating and forgetting mitigation. Recent works (Sun et al., 2024; Smith et al., 2023) address token forgetting but neglect other impacts of finetuning, such as semantic drifting in color, appearance, and visual recognition, which this paper explores.
Assessing forgetting.
The main challenge of continual learning is to learn incrementally and accumulate knowledge of new data while preventing catastrophic forgetting, which is defined as a sudden drop in performance on previously acquired knowledge (McCloskey & Cohen, 1989; McClelland et al., 1995). The vast majority of studies on continual learning focus on, what we here call, closed-world forgetting, where the knowledge of the network can be represented by its performance on a limited set of classes (Lopez-Paz & Ranzato, 2017; De Lange et al., 2021; Masana et al., 2022). However, as argued in the introduction, the growing importance of starting from large pretrained models (also known as foundation models), which have a vast prior knowledge, requires new techniques to assess forgetting. The forgetting of large language models (LLMs) during continual finetuning has received some attention in recent years, showing the importance of pretraining to mitigate forgetting (Cossu et al., 2024), however, they mainly evaluate on down-stream-task performance Scialom et al. (2022). To the best of our knowledge, open-world forgetting has not yet been systematically analyzed for text-based image generation models which multi-modal nature can further worsen the impact of forgetting due to misalignment of the modalities.
3 Customization of Diffusion Models
In this section, we briefly introduce text-to-image (T2I) models and the two main customization methods we will evaluate during our analysis. In addition, we introduce an alternative regularization method to further mitigate forgetting.
3.1 Diffusion Models
Diffusion models are a class of generative models that generate data by gradually denoising a sample from a pure noise distribution. The process is modeled in two stages: a forward process and a reverse process. In the forward process, Gaussian noise is iteratively added to data samples, typically over steps, forming a Markov chain. At each step, the transition is defined as:
(1) |
where controls the noise schedule.
The reverse process denoises the data by learning the conditional probability , typically parameterized by a neural network that predicts the added noise at each step. The model is trained by minimizing a simple mean-squared error between the true and predicted noise:
(2) |
Text-to-image diffusion models employ an additional conditioning vector generated using a text encoder and a text prompt . These models have gained prominence for their ability to generate high-quality, diverse samples, often outperforming other generative models like GANs and VAEs in terms of mode coverage and sample quality (Ho et al., 2020; Song et al., 2021).
3.2 Customization approaches
Diffusion models often require finetuning for specific domains or user needs. This involves introducing new conditioning mechanisms or retraining on specialized datasets. This paper applies two adaptation methods to evaluate finetuning’s impact on image generation models.
Dreambooth (Ruiz et al., 2023b) enables personalization of diffusion models by finetuning them with a small set of images. It reuses an infrequent token of the vocabulary to represent a unique subject, allowing the model to generate images of the subject in varied contexts or styles. This approach induces language drift and reduced output diversity in the model, which is mitigated by replaying class-specific instances alongside the subject training, called prior preservation loss. The final training objective reads
(3) |
where is a weighting parameter, and and come from the prior dataset. DreamBooth is especially useful for personalized content generation where subject fidelity is critical.
Custom diffusion (Kumari et al., 2023) is another approach aimed at efficiently finetuning diffusion models with minimal data and compute. This method observes that the cross-attention layer parameters undergo the most change during personalization, so they propose to only update the key and value projections in these layers. It introduces a token into the text-encoder representing a unique subject, rather than reusing an old one. By freezing the majority of the model’s parameters and focusing updates on a few key layers, Custom Diffusion facilitates rapid customization with less degradation in image quality. Prior preservation loss is maintained, since language drift is still experienced otherwise.
Customized Model Set In our experiments, we will evaluate Dreambooth and Custom Diffusion. We adapt both these models to ten different concepts based on 5 images per concept. The concepts are ‘lamp’, ‘vase’, ‘person2’,‘person3’,‘cat’,‘dog’,‘lighthouse’,‘waterfall’,‘bike’ and ‘car’ taken from CustomConcept101 (Kumari et al., 2023). We will refer to these ten models for both DreamBooth and Custom Diffusion as the Customized Model Set.
3.3 Drift Correction
The two studied approaches, Dreambooth (Ruiz et al., 2023b) and Custom Diffusion (Kumari et al., 2023), apply finetuning to adapt to the new data: they mainly focus on how good the learned model is on the target data, and do not study the possible detrimental effects for other classes. The Dreambooth method includes a method called prior regularization, which by replaying general instances of the concept being learned (see Eq. 3), helps to prevent the model from overfitting to the new data and ensures that the representation of the superclass remains stable. This same mitigation strategy is also applied in custom diffusion (Kumari et al., 2023).
In this paper, we propose another regularization technique that can be applied during new concept learning. The method is remarkably simple and is motivated from continual learning literature. This field has proposed a variety of methods to counter forgetting during the learning of new concepts (De Lange et al., 2021). Regularization methods aim to regularize the learning of new concepts in such a way that it does not change weights which were found relevant for previous tasks. The field differentiates between parameter regularization methods, like EWC (Kirkpatrick et al., 2017) which directly learn an importance weight for all the network parameters, or functional (or data) regularization, like Learning-without-Forgetting (Li & Hoiem, 2017; Pan et al., 2020) which regularizes the weights indirectly by imposing a penalty on changes between the (intermediate) outputs of a previous and current model.
We propose to apply a functional regularization loss to the network during the training of new concepts. Our loss, called drift correction loss, constrains the difference between the outputs of the pre-trained and fine-tuned models when the new concept is not present in the prompt. It has the following form:
(4) |
where the second term is the distillation loss, is a relative weighting parameter and is the base model. This loss helps to maintain consistency in the model’s internal representations while allowing it to learn new information effectively. For the training process, we choose instances from the same class as the concept being learned, similar to those used by prior regularization. The change between our proposed drift correction method (Eq. 4) and the existing prior regularization (Eq. 3) is that we do not require the finetuned network to estimate the true forward noise, but instead we want it to estimate the same noise as the original starting network. We will see that this small change significantly improves stability and mitigates forgetting.
In our evaluation, we provide results for DreamBooth (DB) which includes the prior regularization, for Dreambooth with Drift Correction (DB-DC) which also includes the prior regularization and for DreamBooth with Drift Correction without the prior regularization (DB-DCpr). Similarly, we show results for the various variants of Custom Diffusion (CD, CD-DC, and CD-DCpr).
4 Open-World Forgetting in Generative Model Adaptation
In this section, we explore the effects of finetuning on foundational image generation models, particularly how even slight modifications can significantly impair the model’s ability to retain previously acquired knowledge. We hypothesize that this degradation affects not only the model’s performance on newly introduced tasks, but also its capacity to accurately reproduce or classify previously learned concepts. Given the broad scope of knowledge encompassed by the pretrained model, we refer to this phenomenon as open-world forgetting.
As an initial experiment, to assess open-world forgetting, we evaluate both the original unaltered model (called base model from now on) and the Customized Model Set on 10,000 user prompts from DiffusionDB (Wang et al., 2023) dataset (prompt examples are provided in Appendix B.1). Specifically, we measure the change of the resulting images using the cosine distance between CLIP-I encodings (Radford et al., 2021) when generating images with the same prompt and seed. Distances in the CLIP-I embedding are related to semantic similarity between images, with smaller distances indicating more similar visual content and larger distances suggesting more significant differences in the generated images. The distribution is plotted in Figure 2. For a detailed description of our experimental setup, please refer to Appendix B.1. It is important to note that a personalization method that does not alter the model would yield identical image outputs, resulting in a plot density concentrated at .
When considering Figure 2a, we observe that even though most of the prompts from DiffusionDB are not related with the selected trained concepts, there is a significant part of the distribution that is shifted to the left. This shows indeed that the representations of the original model have changed. Furthermore, further analysis shows that open-world forgetting significantly alters the output in different ways, as illustrated by the samples in Figure 2a. For instance, a sampled pair from the most dissimilar outputs (purple triangle) shows a complete change in content, colors, and scene composition that no longer matches the prompt. In contrast, a very similar pair (yellow star) closely adheres to the original model’s output, with only changes in color or details. Interestingly, when looking at Figure 2b where we apply the proposed Drift Correction to DreamBooth, the distribution shifts to the right, showing that the drift has been reduced considerably.

To better assess the impact of open-world forgetting, we propose to categorize the effects into two distinct types: semantic and appearance drift. Semantic drift implies a change at the class or object level, where one concept is effectively misencoded as another. Appearance drift, on the other hand, refers to shifts in the appearance of a concept that do not necessarily imply a change in recognition (e.g., alterations in color, texture, or scene composition). It is important to note that these two categories are highly correlated, and changes in either of them impact the other.
4.1 Semantic Drift
Semantic drift refers to alterations in a model’s representation that cause the generation of semantically divergent content following customization. In the experiment depicted in Figure 2a, almost all prompts exhibit some level of drift, with a notable long tail of highly dissimilar generations. Many of these pronounced deviations have resulted in the generation of content that semantically no longer align with the input prompt.
To evaluate how semantic drift affects generative models, we use a straightforward approach: we utilize the model’s internal representations on different classification tasks (Mittal et al., 2023; Tang et al., 2023). It is based on a recent insight that showed that diffusion models can be directly applied for zero-class classification, by leveraging the conditional likelihood estimation of the model. Concretely, we use Diffusion Classifier (Li et al., 2023), where a posterior distribution over classes is calculated as:
(5) |
Monte Carlo sampling is performed over and to obtain a classifier from a text-to-image model .
While this method offers a simple, parameter-free approach to evaluating semantic drift, it is worth noting that alternative techniques have been proposed to assess the representation space of diffusion models. These include linear probing on activations (Xiang et al., 2023), analysis of hierarchical features (Mukhopadhyay et al., 2023), and methods requiring a preliminary likelihood maximization stage (Chen et al., 2023a). However, these alternatives often involve additional computational steps or are subject to specific settings, potentially limiting their applicability or introducing complexity to the evaluation process.
We conducted zero-shot classification experiments across multiple datasets spanning diverse domains to quantify the semantic drift of the models. We perform two measurements. First, we measure the average zero-shot classification score for the various models (the results are averaged over the 10 models of the Customized Model Set). Second, we establish the performance of the original pretrained model as the baseline, and measure the presence of semantic drift by calculating the drop in accuracy from the baseline. We also report the worst class drop which is the drop in accuracy of the class that has suffered the largest deterioration due to the adaptation. For further details on the classification method, please refer to Appendix B.2.
CIFAR10 | STL10 | Flowers | Pets | ObjectNet | Food | Aircraft | |
---|---|---|---|---|---|---|---|
Base Model | 81.60 | 93.00 | 50.00 | 86.87 | 28.50 | 71.09 | 23.40 |
\hdashlineDB | 75.92 (32.40) | 91.30 (18.60) | 46.61 (64.00) | 82.61 (36.43) | 25.26 (56.00) | 65.48 (56.00) | 19.36 (58.00) |
DB-DC | 80.98 (14.00) | 93.36 (4.40) | 49.29 (42.00) | 86.64 (17.14) | 27.72 (42.00) | 69.07 (44.00) | 21.42 (48.00) |
DB-DCpr | 80.60 (14.00) | 92.94 (5.20) | 49.06 (40.00) | 86.37 (16.43) | 27.45 (46.00) | 68.79 (44.00) | 21.54 (44.00) |
\hdashlineCD | 79.98 (17.00) | 91.40 (12.20) | 47.65 (66.00) | 83.46 (33.57) | 25.75 (58.00) | 65.25 (56.00) | 19.44 (58.00) |
CD-DC | 82.36 (9.00) | 93.02 (5.00) | 49.33 (42.00) | 86.37 (16.43) | 27.91 (42.00) | 69.19 (44.00) | 21.94 (46.00) |
CD-DCpr | 82.04 (10.80) | 92.76 (6.00) | 49.16 (44.00) | 86.70 (20.00) | 27.77 (42.00) | 68.99 (44.00) | 21.56 (48.00) |
DINO | CLIP-I | CLIP-T | |
---|---|---|---|
DB | 0.42 | 0.68 | 0.79 |
DB-DC | 0.43 | 0.68 | 0.78 |
DB-DCpr | 0.43 | 0.68 | 0.78 |
\hdashlineCD | 0.44 | 0.69 | 0.79 |
CD-DC | 0.44 | 0.69 | 0.79 |
CD-DCpr | 0.44 | 0.69 | 0.79 |
The results in Table 1 are surprising, average zero-shot classification accuracy drop significantly on all the datasets: adapting a huge generative image foundation model to just five images of a new concept has a vast impact throughout the latent space of the diffusion models. When applying DreamBooth, average zero-shot performance drops by over 4% on CIFAR10, Pets, Food and Aircraft. If we look at individual classes, the impact can be much larger. As indicated by the worst class drop, for some classes, zero-shot performance drops by over 60% (e.g. ‘vacuum cleaner’ gets recognized as ‘microwave’, ‘drill’ or ‘laptop’). We show that these drops in performance are mitigated to a large extent by our alternative Drift Correction results (see DB-DC and CD-DC results) and their average zero-shot classification scores are in general within 1% of the base model. Removing the prior regularization from our method (see DB-DCpr and CD-DCpr) leads to only slightly lower results, showing the impact of our proposed regularization method. Also, worst class drop significantly reduces when applying DC, but for some datasets remains still high.
Furthermore, to measure generation quality, diversity, and alignment between image and text of the learned concepts, we employ metrics adapted from personalization methods: DINO, CLIP-I, and CLIP-T. In Table 2 we can see that the proposed regularization method DC does not negatively impact the image generation quality of the learned concepts.111The results with standard deviations for Table 1 and 2 are provided in Appendix B.
4.2 Appearance Drift
While open-world forgetting does not always result in significant changes to the core content of the image, as shown in Figure 2a, it notably affects intra-class variation, color distribution, and texture characteristics. We define these collective changes as appearance drift, a phenomenon that alters the model’s representation space in subtle yet impactful ways. Figure 3 demonstrates two key aspects of appearance drift; intra-class and contextual variation (first row), where different customizations of the base model lead to changes in car brand and background, while maintaining the overall concept of ‘car’. Color shift (second row), where the color palette of the generated images changes significantly, even when the intra-class characteristics and background remain relatively constant.
Appearance drift manifests through alterations in visual attributes at varying degrees of intensity. Finetuning can cause the model to reinterpret these visual features, leading to inconsistencies between original and newly generated outputs. Although initially subtle, appearance drift can substantially impact customized models. For example, when attempting to learn and generate a set of new concepts within the same context (e.g., for synthetic dataset creation or advertising purposes), each customization of the base model may result in color and content changes. This variability makes it challenging to achieve consistent results across multiple iterations. Moreover, as the customization process alters the model’s manifold, the resulting model becomes less reliable in domains outside the scope of the customization training images. This limitation highlights the importance of understanding and mitigating appearance drift in applications of fine-tuned text-to-image models.

How to measure appearance drift?
Quantifying appearance drift presents unique challenges due to the inherent variability in text-to-image model outputs. Traditional metrics like LPIPS (Zhang et al., 2018) and DIST (Ding et al., 2020) are designed for image pair comparisons. However, the inherent variability in T2I model outputs means that images generated from the same prompt can vary significantly due to changes in seed, model weights, and prompt interpretation. Comparing just two images fails to capture the full range of possible outputs and does not adequately represent the model’s capabilities or biases. Consequently, conclusions drawn from such limited comparisons may lack statistical significance.
To address this variability, the research community has employed metrics that measure distances between probability distributions (e.g., FID (Heusel et al., 2017), KID (Bińkowski et al., 2018)) of real-world observations and generated data222In general FID and KID require a set of real images for comparison. In our study, we consider the images generated by the original model as the “real” set, as we are measuring the shift from this initial distribution.. In addition to these metrics, we also propose a new metric that directly measures the color drift between image sets.
Color Drift Index.
With the aim to introduce a clearly interpretable difference measure for the color content of the generated images, we propose a novel approach that measures specific properties in pixel space. Our method focuses on color assessment, as traditional image generation metrics tend to be more sensitive to intra-class and texture variations. Inspired by natural image statistics, we introduce the Color Drift Index (CDI), which estimates the color distribution associated with a specific concept by analyzing a large number of generated images.
We utilize the CIE chromaticity diagram333Our approach offers the added benefit of being applicable across multiple color spaces., where each pixel color is projected onto a lobe-shaped space representing all visible colors. Given a set of images and their density distribution in the CIE chromaticity diagram , we calculate the CDI as the Wasserstein distance (Panaretos & Zemel, 2019) between the color distributions of two sets of images:
(6) |

We evaluate the appearance drift using the CDI together with KID (FID results are presented in Appendix B.3). We conducted a comprehensive experiment adopting the carefully curated selection of common concepts from the dataset of Torralba & Oliva (2003). For each prompt, we generated 1,000 images using both original and the ten models from the Customized Model Set. Figure 4 presents the mean values of the metrics across several adaptations, providing a visual representation of the differences captured by each measure. For a detailed overview of the results, including individual model performances, refer to Table 8 in the appendix. To help interpret the KID and CDI scores, we provide a control setting (blue line). In this configuration, we measure CDI and KID between images generated with the base model but using different seeds (functioning as a lower bound). If we are sampling from the same distribution, the base model should yield lower distances (approaching zero as the number of samples grows) than the customized models (DB and CD).
The results in Figure 4 reveal two key insights. First, the base model consistently produces smaller distance values compared to DB, confirming that the distribution of each concept is indeed changing due to appearance drift. Second, each concept is affected differently by the drift, attributable to the fact that each concept relates to different parts of the model’s manifold. Furthermore, as demonstrated in Appendix C, the magnitude and nature of the drift vary as a function of the content in the replay buffer and training images. Also, importantly, Figure 4 shows that our proposed method (DC) significantly reduces the impact of the appearance drift introduced by customization methods. Especially, the drift measure in KID is considerably reduced.
These findings underscore the complexity and pervasiveness of appearance drift in fine-tuned text-to-image models. They highlight the need for robust mitigation strategies and careful consideration when deploying customized models in real-world applications, emphasizing the importance of ongoing research in this area to ensure the reliability and consistency of generated outputs.
4.3 Local Drift
In this paper, we have focused on drift throughout the whole diffusion model manifold. Previous works, especially those in the machine unlearning community (Gandikota et al., 2023), have concentrated on local drift. When removing a concept from a model, it is believed to mainly impact the representation of closely related concepts (hence the name local drift). Our findings suggest that the effects of finetuning are more pervasive than previously thought, potentially influencing the model’s understanding and representation of far-away categories as well as close-by (local) categories.

Here, we repeated our experiments from Section 4.1 and 4.2 to measure the local semantic and local appearance drift. For this setup, we generated 1,000 samples of the closest concepts (superclasses) to each trained model (see Appendix B.1 for the details) and evaluated the CLIP-I, CDI, FID, and KID metrics. As Figure 5a shows, the semantic drift is showing a significant shift towards the left, indicating that local drift is more pronounced. For appearance drift, Figure 5c depicts a more uniform color and KID shift over all the models; this shows that related concepts are affected with a similar magnitude by appearance drift. Again the application of our proposed Drift Correction method greatly reduces both the local semantic drift (as measured in Figure 5b and it almost removes the local appearance drift as measured by KID, even though some color drift remains (see Figure 5c).
5 Discussion and Conclusion
Our investigation into unintended consequences of generative model adaptation reveals several key contributions: 1) We observe that finetuning foundational generative models leads to substantial open-world forgetting, manifesting as both semantic and appearance drift. 2) We propose to measure semantic drift by considering zero-shot classification on a wide range of image classification tasks, and we propose a novel metric to measure color drift that combined with KID measures appearance drift. 3) We propose a technique to mitigate open-world forgetting using functional regularization. This method shows promise in preserving foundational knowledge while allowing effective customization. These findings provide a guidance and initial step to tackle open-world forgetting and developing more robust and reliable customization techniques.
The last couple of years has seen a growing number of foundation models being released. These models have been enthusiastically adapted by the community, and are often the basis for further finetuning operations to new domains, tasks, etc. The main purpose of our paper is to contribute to the awareness of the unintended changes in the original foundation model knowledge that can occur when adapting these. This phenomenon, coined open-world forgetting, is complicated to assess, and we have proposed several ways to assess it for customized generative models. We hope that our results motivate more research in this direction. Open-world forgetting resembles the finding of a needle in a haystack due to the vast amount of prior knowledge of the foundation models, and we think that other research directions which actively optimize to find the most affected areas might be required. Finally, we consider extending our methodology to other forms of model adaptation, such as unlearning techniques as exiting future work.
Limitations. While our study provides valuable insights into open-world forgetting in generative model adaptation, there are several limitations to consider. First, inherently to the nature of open-world, we cannot exhaustively evaluate the entire conceptual space of the model. Our analysis is limited to a subset of concepts and may not capture all potential instances of forgetting. Second, our study has mainly focused on evaluating the impact of diffusion model customization. Even though we represent a mitigation technique for open-world forgetting (Drift Correction) we think that there are many more ways to address this problem (especially considering the vast continual learning literature) which have not considered in our study.
Ethical Statement
We acknowledge the potential ethical implications of deploying generative models, including issues related to privacy, data misuse, and the propagation of biases. All models used in this paper are publicly available, as well as the base training scripts. We will release the modified codes to reproduce the results of this paper. We also want to point out the potential role of customization approaches in the generation of fake news, and we encourage and support responsible usage of these models. Finally, we think that awareness of open-world forgetting can contribute to safer models in the future, since it encourages a more thorough investigation into the unpredictable changes occurring when adapting models to new data.
Reproducibility Statement
To facilitate reproducibility, we will make the entire source code and scripts needed to replicate all results presented in this paper available after the peer review period. We will release the code for the novel color metric we have introduced. We conducted all experiments using publicly accessible datasets. Elaborate details of all experiments have been provided in the Appendices.
References
- Bińkowski et al. (2018) M Bińkowski, DJ Sutherland, M Arbel, and A Gretton. Demystifying mmd gans. In ICLR, 2018.
- Blattmann et al. (2023) Andreas Blattmann, Tim Dockhorn, Sumith Kulal, Daniel Mendelevitch, Maciej Kilian, Dominik Lorenz, Yam Levi, Zion English, Vikram Voleti, Adam Letts, et al. Stable video diffusion: Scaling latent video diffusion models to large datasets. arXiv preprint arXiv:2311.15127, 2023.
- Caron et al. (2021) Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9650–9660, 2021.
- Chen et al. (2023a) Huanran Chen, Yinpeng Dong, Zhengyi Wang, Xiao Yang, Chengqi Duan, Hang Su, and Jun Zhu. Robust classification via a single diffusion model. arXiv preprint arXiv:2305.15241, 2023a.
- Chen et al. (2023b) Wenhu Chen, Hexiang Hu, Yandong Li, Nataniel Rui, Xuhui Jia, Ming-Wei Chang, and William W Cohen. Subject-driven text-to-image generation via apprenticeship learning. arXiv preprint arXiv:2304.00186, 2023b.
- Cossu et al. (2024) Andrea Cossu, Antonio Carta, Lucia Passaro, Vincenzo Lomonaco, Tinne Tuytelaars, and Davide Bacciu. Continual pre-training mitigates forgetting in language and vision. Neural Networks, 179:106492, 2024.
- Daras & Dimakis (2022) Giannis Daras and Alex Dimakis. Multiresolution textual inversion. In NeurIPS 2022 Workshop on Score-Based Methods, 2022.
- De Lange et al. (2021) Matthias De Lange, Rahaf Aljundi, Marc Masana, Sarah Parisot, Xu Jia, Aleš Leonardis, Gregory Slabaugh, and Tinne Tuytelaars. A continual learning survey: Defying forgetting in classification tasks. IEEE transactions on pattern analysis and machine intelligence, 44(7):3366–3385, 2021.
- Ding et al. (2020) Keyan Ding, Kede Ma, Shiqi Wang, and Eero P Simoncelli. Image quality assessment: Unifying structure and texture similarity. IEEE transactions on pattern analysis and machine intelligence, 44(5):2567–2581, 2020.
- Dong et al. (2022) Ziyi Dong, Pengxu Wei, and Liang Lin. Dreamartist: Towards controllable one-shot text-to-image generation via contrastive prompt-tuning. arXiv preprint arXiv:2211.11337, 2022.
- Gal et al. (2023) Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H Bermano, Gal Chechik, and Daniel Cohen-Or. An image is worth one word: Personalizing text-to-image generation using textual inversion. ICLR, 2023.
- Gandikota et al. (2023) Rohit Gandikota, Joanna Materzynska, Jaden Fiotto-Kaufman, and David Bau. Erasing concepts from diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 2426–2436, 2023.
- Goodfellow et al. (2014) Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, and Yoshua Bengio. Generative adversarial nets. In Neural Information Processing Systems, 2014. URL https://api.semanticscholar.org/CorpusID:261560300.
- Han et al. (2023a) Inhwa Han, Serin Yang, Taesung Kwon, and Jong Chul Ye. Highly personalized text embedding for image manipulation by stable diffusion. arXiv preprint arXiv:2303.08767, 2023a.
- Han et al. (2023b) Ligong Han, Yinxiao Li, Han Zhang, Peyman Milanfar, Dimitris Metaxas, and Feng Yang. Svdiff: Compact parameter space for diffusion fine-tuning. ICCV, 2023b.
- Hertz et al. (2023) Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. Prompt-to-prompt image editing with cross attention control. ICLR, 2023.
- Heusel et al. (2017) Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In NeurIPS, pp. 6626–6637, 2017.
- Ho et al. (2020) Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 6840–6851. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/4c5bcfec8584af0d967f1ab10179ca4b-Paper.pdf.
- Hu et al. (2022) Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=nZeVKeeFYf9.
- Kirkpatrick et al. (2017) James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, et al. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13):3521–3526, 2017.
- Kumari et al. (2023) Nupur Kumari, Bingliang Zhang, Richard Zhang, Eli Shechtman, and Jun-Yan Zhu. Multi-concept customization of text-to-image diffusion. CVPR, 2023.
- Li et al. (2023) Alexander C. Li, Mihir Prabhudesai, Shivam Duggal, Ellis Brown, and Deepak Pathak. Your diffusion model is secretly a zero-shot classifier. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 2206–2217, October 2023.
- Li & Hoiem (2017) Zhizhong Li and Derek Hoiem. Learning without forgetting. IEEE transactions on pattern analysis and machine intelligence, 40(12):2935–2947, 2017.
- Liu et al. (2023) Xingchao Liu, Chengyue Gong, and qiang liu. Flow straight and fast: Learning to generate and transfer data with rectified flow. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=XVjTT1nw5z.
- Lopez-Paz & Ranzato (2017) David Lopez-Paz and Marc’Aurelio Ranzato. Gradient episodic memory for continual learning. Advances in neural information processing systems, 30, 2017.
- Masana et al. (2022) Marc Masana, Xialei Liu, Bartłomiej Twardowski, Mikel Menta, Andrew D Bagdanov, and Joost Van De Weijer. Class-incremental learning: survey and performance evaluation on image classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(5):5513–5533, 2022.
- McClelland et al. (1995) James L McClelland, Bruce L McNaughton, and Randall C O’Reilly. Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. Psychological review, 102(3):419, 1995.
- McCloskey & Cohen (1989) Michael McCloskey and Neal J Cohen. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of learning and motivation, volume 24, pp. 109–165. Elsevier, 1989. URL https://www.sciencedirect.com/science/article/abs/pii/S0079742108605368.
- Mittal et al. (2023) Sarthak Mittal, Korbinian Abstreiter, Stefan Bauer, Bernhard Schölkopf, and Arash Mehrjou. Diffusion based representation learning. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp. 24963–24982. PMLR, 23–29 Jul 2023. URL https://proceedings.mlr.press/v202/mittal23a.html.
- Mou et al. (2023) Chong Mou, Xintao Wang, Liangbin Xie, Jian Zhang, Zhongang Qi, Ying Shan, and Xiaohu Qie. T2i-adapter: Learning adapters to dig out more controllable ability for text-to-image diffusion models. arXiv preprint arXiv:2302.08453, 2023.
- Mukhopadhyay et al. (2023) Soumik Mukhopadhyay, Matthew Gwilliam, Yosuke Yamaguchi, Vatsal Agarwal, Namitha Padmanabhan, Archana Swaminathan, Tianyi Zhou, and Abhinav Shrivastava. Do text-free diffusion models learn discriminative visual representations?, 2023.
- Pan et al. (2020) Pingbo Pan, Siddharth Swaroop, Alexander Immer, Runa Eschenhagen, Richard Turner, and Mohammad Emtiyaz E Khan. Continual deep learning by functional regularisation of memorable past. Advances in neural information processing systems, 33:4453–4464, 2020.
- Panaretos & Zemel (2019) Victor M Panaretos and Yoav Zemel. Statistical aspects of wasserstein distances. Annual review of statistics and its application, 6(1):405–431, 2019.
- Radford et al. (2021) Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748–8763. PMLR, 2021.
- Ramesh et al. (2022) Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125, 2022.
- Rombach et al. (2022) Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10684–10695, 06 2022.
- Ruiz et al. (2023a) Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman. Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. CVPR, 2023a.
- Ruiz et al. (2023b) Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman. Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023b.
- Scialom et al. (2022) Thomas Scialom, Tuhin Chakrabarty, and Smaranda Muresan. Fine-tuned language models are continual learners. arXiv preprint arXiv:2205.12393, 2022.
- Shi et al. (2023a) Jing Shi, Wei Xiong, Zhe Lin, and Hyun Joon Jung. Instantbooth: Personalized text-to-image generation without test-time finetuning. arXiv preprint arXiv:2304.03411, 2023a.
- Shi et al. (2023b) Ruoxi Shi, Hansheng Chen, Zhuoyang Zhang, Minghua Liu, Chao Xu, Xinyue Wei, Linghao Chen, Chong Zeng, and Hao Su. Zero123++: a single image to consistent multi-view diffusion base model, 2023b.
- Shonenkov et al. (2023) Alex Shonenkov, Misha Konstantinov, Daria Bakshandaeva, Christoph Schuhmann, Ksenia Ivanova, and Nadiia Klokova. Deepfloyd-if. https://github.com/deep-floyd/IF, 2023.
- Smith et al. (2023) James Seale Smith, Yen-Chang Hsu, Lingyu Zhang, Ting Hua, Zsolt Kira, Yilin Shen, and Hongxia Jin. Continual diffusion: Continual customization of text-to-image diffusion with c-lora. arXiv preprint arXiv:2304.06027, 2023.
- Song et al. (2021) Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=PxTIG12RRHS.
- Sun et al. (2024) Gan Sun, Wenqi Liang, Jiahua Dong, Jun Li, Zhengming Ding, and Yang Cong. Create your world: Lifelong text-to-image diffusion. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024.
- Tang et al. (2023) Luming Tang, Menglin Jia, Qianqian Wang, Cheng Perng Phoo, and Bharath Hariharan. Emergent correspondence from image diffusion. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=ypOiXjdfnU.
- Torralba & Oliva (2003) Antonio Torralba and Aude Oliva. Statistics of natural image categories. Network: computation in neural systems, 14(3):391, 2003.
- von Platen et al. (2022) Patrick von Platen, Suraj Patil, Anton Lozhkov, Pedro Cuenca, Nathan Lambert, Kashif Rasul, Mishig Davaadorj, Dhruv Nair, Sayak Paul, William Berman, Yiyi Xu, Steven Liu, and Thomas Wolf. Diffusers: State-of-the-art diffusion models. https://github.com/huggingface/diffusers, 2022.
- Voynov et al. (2023) Andrey Voynov, Qinghao Chu, Daniel Cohen-Or, and Kfir Aberman. : Extended textual conditioning in text-to-image generation. arXiv preprint arXiv:2303.09522, 2023.
- Wang et al. (2023) Zijie J. Wang, Evan Montoya, David Munechika, Haoyang Yang, Benjamin Hoover, and Duen Horng Chau. DiffusionDB: A large-scale prompt gallery dataset for text-to-image generative models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023. URL https://aclanthology.org/2023.acl-long.51.
- Xiang et al. (2023) Weilai Xiang, Hongyu Yang, Di Huang, and Yunhong Wang. Denoising diffusion autoencoders are unified self-supervised learners. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023.
- Ye et al. (2023) Hu Ye, Jun Zhang, Sibo Liu, Xiao Han, and Wei Yang. Ip-adapter: Text compatible image prompt adapter for text-to-image diffusion models. arXiv preprint arXiv:2308.06721, 2023.
- Yu et al. (2022) Jiahui Yu, Yuanzhong Xu, Jing Yu Koh, Thang Luong, Gunjan Baid, Zirui Wang, Vijay Vasudevan, Alexander Ku, Yinfei Yang, Burcu Karagol Ayan, et al. Scaling autoregressive models for content-rich text-to-image generation. arXiv preprint arXiv:2206.10789, 2(3):5, 2022.
- Zhang & Agrawala (2023) Lvmin Zhang and Maneesh Agrawala. Adding conditional control to text-to-image diffusion models. arXiv preprint arXiv:2302.05543, 2023.
- Zhang et al. (2018) Richard Zhang, Phillip Isola, Alexei A Efros, Eli Shechtman, and Oliver Wang. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 586–595, 2018.
Appendix A More examples
A.1 Open-world forgetting
In Figure 6, we present more samples of generated images with appearance drift.

A.2 Local drift
In Figure 7 we present examples of local drift for the concept ”dog”.

Appendix B Experiment details
This section outlines our experimental setup, including datasets, metrics, and training configurations.
B.1 Semantic Drift Evaluation
Datasets.
To evaluate open-world forgetting, we select a random subset of 10,000 user prompts from DiffusionDB 2M (Wang et al., 2023) (Table 3). For adaptation training and evaluation, we choose a subset of 10 concepts from CustomConcept101 (Kumari et al., 2023), namely decoritems_vase2, decoritems_lamp1, person_2, person_3, pet_cat5, pet_dog4, transport_bike, transport_car2, scene_lighthouse, scene_waterfall. Each concept contains approximately 3-5 images. For superclass evaluation, we create a dataset of 10 synonyms with respect to each concept, which can be found in Table 4.
DiffusionDB prompts |
---|
“dafne keen, mad max, cinematic shot, 8k resolution” |
“creepy horror movie characters, fog, rain, volumetric lighting, beautiful, golden hour, sharp focus, highly detailed, cgsociety” |
“the railroad is a place of death. it’s where the forgotten and the damned go to die. it’s a place of dark secrets and hidden terror. photorealistic ” |
“samurai jack johnny bravo by salvador dali ” |
“Film still of Emma Watson as Princess Leia in Star Wars (1977)” |
“a detailed figure of indigo montoya, first 4 figures, detailed product photo ” |
“a hyper scary pokemon, horror, creepy, big budget horror movie, by zdzisław beksinski, by dorian cleavenger ” |
“the war between worlds extremely detailed claymation art, dark, moody, foggy ” |
“a painting of Hatsune Miku by H. R. Giger, highly detailed, 4k digital art” |
“a redneck with wings and horns wearing sunglasses and snake skin smoking a blunt, detailed, 4 k, realistic, picture ” |
“fantasy art 4 k ultra detailed photo caricature walter matthau as an fighter pilot ” |
“ CG Homer Simpson as Thanos, cinematic, 4K” |
“Full body portrait of Raven from Teen Titans (2003), digital art by Sakimichan, trending on ArtStation” |
“bigfoot walking down the street in downtown Bremerton Washington” |
“garden layout rendering with flowers and plants native to ottawa canada ” |
“a beautiful planet of guangzhou travel place of interest, chill time. good view, exciting honor. by david inshaw ” |
“an oil painting of Dwayne Johnson instead of Mona Lisa in the famous painting The Joconde painted by Leonardo Da Vinci” |
“film still of danny devito as mario in live action super mario bros movie, 4 k ” |
“a beautiful artist’s rendition of what the stable diffusion algorithm dreams about ” |
Concept | Synonyms |
---|---|
bike | pedal cycle, velociped, roadster, bicycle, push bike, pushbike, cycle,wheels, two-wheeler, pedal bike |
car | jalopy, ride, auto, vehicle, coupe, wheels, automobile, sedan, hatchback, motocar |
cat | feline, grimalkin, mouser, moggy, tabby, puss, kitty, kitten, pussycat, tomcat |
dog | canine, hound, pup, pooch, fido, puppy, mutt, man’s best friend, doggy, cur |
lamp | fixture, chandelier, light, illuminator, lantern, luminaire, glow, torch, sconce, beacon |
lighthouse | light, coastal beacon, navigation light, pharos, seamark, watchover, beacon, guide light, light station, signal tower |
person | gent, bloke, chap, gentleman, lad, guy, male, bro, fellow, dude |
vase | urn, amphora, container, pitcher, carafe, receptacle, jar, vessel, pot, jug |
waterfall | rapids, torrent, flume, cascade, spillway, cataract, plunge, chute, falls, deluge |
Metrics.
We employ three primary metrics to assess image generation quality. CLIP-I is calculated as the average pairwise cosine similarity between CLIP (Radford et al., 2021) embeddings of real and generated images. DINO uses the same pairwise cosine similarity method but with DINO (Caron et al., 2021) ViT-S16 embeddings. This metric is preferred over CLIP-I as it does not ignore differences between subjects of the same class. CLIP-T measures the CLIP embedding cosine similarity between the prompt and the generated image, and is used to evaluate prompt fidelity.
Training configuration.
We adapt models using publicly available scripts from Diffusers (von Platen et al., 2022) for Dreambooth444https://github.com/huggingface/diffusers/blob/main/examples/dreambooth/train_dreambooth_lora.py and Custom Diffusion555https://github.com/huggingface/diffusers/blob/main/examples/custom_diffusion/train_custom_diffusion.py applied to Stable Diffusion v1.5 (Rombach et al., 2022). Both methods use prior regularization unless otherwise stated, which is designed to prevent drifting towards the training concept. The set of images for prior regularization is generated from the base model before training starts. We use the LoRA script versions and refer to the resulting models as DB for DreamBooth and CD for Custom Diffusion. Full finetuning models exhibit the same or worse shortcomings as the LoRA models analyzed and are termed FT for finetuning, such as DB FT in Table 6. Since DB and CD are similar and to ensure a fair comparison, both methods use similar training settings: a learning rate of 1e-4, batch size of 1, 500 training steps, and no augmentations. The prior regularization uses a weighting of 1 and comprises 200 samples of generated images with the prompt “{concept}” for each concept, using default generation settings. For the drift correction method described in Section 3.3, all settings remain the same, and the weighting parameter is set to .
B.2 Diffusion Classifier
We employ the official released code of the Diffusion Classifier method666https://github.com/diffusion-classifier/diffusion-classifier. However, due to computational constraints, we modify some parameters for our explorations. We reduce the keep list to across all datasets while maintaining the trial list at . This significantly reduces computational time while resulting in minimal percentual score uncertainty. Additionally, datasets with many classes or samples were reduced to have a total number of samples of roughly 500 by random selection of the samples of each class. The datasets configuration can be seen in Table 5. It is worth noting that the original ObjectNet has 313 classes, but Diffusion Classifier only uses 113 for testing. We also use a fixed noise for consistent evaluations.
Dataset | Food | CIFAR10 | Aircraft | Pets | Flowers | STL10 | ObjectNet |
---|---|---|---|---|---|---|---|
# classes | 101 | 10 | 100 | 37 | 102 | 10 | 113 |
Samples / class | 5 | 50 | 5 | 14 | 5 | 50 | 5 |
Total samples | 505 | 500 | 500 | 518 | 510 | 500 | 565 |
Food | CIFAR10 | Aircraft | Pets | Flowers | STL10 | ObjectNet | |
---|---|---|---|---|---|---|---|
Base Model | 71.09 | 81.60 | 23.40 | 86.87 | 50.00 | 93.00 | 28.50 |
DB FT | 61.505.95 | 69.866.79 | 16.043.81 | 79.115.29 | 43.186.58 | 87.044.63 | 21.523.41 |
DB | 65.482.59 | 75.925.81 | 19.362.69 | 82.613.33 | 46.612.49 | 91.302.05 | 25.261.85 |
DB-DC | 69.071.58 | 80.982.57 | 21.420.45 | 86.640.92 | 49.291.33 | 93.360.70 | 27.721.56 |
DB-DCpr | 68.791.78 | 80.602.46 | 21.540.77 | 86.370.78 | 49.061.12 | 92.940.95 | 27.451.15 |
CD | 65.252.76 | 79.984.21 | 19.441.97 | 83.462.97 | 47.652.34 | 91.401.60 | 25.752.16 |
CD-DC | 69.191.73 | 82.361.91 | 21.941.51 | 86.371.28 | 49.331.16 | 93.020.97 | 27.911.30 |
CD-DCpr | 68.991.73 | 82.042.23 | 21.561.94 | 86.701.22 | 49.161.31 | 92.760.85 | 27.771.58 |
As further illustration in Figure 8 of the zero-shot classification accuracy, we provide the results for one of the models, namely “decoritems_lamp1” for all the datasets for DreamBooth. We can observe that the adaptation leads to drops on most classes (identified in red) but can also occasionally result in a performance increase (in green).

DINO | CLIP-I | CLIP-T | |
---|---|---|---|
DB | 0.42410.1503 | 0.67640.1046 | 0.78960.0296 |
DB-DC | 0.42830.1584 | 0.68170.1097 | 0.77990.0324 |
DB-DCpr | 0.43150.1585 | 0.68410.1086 | 0.77760.0322 |
CD | 0.44220.1378 | 0.69340.0902 | 0.79160.0266 |
CD-DC | 0.43810.4381 | 0.69250.0888 | 0.78990.0280 |
CD-DCpr | 0.43820.1351 | 0.69350.0872 | 0.78720.0264 |
B.3 Appearance drift full tables
See Table 8 for the full results which have been used for the generation of Figure 4. The prompt are ‘0:Face’,‘1:Pedestrian’,‘2:Car’,‘3:Cow’,‘4:Hand’,‘5:Chair’,‘6:Mountain’, ‘7:Beach’,‘8:Forest’,‘9:Highway’,‘10:Street’,‘11:Indoor’, ‘12:Animal in natural scene’, ‘13:Tree in urban scene’, ‘14:Close-up person in urban scene’,‘15:Far pedestrian in urban scene’,‘16:Car in urban scene’,‘17:Lamp in indoor scene’, ‘18:empty prompt’.
Prompt | Vanilla | DB | DB-DC | CD | CD-DC | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
CDI | FID | KID | CDI | FID | KID | CDI | FID | KID | CDI | FID | KID | CDI | FID | KID | |
00 | 0.10 | 24.75 | 0.00 | 0.65 | 43.74 | 2.00 | 0.23 | 21.70 | 0.16 | 0.48 | 37.01 | 1.22 | 0.24 | 24.90 | 0.29 |
01 | 0.14 | 44.55 | 0.00 | 1.01 | 79.78 | 3.22 | 0.43 | 46.16 | 0.50 | 0.82 | 72.44 | 2.63 | 0.33 | 51.68 | 0.79 |
02 | 0.12 | 19.70 | 0.01 | 0.35 | 28.31 | 0.90 | 0.18 | 18.37 | 0.12 | 0.31 | 27.14 | 0.58 | 0.19 | 20.11 | 0.15 |
03 | 0.09 | 29.17 | 0.01 | 0.36 | 41.81 | 1.22 | 0.14 | 27.40 | 0.13 | 0.40 | 42.25 | 1.13 | 0.23 | 31.62 | 0.37 |
04 | 0.17 | 35.93 | 0.01 | 0.59 | 40.77 | 0.59 | 0.23 | 30.53 | 0.07 | 0.57 | 42.65 | 0.70 | 0.32 | 33.76 | 0.22 |
05 | 0.14 | 19.65 | 0.01 | 0.76 | 28.15 | 0.61 | 0.27 | 17.93 | 0.08 | 0.58 | 26.45 | 0.57 | 0.34 | 20.09 | 0.17 |
06 | 0.09 | 34.22 | 0.00 | 0.48 | 52.80 | 1.46 | 0.26 | 34.55 | 0.32 | 0.57 | 60.90 | 2.16 | 0.34 | 41.52 | 0.72 |
07 | 0.14 | 34.89 | 0.05 | 0.36 | 42.58 | 0.64 | 0.25 | 29.45 | 0.14 | 0.41 | 48.98 | 1.19 | 0.22 | 33.66 | 0.26 |
08 | 0.10 | 25.50 | 0.00 | 0.48 | 52.42 | 2.79 | 0.25 | 26.78 | 0.57 | 0.42 | 55.33 | 2.95 | 0.29 | 32.76 | 1.01 |
09 | 0.12 | 27.51 | 0.01 | 0.51 | 41.82 | 1.73 | 0.30 | 28.22 | 0.30 | 0.43 | 41.37 | 1.40 | 0.33 | 27.80 | 0.52 |
10 | 0.10 | 24.34 | 0.00 | 0.62 | 47.28 | 2.78 | 0.34 | 26.27 | 0.68 | 0.69 | 44.67 | 2.09 | 0.42 | 28.39 | 0.72 |
11 | 0.06 | 33.56 | 0.02 | 0.38 | 46.44 | 1.39 | 0.19 | 30.86 | 0.27 | 0.53 | 52.92 | 1.86 | 0.26 | 35.79 | 0.53 |
12 | 0.11 | 29.75 | 0.01 | 0.43 | 43.92 | 1.20 | 0.24 | 28.67 | 0.21 | 0.41 | 49.51 | 1.69 | 0.27 | 36.02 | 0.68 |
13 | 0.13 | 20.26 | 0.00 | 0.66 | 44.59 | 1.50 | 0.64 | 20.36 | 0.47 | 0.62 | 43.14 | 1.40 | 0.42 | 30.33 | 0.37 |
14 | 0.12 | 40.60 | 0.01 | 0.53 | 47.32 | 0.58 | 0.33 | 36.76 | 0.16 | 0.60 | 48.05 | 0.66 | 0.33 | 38.68 | 0.18 |
15 | 0.07 | 36.90 | 0.01 | 0.96 | 60.22 | 1.96 | 0.48 | 38.81 | 0.51 | 0.75 | 59.16 | 1.92 | 0.46 | 40.97 | 0.51 |
16 | 0.09 | 26.65 | 0.00 | 0.68 | 37.07 | 1.02 | 0.41 | 26.12 | 0.26 | 0.66 | 36.38 | 0.90 | 0.41 | 28.21 | 0.29 |
17 | 0.13 | 27.73 | 0.01 | 0.59 | 32.44 | 0.67 | 0.30 | 25.44 | 0.23 | 0.52 | 33.93 | 0.74 | 0.27 | 25.77 | 0.14 |
18 | 0.18 | 59.52 | 0.02 | 0.29 | 62.02 | 0.59 | 0.23 | 37.99 | 0.00 | 0.31 | 68.52 | 0.93 | 0.21 | 48.00 | 0.13 |
Appendix C Ablations
C.1 Does finetuning lead to loss of diversity?
Finetuning large foundational models on a limited set of images (typically around 5) of a specific subject can lead to overfitting, a phenomenon observed in previous studies such as DreamBooth. This overfitting often results in a loss of diversity in generated images and a noticeable shift towards the characteristics of the training subject. While prior regularization techniques have been employed to mitigate this shift, they have not fully resolved the issue, as our analysis demonstrates.
To assess the impact of finetuning on diversity, we adapt the metric introduced in the DreamBooth study. This metric quantifies diversity by calculating the average Learned Perceptual Image Patch Similarity (LPIPS) between generated images of the same subject using identical prompts. A higher LPIPS score indicates greater diversity among the generated images.
Our proposed method not only improves the mitigation of subject shifting, as evidenced in Figures 4 and 5, but also maintains the diversity of the original model. To validate this, we conducted an extensive evaluation using 100 different prompts, each generating 100 images. These prompts were sourced from the DiffusionDB subset, as detailed in Appendix B.1.
Figure 9 presents the results of our diversity analysis. The data demonstrates that our method preserves diversity at a level comparable to, or even exceeding, previous approaches. This finding is particularly significant as it indicates that our technique not only addresses the shifting problem more effectively but does so without compromising the model’s ability to generate diverse outputs.
The preservation of diversity while improving subject fidelity represents a crucial advancement in finetuning methodologies for large generative models. It ensures that the fine-tuned model retains its creative capacity and versatility across a wide range of prompts and subjects, even as it gains enhanced capabilities in representing specific training subjects. This balance between specificity and diversity is essential for the practical application of fine-tuned models in various creative and technical domains.

C.2 Increasing the buffer size reduces drift
To investigate the impact of buffer size on mitigating open-world forgetting, we conducted experiments varying the number of images in the replay buffer during model adaptation. Table 9 presents the results of this analysis, showing the effect of buffer size on both semantic drift (measured by zero-shot CIFAR10 classification accuracy) and appearance drift (measured by Color Drift Index, CDI).
Metric | 0 | 50 | 100 | 200 | 500 | 1000 | 2000 |
---|---|---|---|---|---|---|---|
Acc CIFAR10 | 63.2414.61 | 76.106.66 | 75.626.48 | 75.925.81 | 76.365.28 | 76.925.87 | 77.126.20 |
CDI | 0.87 | 0.64 | 0.58 | 0.56 | 0.70 | 0.60 | 0.51 |
The experiment suggests that incorporating a replay buffer, even of modest size, is beneficial for mitigating open-world forgetting, particularly in terms of semantic drift. However, the benefits of increasing buffer size show diminishing returns, especially for semantic preservation. For appearance drift, while there is a general trend towards improvement with larger buffers, significant drift persists regardless of buffer size.
C.3 Influence of the training images
Our experiments reveal that the characteristics of the training images used during model adaptation can significantly impact the nature and extent of appearance drift. To illustrate this effect, we conducted an experiment in Figure 10 focusing on how the background color in training samples influences the color distribution of generated images. The results show that the background color of the training images has a noticeable impact on the color distribution of the generated images, even when generating images of unrelated concepts.

These findings highlight the importance of carefully considering the visual characteristics of training images when adapting generative models. The background, lighting, and overall composition of training samples can have far-reaching effects on the model’s output distribution, extending beyond the specific concept being learned.