This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

[1,2]\fnmQuentin \surJodelet

[1]\orgdivDepartment of Computer Science, \orgnameTokyo Institute of Technology, \orgaddress\streetW8-59 2-12-1 Ookayama, \cityMeguro, \postcode152-8552, \stateTokyo, \countryJapan

2]\orgdivArtificial Intelligence Research Center, \orgnameAIST, \orgaddress\street2-4-7 Aomi, \cityKoto, \postcode135-0064, \stateTokyo, \countryJapan

Future-Proofing Class-Incremental Learning

[email protected]    \fnmXin \surLiu [email protected]    \fnmYin Jun \surPhua [email protected]    \fnmTsuyoshi \surMurata [email protected] * [
Abstract

Exemplar-Free Class Incremental Learning is a highly challenging setting where replay memory is unavailable. Methods relying on frozen feature extractors have drawn attention recently in this setting due to their impressive performances and lower computational costs. However, those methods are highly dependent on the data used to train the feature extractor and may struggle when an insufficient amount of classes are available during the first incremental step. To overcome this limitation, we propose to use a pre-trained text-to-image diffusion model in order to generate synthetic images of future classes and use them to train the feature extractor. Experiments on the standard benchmarks CIFAR100 and ImageNet-Subset demonstrate that our proposed method can be used to improve state-of-the-art methods for exemplar-free class incremental learning, especially in the most difficult settings where the first incremental step only contains few classes. Moreover, we show that using synthetic samples of future classes achieves higher performance than using real data from different classes, paving the way for better and less costly pre-training methods for incremental learning.

keywords:
Class Incremental Learning, Continual Learning, Image Classification, Image Generation

1 Introduction

The version of record of this article, first published in “Machine Vision and Applications”, is available online at Publisher’s website: https://doi.org/10.1007/s00138-024-01635-y

While humans naturally learn and accumulate knowledge continuously during their whole lifespan, deep neural networks struggle to do the same. Continuously training deep neural networks on new data will inevitably induce a severe decrease of the performance on the previously learned data. This phenomenon, named catastrophic forgetting [1, 2], has received more and more attention during the past decade, and mitigating it is the core problem of incremental learning and continual learning. Among the different settings of incremental learning, Class-Incremental Learning (CIL) is among the most challenging ones. In this setting, the model is incrementally trained on a few number of classes at a time, without having access to the previously learned ones. The objective is to train a unified model that performs well on every encountered class.

Rehearsal memory [3, 4, 5] is one of the most commonly used methods to mitigate catastrophic forgetting and is used by most of the approaches for Class-Incremental Learning. It consists of a small buffer containing few representative exemplars of each previously learned class that are later used for replay while learning new classes. However, there remain situations where using a rehearsal memory is not possible, either due to privacy concerns regarding the training data (especially for tasks such as person re-identification) or due to limitations in the storage space. This led to the emergence of a new setting named Exemplar-Free Class-Incremental Learning (EFCIL) or Class-Incremental Learning Without Memory. Methods for EFCIL often freeze the feature extractor after the first incremental step [6, 7], achieving competitive performances while allowing for a faster and less computationally expensive learning of new classes compared to methods that have to backpropagate through the entire deep neural network. However, this may also lead to poor performance when the number of classes available in the initial step to train the feature extractor is too limited.

To address this limitation, several works have proposed to pre-train the feature extractor [5, 8] using additional data from classes similar yet different from those in the target dataset. However, with the advent of foundational diffusion model for image generation, pre-training the feature extractor on a more targeted dataset, specifically classes that are highly likely to be encountered during the future incremental steps, has the potential to lead to better improvements in this setting. While the aspect of using future classes to pre-train the model beforehand seems counter to the concept of Class-Incremental Learning, we argue that it is indeed a realistic setting. In many real world applications, collecting training data for every possible object in an ever changing world to train an omniscient classifier is impossible. It is much more realistic to train a classifier that is more specific to the needs of the user. Even under such a scenario, collecting every possible object that the user needs is an insurmountable task. Therefore, it is much more desirable to attempt to predict specific objects that are likely to be encountered, pre-train the feature extractor, and then perform Incremental Learning while it is deployed to the user. In our case, using a diffusion model that generates training data makes it possible to pre-train especially when objects can be difficult or impossible to obtain.

In this work, we propose to leverage pre-trained text-to-image diffusion models to better prepare Exemplar-Free Class-Incremental Learning methods relying on a frozen feature extractor for the future. The diffusion models generate synthetic samples of possible future classes that will be jointly used with actual real data during the first incremental step to train a feature extractor that better generalizes to future classes. Experiments using various challenging settings and datasets for EFCIL highlight how our method can be combined with state-of-the-art approaches and results in significant improvement of the average incremental accuracy. Moreover, it appears that using synthetic images of future classes achieves higher performance than using real images from different yet similar classes.

Our contributions can be summarized into three points:

  • We propose a framework relying on pre-trained diffusion models to better prepare Exemplar-Free Class-Incremental Learning methods for the future, reaching new state-of-the-art performances.

  • While previous approaches for Incremental Learning used pre-trained diffusion models for generating samples from past classes, our proposed method aims at generating samples from future classes instead.

  • We experimentally show that using synthetic samples of future classes for training the feature extractor achieves higher performance than using real images of classes different from those in the target dataset.

2 Related Work

2.1 Class-Incremental Learning (CIL) and Exemplar-Free Class-Incremental Learning (EFCIL)

While Class-Incremental Learning may be applied to various tasks and settings such as semantic segmentation [9, 10] or federated learning [11, 12], in this work we only consider the image classification task which is the most extensively studied. Most approaches for Class-Incremental Learning [13, 14] leverage knowledge distillation [3, 15, 16, 17], a rehearsal memory and a method to mitigate the bias toward the new classes [15, 18, 19, 20] induced by the limited size of the replay memory. Recent works [21, 22] proposed to extend the feature extractor at each incremental step to learn the new classes while preserving past knowledge by freezing previously learned weights.

Exemplar-Free Class-Incremental Learning is a more challenging setting in which there is no rehearsal memory, making the replay of samples from the previously encountered classes impossible. Similarly to approaches designed for Class-Incremental Learning, numerous methods for Exemplar-Free Class-Incremental Learning also continuously finetune the feature extractor and the classifier [23, 24, 25, 26, 27]. Several authors have highlighted the risk of overfitting on the current task and the importance of having general and transferable feature representation for EFCIL. To address this concern, approaches often rely on self-supervised learning [23, 26] or by creating augmented classes using Mixup [28]. Recently, methods relying on a fixed feature extractor have gained more attention [6, 29]. Those methods either rely on a pre-trained model [30, 31, 32, 8] or by freezing the feature extractor after learning it during the initial step [6, 7, 33]. Petit et al. [7] proposed FeTrIL which achieves state-of-the-art performances by relying on a frozen feature extractor and a linear classifier trained using actual features of new classes and pseudo-features of past classes generated by a geometric translation of new class features. FeCAM [33] is a recently proposed approach for Exemplar-Free Class-Incremental Learning with a fixed feature extractor that rely on the Mahalanobis distance combined with Tukey’s transformation, covariance shrinkage, and correlation normalization. Those methods exhibit numerous advantages including a less computationally expensive training procedure that allows for for a deployment on edge devices. However, they also depend on the initial training strategy used [34, 35] and may achieve poor performance if the quantity of data available for training the feature extractor during the initial step is too limited.

2.2 Training on synthetic images

While few attempts have been made using GAN-based models [36, 37], the use of synthetic images for training deep learning models has recently gained momentum following the advances in diffusion models [38, 39] and their wide availability. Diffusion models generate synthetic samples using an iterative denoising process conditioned on an input, usually a text prompt. Imagen [40] performs the denoising task in the pixel space while Stable Diffusion [41] and DALL-E2 [42] perform it in the latent space in order to decrease the computational cost of the generative process. In one of the pioneer works, Sarıyıldız et al. [43] studied the use of the pre-trained text-to-image diffusion model Stable Diffusion to generate a synthetic replacement for the original training dataset. Their experiments highlighted that while the model trained using the real train dataset reaches a higher accuracy on the real test dataset compared to the one trained on synthetic images, they perform on par for transfer learning tasks. Concurrently, Zhang et al. [44] proposed to expand real datasets using synthetic samples generated with various diffusion models and achieved significant improvement on various image classification datasets. Several authors proposed to use synthetic images generated by pre-trained diffusion models for different problems of computer vision such as zero-shot and few-shot learning [45], long-tailed image classification [46], incremental learning [47, 48], out-of-distribution detection [49], or semantic segmentation [50]. Tian et al. [51] showed that the representation learned using contrastive learning on synthetic images generated by Stable Diffusion outperforms state-of-the-art methods trained on real datasets.

2.3 Additional data for Incremental Learning

Several authors [52, 53, 54, 55, 56] have considered the use of additional data for Class-Incremental Learning by leveraging datasets of real images belonging to classes different from the ones that the model has to learn. Most of these approaches assume that the additional dataset is unlabeled, restricting the use of those additional data samples to the distillation loss [52, 53, 56]. Jodelet et al. [47] first proposed to use large pre-trained text-to-image diffusion models for Incremental Learning in order to generate additional synthetic images for the classes previously encountered by the model. Using a textual prompt based on the class name, this method generates labeled samples that can be used for the distillation loss but also for any supervised losses such as the classification loss. More recently, pre-trained diffusion models were also used for Class-Incremental semantic segmentation [57] and object detection [58]. However, there remains a significant distribution gap between the real samples and the synthetic ones [47] whose impact on the classifier makes the use of pre-trained diffusion model for replaying samples from past classes especially difficult for Exemplar-Free Class-Incremental Learning.

2.4 Preparing for the future

While most of the methods for Class-Incremental Learning only focus on how to prevent the forgetting of past data, some authors have studied how to better prepare the model for future updates. Douillard et al. [59] took inspiration from zero-shot learning and proposed to incorporate into the classifier feature vectors of unseen future classes during training. Those possible future vectors, named ghost features, are produced by a generator using actual features of seen classes and prior information about all the classes. [60] proposed to use a pre-allocated Regular Polytope Classifier which can exploit future unseen classes as negative examples. Zhou et al. [61] proposed to reserve embedding space for future classes by pre-assigning virtual prototypes and by forecasting future classes using manifold Mixup [28]. Bellitto et al. [55] proposed to use an auxiliary dataset of real images containing classes unrelated to the main task in order to prepare the model for the future. Those additional data samples are used to learn more general and stable representation that can be leveraged when learning future classes.

Compared to previously mentioned methods, in this work, we aim to leverage pre-trained text-to-image diffusion models not for generating additional past data but for generating synthetic samples for possible future classes. We study how those synthetic images of future classes can be used during the initial incremental step and to what extent they can help to better prepare Exemplar-Free incremental learning methods relying on a frozen feature extractor.

3 Proposed method

The objective of Class-Incremental Learning (CIL) is to learn a unified model θ\theta on a gradually increasing set of classes. θ\theta can be decomposed into {Φ,𝒲}\{\Phi,\mathcal{W}\} with Φ\Phi the feature extractor and 𝒲\mathcal{W} the classifier. The training procedure is divided into TT incremental steps, each step consisting of a dataset 𝒟t\mathcal{D}_{t} containing samples from the set of new classes 𝒴t\mathcal{Y}_{t} such that i,j{1,2,,T},𝒴i𝒴j=\forall i,j\in\{1,2,...,T\},\mathcal{Y}_{i}\cap\mathcal{Y}_{j}=\emptyset for iji\neq j. Contrary to standard Class-Incremental Learning, in the Exemplar-Free Class-Incremental Learning (EFCIL) setting, the model does not have access to any replay memory: during each incremental step tt, the model θt\theta_{t} is trained solely on the current dataset 𝒟t\mathcal{D}_{t}. After each incremental step, the model is evaluated on the test dataset of all the classes observed so far without having any step descriptor.

3.1 Future-Proofing Class-Incremental Learning

Refer to caption
Figure 1: Diagram representing our proposed method Future-Proof Class Incremental Learning (FPCIL) which prepares the feature extractor for the future by jointly training it on an auxiliary dataset in addition to the current dataset during the initial step. The auxiliary dataset contains synthetic samples of future classes generated using a pre-trained text-to-image diffusion model. After the first incremental step, the feature extractor ϕ\phi is frozen and the weights of the classifier w~\tilde{w} corresponding to the future classes are removed. Afterward, in the following incremental step, only the classifier 𝒲\mathcal{W} is trained using dedicated method (e.g. FeTrIL, FeCAM).

In order to create a future-proof model, as illustrated in Figure 1, we introduce Future-Proofing Class Incremental Learning (FPCIL), a straightforward approach which can be combined with any method for Class-Incremental Learning relying on a frozen feature extractor. During the first incremental step, when training the feature extractor, in addition to the initial step dataset 𝒟1\mathcal{D}_{1}, an additional dataset 𝒮\mathcal{S} containing |𝒴𝒮|\lvert\mathcal{Y}_{\mathcal{S}}\rvert different classes is used. While any labeled datasets could be used for 𝒮\mathcal{S}, we propose a novel approach which consists in using images from classes that will be encountered by the model during the future incremental steps. In detail, the objective of the initial incremental step is extended from solely learning the initial classes 𝒴1\mathcal{Y}_{1} into learning 𝒴1𝒴𝒮\mathcal{Y}_{1}\cup\mathcal{Y}_{\mathcal{S}} where 𝒴𝒮=i=2T𝒴i\mathcal{Y}_{\mathcal{S}}=\bigcup_{i=2}^{T}\mathcal{Y}_{i} are the future classes. The feature extractor is jointly trained with a classifier on 𝒟1𝒮\mathcal{D}_{1}\cup\mathcal{S}. At the end of the initial step, the classifier is restricted to the base classes 𝒴1\mathcal{Y}_{1} by dropping the weights corresponding to the future classes from the auxiliary dataset 𝒮\mathcal{S}, and the feature extractor is frozen. Then, the training procedure continues according to the baseline method for the remaining incremental steps. Our proposed method is detailed in Algorithm 1.

Input: Data-flow {𝒟i}i=1T\{\mathcal{D}_{i}\}_{i=1}^{T} and pre-trained generative model G.
Ouput: Models {θi}i=1T\{\theta_{i}\}_{i=1}^{T} composed of {Φi,𝒲i}\{\Phi_{i},\mathcal{W}_{i}\}.
for t{1,2,,T}t\in\{1,2,...,T\} do
       if t=1t=1 then
             // Predict future classes based on initial classes using oracle or GPT
             𝒴𝒮\mathcal{Y}_{\mathcal{S}}\leftarrow PredictFuture(𝒴1\mathcal{Y}_{1})
            
             // Generate synthetic samples of future classes
             𝒮\mathcal{S}\leftarrow GenerateSynthetic(G, 𝒴𝒮\mathcal{Y}_{\mathcal{S}})
            
            
            // Train the model using real data of the first step and synthetic data of future steps
             Train θ1\theta_{1} using 𝒟1𝒮\mathcal{D}_{1}\cup\mathcal{S}
             Freeze Φ1\Phi_{1} and remove 𝒴𝒮\mathcal{Y}_{\mathcal{S}} weights from 𝒲1\mathcal{W}_{1}
            
      else
             // Incrementally update the classifier using dedicated method (e.g. FeTrIL, FeCAM)
             Train 𝒲t\mathcal{W}_{t} using 𝒟t\mathcal{D}_{t}
            
      
Algorithm 1 Future-Proof Class-Incremental Learning (FPCIL)

In the initial step, our proposed method FPCIL trains on both the initial training dataset 𝒟1\mathcal{D}_{1} and an additional dataset 𝒮\mathcal{S} which contains training data for predicted future classes. As only the data for the current step are available in the Class-Incremental Learning setting, we propose to use synthetic data instead of the real ones. We use a pre-trained text-to-image diffusion model to generate synthetic images from the future classes during the initial step. This means that only the labels of the future classes are necessary. Compared to recent approaches for Incremental Learning relying on a pre-trained diffusion model [47, 57, 58] for generating synthetic samples of past classes, our proposed method aims at generating synthetic samples of future classes instead. By primarily using the synthetic data to train the feature extractor, our proposed method circumvents the impact of the distribution gap between real and synthetic samples generated by the diffusion model, making it applicable to Exemplar-Free Class-Incremental Learning.

While the requirement of having access to the label of future classes may seem counter to the concept of Class-Incremental Learning, we argue that it is practical and compatible with most real-world applications of Class-Incremental Learning. In general, during the developing a deep learning model meant for continuously accumulating new knowledge while achieving a given task in an ever-evolving environment, both the task and the environment are known. Therefore, it is reasonable to assume that it is possible, based on this initial knowledge of the environment and task, to predict their possible future evolution. However, acquiring beforehand real data corresponding to these future states of the environment and task may be difficult, expensive, or even impossible in some cases which lead us to use synthetic samples instead. As a concrete example, we propose to consider the case of an agent that would be deployed on an embedded device for assisting its user inside his home. The agent would have to be able to learn new concepts based on data directly collected from its environment depending on the particular need of its user. It would have to rely on incremental learning as frequent retraining from scratch would not be possible on either a cloud server as uploading user’s data would be difficult due to privacy reasons, or in local due to the limited computational and memory capacity of the embedded device. Therefore, using our proposed method it is possible to train an initial model precisely tailored for each user (or group of users) before deployment, by relying on available knowledge about the user and its environment (interests, location, particular requests, etc.) to preemptively learn specific concepts that may appear in the future in addition to generic concepts. Since building a specific dataset of real samples for all possible needs of all possible users would not be tractable due to the significant cost in time and resources of the data collection process, the use of synthetic data appears as a faster and cheaper alternative. As a second example, we consider the case of an agent deployed on an embedded device for performing a task in a hard to access and potentially hostile environment where the agent has to be autonomous due to unreliable remote control possibility (deep sea, space, etc). The nature of the environment intrinsically limits the capacity to collect large training datasets in advance but its properties may be known. Using our proposed method, the agent could be trained before its deployment using synthetic data from generative diffusion models, or more realistically in this scenario using simulation-based tools. The agent would then be able to incrementally learn and adjust itself to the environment using real data collected while performing the task.

It should also be noted that based on our experiments detailed in the ablation study, the classes in the auxiliary dataset 𝒮\mathcal{S} do not need to be perfectly predicted to have a positive impact on the performance. For clarity, we refer to FPCIL as “FPCIL-Oracle” when the list of future classes has been perfectly predicted and “FPCIL-Partial (k%)” otherwise, with kk the percentage of future classes correctly predicted.

Predictions of the future classes can be done through expert knowledge or through automated methods. Here, we introduce a simple approach for automatically predicting the future classes based on recently popular large language models. Our proposed method relies on the GPT model [62] “GPT-3.5 Turbo Instruct” used through the OpenAI API. The large language model is queried for a completion task using the prompt “The dataset contains the following 100 classes: c1c_{1}, […], ckc_{k},” where c1c_{1} to ckc_{k} are the names of the kk classes from the initial step, with kk equals to |𝒴1|\lvert\mathcal{Y}_{1}\rvert. The number “100” is used here to encourage the model to generate a longer list of possible future classes as initial experiments without specifying a number would result in only a few number of guesses. In practice, the output of GPT has to be parsed to extract the predicted future classes. The completion function of GPT often continues the prompt as if it was part of a website or a scientific paper by appending imaginary details such as the number of images per class in the dataset, nonexistent download url for the dataset, or architecture of the model. Due to the random nature of the output from GPT, the prompt is queried 10 times and the predicted classes are ordered according to the number of times they have been predicted. The auxiliary dataset 𝒮\mathcal{S} is then restricted to the classes which were the most often predicted. We evaluated three versions with an increasing level of restriction: “FPCIL-GPT”, “FPCIL-GPT R1”, and “FPCIL-GPT R2”. Those approaches based on GPT also serve to imitate a more realistic application of Future-Proof Class-Incremental Learning by predicting various number of future classes with various accuracy.

3.2 Generating future data

Refer to caption
Figure 2: Synthetic images of future classes for CIFAR100 generated using different diffusion models. Images are displayed before being resized for better appreciation. First row has been generated by Stable Diffusion 1.4 with a guidance scale of 2.02.0, second row using Stable Diffusion 1.4 with a guidance scale of 7.57.5, and last row using DALL-E2. From left to right: butterfly, can, house, lion, mountain, orange, otter, pine tree, rose, and streetcar.

Leveraging recent advances in diffusion models and the wide availability of generative text-to-image models pre-trained on natural images, we propose to use them for generating the synthetic images dataset of future classes following the procedure initially defined by Sarıyıldız et al. [43].

The auxiliary dataset 𝒮\mathcal{S} is generated using a pre-trained diffusion model GG conditioned solely on a textual prompt. As proposed in [43], the default prompt used is c,dcc,d_{c} with cc the name of the class and dcd_{c} a simple description of the class. Adding the description of the class into the prompt is intended to ensure that the generated images belong to the targeted class and not to any homonyms: for example “crane” in the ImageNet dataset, which may refer to either the bird or the machine used to move heavy objects. The definition dcd_{c} also tends to restrain the domain of the generated images even though the domain could also be enforced by adding it to the prompt. To automatically create the prompts, we rely on WordNet [63] and use the lemmas of the synset as cc and the definition of the synset as dcd_{c}. For each class, nn synthetic images are generated with nn equals to the number of images per class in the target dataset. Compared to traditional pre-training and concurrent method [55], our proposed method does not require the costly and fastidious process of collecting, curating, and annotating a dataset of real images. Examples of synthetic images generated for CIFAR100 are displayed in Figure 2.

The guidance scale is an important hyper-parameter of diffusion models relying on classifier-free guidance [64]. To some extent, it can be used to control the trade-off between the diversity and the quality of the generated synthetic images. For Stable Diffusion, the default recommended value for the guidance scale is often 7.07.0 but following [47, 43] we use the lower value of 2.02.0 as a higher diversity is beneficial for pre-training.

While this generation procedure allows our proposed method to achieves significant improvement compared to baselines, there remains an important distribution gap between the real and synthetic data [47]. Reducing this distribution gap and its impacts on the training of the feature extractor would further improve the performance of our proposed method. As possible future research direction, the diffusion model could be finetuned using the data from the first incremental step, a more sophisticated prompting mechanism could be implemented, or a selection process could be used to discard the synthetic samples too far away from the distribution of the target dataset.

Table 1: Definition of the different auxiliary datasets used during the first incremental step for each setting. “Class overlap” is the percentage of classes (actual number of classes in parentheses) from the future incremental steps of the target dataset that are in the auxiliary dataset 𝒮\mathcal{S}, i.e. the number of correctly predicted future classes.
Target dataset Auxiliary dataset 𝒮\mathcal{S} Nature of data Setting Number of classes |𝒴𝒮|\lvert\mathcal{Y}_{\mathcal{S}}\rvert Samples per class nn Class overlap with future steps of target dataset Note
CIFAR100 (100 classes) CIFAR10 Real All 10 5000 0%
FPCIL-Partial (k%) Synthetic B50 Inc 50 500 k% k% of the classes have been randomly selected from the target dataset, the remaining have been randomly selected from the wrong predictions of GPT-3.5. For k=100, we use the denomination “FPCIL-Oracle”.
B0 Inc10 90 500 k%
B0 Inc5 95 500 k%
FPCIL-GPT Synthetic B0 Inc10 150 500 58.9% (53) Classes the most often predicted by GPT-3.5 based on the classes encountered in the first incremental step.
B0 Inc5 166 500 53.7% (51)
FPCIL-GPT R1 Synthetic B0 Inc10 90 500 47.8% (43) Subset of “FPCIL-GPT” containing a more restricted set of classes.
B0 Inc5 102 500 33.7% (32)
FPCIL-GPT R2 Synthetic B0 Inc10 53 500 33.3% (30) Subset of “FPCIL-GPT” containing a more restricted set of classes. More restrictive than R1.
B0 Inc5 58 500 21.0% (20)
ImageNet-Subset (100 classes) ImageNet-Compl. Real B50 Inc 50 \sim1300 0% Classes have been randomly sampled from the remaining 900 classes of ILSVRC 2012. No class overlap with ImageNet-Subset.
B0 Inc10 90 \sim1300 0%
B0 Inc5 95 \sim1300 0%
ImageNet-Compl. X2 Real B50 Inc 100 \sim1300 0% Classes have been randomly sampled from the remaining 900 classes of ILSVRC 2012. No class overlap with ImageNet-Subset.
B0 Inc10 180 \sim1300 0%
B0 Inc5 190 \sim1300 0%
ImageNet-Compl. Animals Real B0 Inc10 90 \sim1300 0% Classes have been randomly sampled among the “animals” classes from the remaining 900 classes of ILSVRC 2012. No class overlap with ImageNet-Subset.
B0 Inc5 95 \sim1300 0%
FPCIL-Oracle Synthetic B50 Inc 50 1300 100% (50) Future classes perfectly predicted.
B0 Inc10 90 1300 100% (90)
B0 Inc5 95 1300 100% (95)
FPCIL-Partial (0%) Synthetic B50 Inc 50 1300 0% Same classes as “ImageNet-Compl.”.
B0 Inc10 90 1300 0%
B0 Inc5 95 1300 0%
FPCIL-Partial Animals (0%) Synthetic B0 Inc10 90 1300 0% Same classes as “ImageNet-Compl. Animals”.
B0 Inc5 95 1300 0%

4 Experiments

4.1 Experimental setup

4.1.1 Datasets

The experiments are conducted on two large scale datasets: CIFAR100 [65] and ImageNet-Subset. CIFAR100 is a dataset containing 60,000 32×\times32 RGB images evenly divided among 100 classes with 500 training images and 100 testing images per class. ImageNet-Subset is subset of the high-resolution images dataset ImageNet (ILSVRC 2012) [66]. This subset contains 100 randomly selected classes with about 1,300 training images and 50 testing images per class. Following Rebuffi et al. [3], the classes are ordered using numpy with the random seed 1993.

Two training procedures are used: Training from Scratch (TFS) and Training From Half (TFH). For the Training from Scratch procedure, the classes are evenly divided into TT (10, or 20) incremental steps, each incremental step containing the same number of classes. For the Training From Half procedure, following [15], the first step contains half of the classes and the remaining classes are evenly divided into the following T1T-1 incremental steps. Following Zhou et al. [14], the two training procedures are referred to using the common naming system “B-ii Inc-jj with ii the number of classes in the first step and jj the number of classes each of the following incremental step; ii equals to zero is used to refer to the Learning From Scratch procedure.

4.1.2 Model training

Following common practice for Exemplar-Free Class-Incremental Learning [7, 26, 27], ResNet-18 is used for both CIFAR100 and ImageNet-Subset. The feature extractor is trained solely during the first incremental step and is frozen for the rest of the training procedure. The backbone is trained with SDG using the standard Softmax Cross-Entropy loss and AutoAugment [67] for 160 epochs with a batch size of 128 and a weight decay of 5e45e^{-4}. The initial learning rate is set to 0.1 and decrease to zero using a cosine annealing schedule. For the classifier, we use a fully-connected linear layer trained following FeTrIL [7]. The classifier is trained using SGD for 50 epochs at each incremental step, with an initial learning rate of 0.1 and cosine annealing schedule. In the ablation study, we also report results using a Nearest Class Mean (NCM) Classifier and FeCAM [33]. If not stated otherwise, the reported results for the different baselines have been reproduced by our own experiments. Experiments are conducted using Pytorch [68] and PyCIL [69].

Table 2: Performances on CIFAR100 with the Learning from Half and Learning From Scratch procedures. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-18 as the backbone. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold and second best is underlined. Results averaged over 3 random runs.
Methods CIFAR100 - LFH CIFAR100 - LFS
B50 Inc10 B50 Inc5 B50 Inc2 B0 Inc10 B0 Inc5
AVG Last AVG Last AVG Last AVG Last AVG Last
FeTrIL [7] 68.05 58.69 66.81 58.22 65.46 56.94 50.50 33.86 40.80 24.65
FeTrIL w/ Aux. CIFAR10 70.60 60.82 69.34 60.03 68.23 59.26 58.56 41.72 52.42 35.03
FeTrIL w/ FPCIL-Oracle 73.44 65.71 72.81 65.20 72.40 64.76 68.09 55.43 62.92 49.67
FeTrIL w/ FPCIL-Partial (0%) 70.25 61.41 69.27 60.56 68.13 59.67 62.63 47.71 57.46 42.71
FeTrIL w/ FPCIL-GPT - - - - - - 67.18 53.89 62.15 48.67
FeTrIL w/ FPCIL-GPT R1 - - - - - - 66.79 53.02 60.65 46.08
FeTrIL w/ FPCIL-GPT R2 - - - - - - 65.25 51.01 59.89 44.33
Table 3: Performances on ImageNet-Subset with the Learning from Half and Learning From Scratch procedures. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-18 as the backbone. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold and second best is underlined. Results averaged over 3 random runs.
Methods ImageNet-Subset - LFH ImageNet-Subset - LFS
B50 Inc10 B50 Inc5 B50 Inc2 B0 Inc10 B0 Inc5
AVG Last AVG Last AVG Last AVG Last AVG Last
FeTrIL [7] 72.97 64.49 72.01 63.54 71.02 62.19 51.64 34.53 39.64 22.42
FeTrIL w/ Aux. ImageNet-Compl. 75.19 67.81 74.08 66.96 73.25 65.71 70.62 59.53 68.14 57.45
FeTrIL w/ Aux. ImageNet-Compl. X2 76.33 69.52 75.30 68.53 74.69 67.89 73.96 64.68 73.01 62.98
FeTrIL w/ FPCIL-Oracle 77.21 71.80 76.81 71.51 76.84 71.17 74.04 65.65 72.64 64.29
FeTrIL w/ FPCIL-Partial (0%) 74.90 67.27 73.93 66.37 73.18 65.25 68.89 57.21 65.94 55.09

4.1.3 Synthetic generation

Two diffusion models are used for generating the synthetic samples during training: Stable Diffusion [41] and DALL-E2 [42]. If not specified otherwise, the default diffusion model is Stable Diffusion v1.4 111https://huggingface.co/CompVis/stable-diffusion-v1-4 used with a guidance scale of 2.02.0 for 50 steps. Following [43], c,dcc,\;d_{c} is used for the prompt. The same datasets of synthetic images are used for all the experiments. Those datasets have been generated prior to incremental learning experiments, at size 512×\times512 using a fixed seed and contains 500 images per class for CIFAR100 and 1300 images per class for ImageNet-Subset. For CIFAR100, the synthetic images are resized to 32×\times32 before being fed to the model. For DALL-E2, because the model is used through the official API of OpenAI, it is not possible to control the different hyper-parameters of the model; the default, unknown, values are used for the guidance scale, the random seed and the number of denoising step. A slightly different prompt is used for DALL-E2, “a photo of a c,dcc,\;d_{c}, as the default one would generate unsatisfactory images for some classes. In order to comply with the moderation policy of the API, few definitions dcd_{c} have been changed.

4.1.4 Baselines

As a comparison baseline, a dataset containing real images from classes similar yet different to the ones in the target dataset is used as auxiliary dataset 𝒮\mathcal{S} during the first incremental step. The purpose of this baseline is to compare our proposed method with the more standard pre-training approach. It is similar to the method proposed by Bellitto et al. [55] adapted to Exemplar-Free Class-Incremental Learning with a frozen feature extractor. For CIFAR100, we use CIFAR10 for the auxiliary dataset as both datasets were created using the same annotation and curation methodology from the TinyImages Dataset [70]. However, while CIFAR10 contains more samples than the datasets generated using our proposed method FPCIL, it is important to note that CIFAR10 contains only 10 classes. For ImageNet-Subset, the auxiliary dataset “ImageNet-Compl.” was created by randomly selecting classes from the complement of ImageNet-Subset in ImageNet. “ImageNet-Compl.” contains the same number of classes as the datasets generated by our proposed method FPCIL and approximately the same number of samples. This is a strong baseline as both subsets are sampled from the same dataset and, while they do not contain the exact same classes, classes in both datasets are from the same categories and some are highly similar: for example both datasets contain different breeds of dogs or species of insects. We additionally use “ImageNet-Compl. X2” which contains two times more classes than ImageNet-Compl. We also report the performance of “FPCIL-Partial (k%)” which generates a dataset of the same size as FPCIL-Oracle but with only kk% of the classes correctly predicted. For CIFAR100, this auxiliary dataset is built by randomly sampling kk% of the classes from FPCIL-Oracle and the remaining ones from the wrong prediction of GPT-3.5. For ImageNet-Subset, “FPCIL-Partial (0%)” contains synthetic images belonging to the same classes as the ones in ImageNet-Compl. The definition of all the different auxiliary datasets used are summarized in Table 1.

4.1.5 Evaluation

Models are evaluated and compared using both the “Final Accuracy” and the “Average Incremental Accuracy” [3]. The Average Incremental Accuracy is defined as the average of the Top-1 accuracy of the model on the test dataset of all the classes learned so far, at the end of each incremental step including the first one.

4.2 Results

Average incremental accuracy and final accuracy of our proposed method FPCIL are reported for CIFAR100 and ImageNet-Subset in Table 2 and Table 3 for both the Learning From Scratch and Learning From Half procedures.

Our proposed approach FPCIL considerably improves the average incremental accuracy of the state-of-the-art method FeTrIL [7], from 5.39 percentage points (p.p.) up to 22.12p.p. on CIFAR10 and from 4.24p.p. up to 33.00p.p. on ImageNet-Subset. This improvement is notably higher for the more difficult Learning From Scratch procedure which is particularly challenging for FeTrIL due to the limited quantity of data available during the first incremental step. Compared to using an auxiliary dataset of real images from classes different from those in the target dataset, our method also achieves significantly higher accuracy. This highlights the fundamental advantage of our proposed method: while the standard pre-training procedure improves the generalization capability of the feature extractor by leveraging additional data belonging to classes different from those in the target dataset, our proposed method aims to learn a feature extractor specifically designed for the incoming task by relying on data from future classes instead. Even if those data from future classes are synthetic, it allows to learn a feature extractor more suitable for the future tasks resulting in a significantly higher accuracy compared standard pre-training on real data. Moreover, on ImageNet-Subset, our proposed method FPCIL even outperforms a two times larger auxiliary dataset of real images, demonstrating that our proposed algorithm is more data-efficient that standard pre-training. Furthermore synthetic datasets can be rapidly generated whereas auxiliary datasets of real data require a long and costly collection, curation, and annotation process. Finally, even in the case where the future classes have not been accurately predicted, the auxiliary dataset of synthetic images performs on par with the standard pre-training approach relying on real datasets: for the Learning From Half procedure, the difference of average incremental accuracy between using ImageNet-Compl or its synthetic counterpart is at most 0.29p.p.. This makes our proposed method beneficial in every setting.

On CIFAR100, our more practical approach FPCIL-GPT also reaches a higher accuracy than the baseline method FeTrIL yet slightly lower than the FPCIL-Oracle. On CIFAR100 in the Learning From Scratch procedure with 10 steps of 10 classes (B0 Inc10), the less restricted FPCIL-GPT predicted a total of 150 classes with 53 that actually are among the 90 future classes while the most restricted FPCIL-GPT R2 predicted only 53 classes with 30 among the 90 future classes. All the details regarding future class predictions using GPT are reported in Table 1. While the more restricted version FPCIL-GPT R1 and R2 tend to predict a lower number of classes that are not among the future classes, the less restricted version predicts a higher number of future classes, achieving higher performances. There is a trade-off between the precision and recall in the prediction of future classes, while predicting more classes improve the performance it also increase the computational cost of the first incremental step due to the cost of generating synthetic images and training the feature extractor. In the ablation study, we further analyze the impact of the correctness of the prediction of future classes.

Table 4: Performances on CIFAR100 with the Learning From Scratch procedure depending on the ratio of correctly predicted future classes in the complementary dataset used during the initial step. Synthetics samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-18 as the backbone. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold. Results averaged over 3 random runs.
Methods CIFAR100 - B0 Inc10
AVG Last
FeTrIL [7] 50.50 33.86
FeTrIL w/ FPCIL-Oracle (100%) 68.09 55.43
FeTrIL w/ FPCIL-Partial (66%) 66.06 52.21
FeTrIL w/ FPCIL-Partial (50%) 64.79 50.76
FeTrIL w/ FPCIL-Partial (33%) 64.32 49.90
FeTrIL w/ FPCIL-Partial (0%) 62.63 47.71
Table 5: Performances on ImageNet-Subset with the Learning From Scratch procedure depending on the complementary dataset used during initial step. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-18 as the backbone. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold. Results averaged over 3 random runs.
Methods ImageNet-Subset - LFS
B0 Inc10 B0 Inc5
AVG Last AVG Last
FeTrIL [7] 51.64 34.53 39.64 22.42
FeTrIL w/ FPCIL-Oracle 74.04 65.65 72.64 64.29
FeTrIL w/ Aux. ImageNet-Compl. Animals 68.74 57.87 64.09 54.29
FeTrIL w/ Aux. ImageNet-Compl. 70.62 59.53 68.14 57.45
FeTrIL w/ FPCIL-Partial Animals (0%) 66.08 54.97 61.63 51.11
FeTrIL w/ FPCIL-Partial (0%) 68.89 57.21 65.94 55.09

4.3 Ablation study

4.3.1 Accurately predicting the future

In Table 4, we compare the performance of our proposed methods combined with FeTrIL on CIFAR100 B0 Inc10 depending on the ratio of correctly predicted future classes in the auxiliary dataset used during the initial step. Experimentally, for a fixed dataset size, the average incremental accuracy increase with the correctness of the prediction of the future classes. This confirms the importance of predicting the future classes compared to the standard pre-training procedure which relies on real data of classes different from those in the target dataset that will be encountered by the continual learner in the future. Moreover, even in the case where no future class has been correctly predicted, our proposed method still improves the performance of the baseline approach as, in most cases, additional data help to learn a feature extractor with better generalization capability.

Additionally, in Table 5 we report the accuracy for an alternative version of ImageNet-Compl and FPCIL-Partial named respectively “ImageNet-Compl. Animals” and “FPCIL-Partial Animals”. ImageNet-Compl Animals is built by sampling classes from the complement to ImageNet-Subset in ImageNet while restricting the classes to animals; FPCIL-Partial Animals contains the same classes that ImageNet-Compl Animals but the images are generated using a diffusion model. These two more specialized datasets achieve notably lower accuracy than their more general version. This further emphasizes the importance of predicting the future classes and not limiting the scope of the auxiliary dataset.

4.3.2 Quality and diversity of synthetic samples

Table 6: Performances on CIFAR100 with the Learning from Half and Learning From Scratch procedures depending on the Diffusion Model used for generating the synthetic data: Stable Diffusion with a guidance scale of 2.02.0, Stable Diffusion with a guidance scale of 7.57.5, and DALL-E2. Using ResNet-18 as the backbone. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold. Results averaged over 3 random runs.
Methods CIFAR100
B50 Inc5 B0 Inc10 B0 Inc5
AVG Last AVG Last AVG Last
FeTrIL [7] 66.81 58.22 50.50 33.86 40.80 24.65
FeTrIL w/ FPCIL-Oracle SD g=2.0g=2.0 72.81 65.20 68.09 55.43 62.92 49.67
FeTrIL w/ FPCIL-Oracle SD g=7.5g=7.5 72.22 64.40 66.48 53.36 61.63 48.50
FeTrIL w/ FPCIL-Oracle DALL-E2 71.08 63.02 63.94 50.39 59.76 46.34
Table 7: Performances on CIFAR100 with the Learning from Half and Learning From Scratch procedures for FeTrIL depending on the number of synthetic samples nn generated for each future class. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-18 as the backbone. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Results averaged over 3 random runs.
Methods CIFAR100
B50 Inc5 B0 Inc10 B0 Inc5
AVG Last AVG Last AVG Last
FeTrIL [7] 66.81 58.22 50.50 33.86 40.80 24.65
FeTrIL w/ FPCIL-Oracle n=50n=50 68.31 59.67 60.45 45.56 55.52 40.89
FeTrIL w/ FPCIL-Oracle n=250n=250 71.27 63.23 65.53 52.40 61.45 48.38
FeTrIL w/ FPCIL-Oracle n=500n=500 72.81 65.20 68.09 55.43 62.92 49.67
FeTrIL w/ FPCIL-Oracle n=1000n=1000 73.95 66.77 70.24 58.37 64.89 52.73

In Table 6, the performance of FPCIL Oracle is reported on CIFAR100 depending on the diffusion model used to generate the synthetic images of future classes. By default in our study, we used Stable Diffusion with a guidance scale of 2.02.0 [43] to favor the diversity to the detriment of the quality. We verified this assumption by comparing it against Stable Diffusion with a higher guidance scale and DALL-E2 which both tend to favor more the quality than the diversity. Experiments confirm that when pre-training the feature extractor on future classes, the diversity is more important than the quality of the generated samples: Stable Diffusion with a small guidance scale achieves the highest results, ahead of DALL-E2 and Stable Diffusion with the default guidance scale.

In Table 7, the performance of FPCIL Oracle is reported on CIFAR100 for different number of synthetic samples nn generated for each future classes. While the default value of nn equals 500 was used in the other experiments in order to match the size of the target dataset, it appears that just 50 samples per class is enough to improve the performance of the baseline and even outperform the use of eternal dataset of unrelated classes such as CIFAR10 which is more than 10 times bigger. Moreover, it is possible to further improve the performance of our proposed method by generating more synthetic data if the constraints of the system permits it.

In addition to the trade-of between quality and diversity, it is important to emphasize that a significant gap remains between real images and the synthetic samples generated using pre-trained diffusion models. As illustrated in Table 5, while Aux. ImageNet-Compl. and FPCIL-Partial (0%) contain the same classes and almost the same number of samples, there is a difference of about 2 percentage points of accuracy between the feature extractor trained on those datasets. The exact same observation can be made for Aux. ImageNet-Compl. Animals and FPCIL-Partial Animals (0%). Similarly, while achieving significantly higher results than the baseline, FPCIL-Oracle still achieves lower performances on both CIFAR100 and ImageNet-Subset compared to a upper-bound model which would have been trained on the complete target dataset at once.

Table 8: Performances on CIFAR100 with the Learning from Half and Learning From Scratch procedures for FeTrIL depending on the nature (real or synthetic) of the data used to train the feature extractor during the first incremental step. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-18 as the backbone. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold. Results averaged over 3 random runs.
Methods CIFAR100
Initial classes Future classes B50 Inc5 B0 Inc10 B0 Inc5
AVG Last AVG Last AVG Last
FeTrIL [7] real - 66.81 58.22 50.50 33.86 40.80 24.65
FeTrIL w/ FPCIL-Oracle var1 - synth. 47.00 44.10 64.05 53.30 64.07 52.44
FeTrIL w/ FPCIL-Oracle var2 synth. synth. 58.04 53.52 65.56 54.42 65.24 53.51
FeTrIL w/ FPCIL-Oracle real synth. 72.81 65.20 68.09 55.43 62.92 49.67
Table 9: Performances on CIFAR100  LABEL:sub@tab:backbone:lfh B50 Inc5 and  LABEL:sub@tab:backbone:lfs B0 Inc10 depending on the backbone used. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold. Results averaged over 3 random runs.
Methods CIFAR100 - B50 Inc5
ResNet-32 ResNet-18
AVG Last AVG Last
FeTrIL [7] 64.27 55.19 66.81 58.22
FeTrIL w/ Aux. CIFAR10 64.48 55.85 69.34 60.03
FeTrIL w/ FPCIL-Oracle 66.24 58.88 72.81 65.20
FeTrIL w/ FPCIL-Partial (0%) 63.80 55.67 69.27 60.56
(a) CIFAR100 - B50 Inc5
Methods CIFAR100 - B0 Inc10
ResNet-32 ResNet-18
AVG Last AVG Last
FeTrIL [7] 46.32 29.80 50.50 33.86
FeTrIL w/ Aux. CIFAR10 56.70 39.42 58.56 41.72
FeTrIL w/ FPCIL-Oracle 62.73 48.96 68.09 55.43
FeTrIL w/ FPCIL-Partial (0%) 59.68 45.30 62.63 47.71
(b) CIFAR100 - B0 Inc10

4.3.3 Combining real and synthetic data

During the first incremental step, our proposed method FPCIL combines the real samples of the initial classes with synthetic samples of possible future classes to train the feature extractor. In Table 8, different variations of our proposed method are compared on CIFAR100 to estimate the contribution of each component. “FPCIL var1” does not use the real samples from the first classes and trains the feature extractor solely using synthetic samples of future classes. “FPCIL var2” also generates synthetic samples of the initial classes and train the feature extractor using synthetic samples of both initial and future classes. In both cases, the classifier is then retrained at the end of the first incremental step using the real samples of the initial classes. Our proposed approach achieves the best performance overall, confirming the importance of using the available real data and not relying only on synthetic data. Interestingly, when the number of real data in the first incremental step is extremely small, replacing them by synthetic samples improves the overall performance while decreasing the final accuracy on the first classes. This may be due to bias induced in the learned representation by the distribution gap between real and synthetic samples being exacerbated in the most extreme settings.

4.3.4 Architecture of feature extractor

Table 9 compares the performance of our method for two different feature extractors: the 11.2 million parameters ResNet-18 and the 0.46 million parameters ResNet-32. Due to the addition of the auxiliary dataset to the often limited dataset of the initial step, our method takes advantage of a larger feature extractor and results in a large improvement with the baseline compared to backbones with less parameters. Nonetheless, FPCIL remains competitive with smaller deep neural networks.

4.3.5 Pre-trained model and finetuning

Table 10: Performances on CIFAR100 with the Learning from Half and Learning From Scratch procedures for FeTrIL using a ResNet-18 feature extractor pre-trained on ImageNet. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-18 as the backbone. Input data are upscaled before being fed to the model in order to match the pre-training configuration. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold. Results averaged over 3 random runs.
Methods CIFAR100
B50 Inc5 B0 Inc10 B0 Inc5
AVG Last AVG Last AVG Last
FeTrIL [7] 66.73 57.98 51.05 34.43 42.71 26.19
FeTrIL w/ Pre-training 64.16 59.85 71.05 60.97 70.70 59.81
FeTrIL w/ Pre-training + Finetune 76.68 70.55 66.50 54.73 63.05 51.77
FeTrIL w/ FPCIL-Oracle 73.59 65.83 68.10 55.20 63.38 50.54
FeTrIL w/ Pre-training + FPCIL-Oracle 79.06 73.87 76.60 67.14 74.83 64.30

In addition to the main experiments in which the feature extractor is trained from scratch, we also conducted experiments with a pre-trained feature extractor. Using a feature extractor pre-trained on a large external dataset is an effective approach when the low number of classes in the initial incremental step is not enough to train the feature extractor from scratch.

Compared to experiments where the feature extractor is trained from scratch, when using a pre-trained model the initial learning rate for the feature extractor is set to 0.001 and the number of epochs is reduced to 80, other hyper-parameters are kept unchanged. In Table 10, the performances of FeTrIL on CIFAR100 using both a ResNet-18 feature extractor trained from scratch during the first incremental step and a ResNet-18 pre-trained on ImageNet. While it appears that slightly finetuning the feature extractor using the data from the initial incremental step is detrimental if the number of classes in too small, it may also further improve the performance when enough classes are available in the first incremental step, especially if there is a large domain gap between the target dataset and the pre-training dataset. In comparison, FPCIL can be combined with pre-training independently of the number of classes in the first incremental step, increasing the average incremental accuracy by 2.38p.p. up to 5.55p.p..This highlights the versatility of our proposed method FPCIL which can be used either for training the feature extractor from scratch or for finetuning a feature extractor pre-trained on a large dataset, significantly improving the performance of the baseline in both cases.

4.3.6 Comparison with additional baselines

Table 11: Performances on CIFAR100 with the Learning from Half and Learning From Scratch procedures for FeTrIL, FeCAM, and Nearest Class Mean (NCM) Classifier. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-18 as the backbone. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold. Results averaged over 3 random runs.
Methods CIFAR100
B50 Inc5 B0 Inc10 B0 Inc5
AVG Last AVG Last AVG Last
NCM 65.24 55.23 40.48 23.37 30.41 15.00
NCM w/ FPCIL-Oracle 71.61 63.21 60.98 46.39 55.69 41.10
Improvement in p.p. +6.37 +7.98 +20.50 +23.02 +25.28 +26.10
FeTrIL [7] 66.81 58.22 50.50 33.86 40.80 24.65
FeTrIL w/ FPCIL-Oracle 72.81 65.20 68.09 55.43 62.92 49.67
Improvement in p.p. +6.00 +6.98 +17.59 +21.57 +22.12 +25.02
FeCAM [33] 70.79 63.04 52.85 38.47 43.30 28.88
FeCAM w/ FPCIL-Oracle 76.77 70.49 71.14 59.95 68.31 56.05
Improvement in p.p. +5.98 +7.45 +18.29 +21.48 +25.01 +27.17
Table 12: Performances on CIFAR100 with the Learning From Scratch procedures. Synthetic samples generated using Stable Diffusion with a guidance scale of 2.02.0 . Using ResNet-32 as the backbone for CIFAR100 and ResNet-18 for ImageNet-Subset. Results marked with \dagger are reported from [71]. “AVG” is Average Incremental Accuracy (Top-1) and “Last” is the final overall accuracy of the model after the last incremental step. Best result is marked in bold. Results averaged over 3 random runs.
Methods CIFAR100 ImageNet-Subset
B0 Inc10 B0 Inc5 B0 Inc10 B0 Inc5
AVG Last AVG Last AVG Last AVG Last
DiffClass\dagger [71] 68.05 58.40 67.10 57.11 73.87 67.02 72.51 68.68
FeTrIL [7] 46.32 29.80 37.75 21.26 51.64 34.53 39.64 22.42
FeTrIL w/ FPCIL-Oracle 62.73 48.96 59.89 46.62 74.04 65.65 72.64 64.29
FeCAM [33] 44.12 27.83 35.57 20.04 55.37 40.03 43.48 29.11
FeCAM w/ FPCIL-Oracle 64.47 52.01 63.25 50.00 77.78 69.85 77.41 69.17

Additionally to FeTrIL [7], we report in Table 11 the performance of our method combined with a Nearest Class Mean (NCM) Classifier and FeCAM [33]. NCM is a simple yet effective method [34, 29] for Class-Incremental Learning with a fixed feature extractor that does not require memory. FeCAM is a recently proposed approach for Exemplar-Free Class-Incremental Learning achieving new state-of-the-art performances. Similarly to FeTrIL, the performance of both NCM and FeCAM is highly dependent on the quality of the features produced by the backbone. It struggles with the Learning From Scratch procedure where the number of classes and data samples used to train the feature extractor is limited. Our method improves the performance of the NCM classifier and FeCAM in every considered setting and achieves the most significant improvements with the Learning From Scratch procedure on CIFAR100 B0 Inc5, reaching a final accuracy up to two times higher than the baseline approach.

Furthermore, in Table 12 we compare our approach with the concurrently released method DiffClass [71] which leverages the Stable Diffusion model for generating samples from past classes. While our method takes advantage of a larger backbone and can achieve higher performances as detailed in the ablation study, the comparison on CIFAR100 is done using ResNet-32 for the backbone as the results for DiffClass using ResNet-18 are not available. Experimentally, DiffClass achieves higher performance on CIFAR100 but is outperformed by our approach on ImageNet-Subset. It should be noted that Meng et al. [71] proposed several new components to mitigate the distribution gap between real and synthetic samples [47]. Some of those contributions, such as a multi-distribution matching technique to finetune the diffusion model and a selective synthetic image augmentation technique, could be adapted and combined with our proposed approach FPCIL to further improve its performance. Finally, our proposed method FPCIL conducts the computationally expensive task of generating synthetic samples using diffusion model only during the first incremental step, before freezing the feature extractor and only updating the classifier afterward. This allows us to potentially conduct the first incremental step on specialised hardware in a data center before deploying the model on devices with limited hardware where each following incremental step could be conducted in a matter of minutes on a CPU using approaches such as FeTrIL or FeCAM. In comparison, DiffClass finetunes the diffusion model, generates synthetic samples of past classes, and finetunes the whole feature extractor at each incremental step, resulting in much larger computational cost and requiring specialised hardware during whole training procedure. This makes our approach preferable if applicable.

5 Conclusion

In this work, we introduced a new method for Exemplar-Free Class-Incremental Learning which leverages large pre-trained diffusion model to generate images from future classes. Experimental results highlight how our method can significantly improve the accuracy of state-of-the-art approaches while only modifying the initial step. We found that our method requires less additional data than traditional methods relying on real curated datasets. While we focused our present study to feature extractors trained from scratch during the first incremental step, in future work, we will further investigate how synthetic images of future classes can be used to adapt a general pre-trained foundation model. Moreover, we will also explore how synthetic data of future classes can be leveraged during the whole training procedure, not only during the first incremental step.

\bmhead

Acknowledgements

This work is partly supported by JSPS Grant-in-Aid for Scientific Research (grant number 23H03451, 21K12042) and the New Energy and Industrial Technology Development Organization (Grant Number JPNP20017).

\bmhead

Credit Author Statement Quentin Jodelet: Conceptualization, Investigation, Software, Writing - Original Draft. Xin Liu: Conceptualization, Supervision, Resources, Funding acquisition, Writing - Review & Editing. Yin Jun Phua: Conceptualization, Supervision, Writing - Review & Editing. Tsuyoshi Murata: Supervision, Resources, Funding acquisition, Writing - Review & Editing.

\bmhead

Data Availability

Benchmark datasets CIFAR-100, and ImageNet are publicly available. Diffusion models used for generating synthetic datasets are either publicly available or behind a paid API.

Declarations

\bmhead

Conflict of interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  • \bibcommenthead
  • McCloskey and Cohen [1989] McCloskey, M., Cohen, N.J.: Catastrophic interference in connectionist networks: The sequential learning problem. Psychology of Learning and Motivation, vol. 24, pp. 109–165. Academic Press (1989). https://doi.org/10.1016/S0079-7421(08)60536-8 . http://www.sciencedirect.com/science/article/pii/S0079742108605368
  • Ratcliff [1990] Ratcliff, R.: Connectionist models of recognition memory: constraints imposed by learning and forgetting functions. Psychological review 97(2), 285 (1990)
  • Rebuffi et al. [2017] Rebuffi, S.-A., Kolesnikov, A., Sperl, G., Lampert, C.H.: icarl: Incremental classifier and representation learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)
  • Chaudhry et al. [2019] Chaudhry, A., Rohrbach, M., Elhoseiny, M., Ajanthan, T., Dokania, P.K., Torr, P.H.S., Ranzato, M.: On Tiny Episodic Memories in Continual Learning (2019)
  • Lomonaco et al. [2022] Lomonaco, V., Pellegrini, L., Rodriguez, P., Caccia, M., She, Q., Chen, Y., Jodelet, Q., Wang, R., Mai, Z., Vazquez, D., Parisi, G.I., Churamani, N., Pickett, M., Laradji, I., Maltoni, D.: Cvpr 2020 continual learning in computer vision competition: Approaches, results, current challenges and future directions. Artificial Intelligence 303, 103635 (2022) https://doi.org/%****␣sn-article.bbl␣Line␣125␣****10.1016/j.artint.2021.103635
  • Hayes and Kanan [2020] Hayes, T.L., Kanan, C.: Lifelong machine learning with deep streaming linear discriminant analysis. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 220–221 (2020)
  • Petit et al. [2023] Petit, G., Popescu, A., Schindler, H., Picard, D., Delezoide, B.: Fetril: Feature translation for exemplar-free class-incremental learning. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 3911–3920 (2023)
  • Zhou et al. [2024] Zhou, D.-W., Sun, H.-L., Ning, J., Ye, H.-J., Zhan, D.-C.: Continual Learning with Pre-Trained Models: A Survey (2024)
  • Douillard et al. [2021] Douillard, A., Chen, Y., Dapogny, A., Cord, M.: Plop: Learning without forgetting for continual semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4040–4050 (2021)
  • Cong et al. [2024] Cong, W., Cong, Y., Dong, J., Sun, G., Ding, H.: Gradient-semantic compensation for incremental semantic segmentation. IEEE Transactions on Multimedia 26, 5561–5574 (2024) https://doi.org/10.1109/tmm.2023.3336243
  • Dong et al. [2022] Dong, J., Wang, L., Fang, Z., Sun, G., Xu, S., Wang, X., Zhu, Q.: Federated class-incremental learning. In: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, ??? (2022). https://doi.org/10.1109/cvpr52688.2022.00992 . http://dx.doi.org/10.1109/CVPR52688.2022.00992
  • Dong et al. [2023] Dong, J., Zhang, D., Cong, Y., Cong, W., Ding, H., Dai, D.: Federated incremental semantic segmentation. In: 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, ??? (2023). https://doi.org/10.1109/cvpr52729.2023.00383 . http://dx.doi.org/10.1109/CVPR52729.2023.00383
  • Belouadah et al. [2021] Belouadah, E., Popescu, A., Kanellos, I.: A comprehensive study of class incremental learning algorithms for visual tasks. Neural Networks 135, 38–54 (2021) https://doi.org/10.1016/j.neunet.2020.12.003
  • Zhou et al. [2023] Zhou, D.-W., Wang, Q.-W., Qi, Z.-H., Ye, H.-J., Zhan, D.-C., Liu, Z.: Deep Class-Incremental Learning: A Survey (2023)
  • Hou et al. [2019] Hou, S., Pan, X., Loy, C.C., Wang, Z., Lin, D.: Learning a unified classifier incrementally via rebalancing. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019)
  • Douillard et al. [2020] Douillard, A., Cord, M., Ollion, C., Robert, T., Valle, E.: Podnet: Pooled outputs distillation for small-tasks incremental learning. In: Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XX 16, pp. 86–102 (2020). Springer
  • Dong et al. [2023] Dong, J., Liang, W., Cong, Y., Sun, G.: Heterogeneous forgetting compensation for class-incremental learning. In: 2023 IEEE/CVF International Conference on Computer Vision (ICCV). IEEE, ??? (2023). https://doi.org/10.1109/iccv51070.2023.01078 . http://dx.doi.org/10.1109/ICCV51070.2023.01078
  • Wu et al. [2019] Wu, Y., Chen, Y., Wang, L., Ye, Y., Liu, Z., Guo, Y., Fu, Y.: Large scale incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 374–382 (2019)
  • Jodelet et al. [2021] Jodelet, Q., Liu, X., Murata, T.: Balanced softmax cross-entropy for incremental learning. In: Farkaš, I., Masulli, P., Otte, S., Wermter, S. (eds.) Artificial Neural Networks and Machine Learning – ICANN 2021, pp. 385–396. Springer, Cham (2021)
  • Ahn et al. [2021] Ahn, H., Kwak, J., Lim, S., Bang, H., Kim, H., Moon, T.: Ss-il: Separated softmax for incremental learning. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 844–853 (2021)
  • Yan et al. [2021] Yan, S., Xie, J., He, X.: Der: Dynamically expandable representation for class incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3014–3023 (2021)
  • Zhou et al. [2023] Zhou, D.-W., Wang, Q.-W., Ye, H.-J., Zhan, D.-C.: A model or 603 exemplars: Towards memory-efficient class-incremental learning. In: ICLR (2023)
  • Wu et al. [2021] Wu, G., Gong, S., Queen, P.L.: Striking a balance between stability and plasticity for class-incremental learning. In: 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 1104–1113 (2021). https://doi.org/%****␣sn-article.bbl␣Line␣400␣****10.1109/ICCV48922.2021.00116
  • Zhu et al. [2021] Zhu, F., Cheng, Z., Zhang, X.-y., Liu, C.-l.: Class-incremental learning via dual augmentation. In: Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems (2021). https://openreview.net/forum?id=8dqEeFuhgMG
  • Jodelet et al. [2022] Jodelet, Q., Liu, X., Murata, T.: Balanced softmax cross-entropy for incremental learning with and without memory. Computer Vision and Image Understanding 225, 103582 (2022) https://doi.org/10.1016/j.cviu.2022.103582
  • Zhu et al. [2021] Zhu, F., Zhang, X.-Y., Wang, C., Yin, F., Liu, C.-L.: Prototype augmentation and self-supervision for incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5871–5880 (2021)
  • Zhu et al. [2022] Zhu, K., Zhai, W., Cao, Y., Luo, J., Zha, Z.-J.: Self-sustaining representation expansion for non-exemplar class-incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9296–9305 (2022)
  • Zhang et al. [2017] Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: Beyond Empirical Risk Minimization (2017)
  • Ostapenko et al. [2022] Ostapenko, O., Lesort, T., Rodríguez, P., Arefin, M.R., Douillard, A., Rish, I., Charlin, L.: Continual Learning with Foundation Models: An Empirical Study of Latent Replay (2022)
  • Wang et al. [2022a] Wang, Z., Zhang, Z., Ebrahimi, S., Sun, R., Zhang, H., Lee, C.-Y., Ren, X., Su, G., Perot, V., Dy, J., et al.: Dualprompt: Complementary prompting for rehearsal-free continual learning. In: European Conference on Computer Vision, pp. 631–648 (2022). Springer
  • Wang et al. [2022b] Wang, Z., Zhang, Z., Lee, C.-Y., Zhang, H., Sun, R., Ren, X., Su, G., Perot, V., Dy, J., Pfister, T.: Learning to prompt for continual learning. In: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, ??? (2022). https://doi.org/10.1109/cvpr52688.2022.00024 . http://dx.doi.org/10.1109/CVPR52688.2022.00024
  • Smith et al. [2023] Smith, J.S., Karlinsky, L., Gutta, V., Cascante-Bonilla, P., Kim, D., Arbelle, A., Panda, R., Feris, R., Kira, Z.: Coda-prompt: Continual decomposed attention-based prompting for rehearsal-free continual learning. In: 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, ??? (2023). https://doi.org/10.1109/cvpr52729.2023.01146 . http://dx.doi.org/10.1109/CVPR52729.2023.01146
  • Goswami et al. [2023] Goswami, D., Liu, Y., Twardowski, B., Weijer, J.: Fecam: Exploiting the heterogeneity of class distributions in exemplar-free continual learning. In: Advances in Neural Information Processing Systems (NeurIPS) (2023)
  • Janson et al. [2022] Janson, P., Zhang, W., Aljundi, R., Elhoseiny, M.: A simple baseline that questions the use of pretrained-models in continual learning. In: NeurIPS 2022 Workshop on Distribution Shifts: Connecting Methods and Applications (2022). https://openreview.net/forum?id=dnVNYctP3S
  • Petit et al. [2024] Petit, G., Soumm, M., Feillet, E., Popescu, A., Delezoide, B., Picard, D., Hudelot, C.: An analysis of initial training strategies for exemplar-free class-incremental learning. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 1837–1847 (2024)
  • Besnier et al. [2020] Besnier, V., Jain, H., Bursuc, A., Cord, M., Pérez, P.: This dataset does not exist: training models from generated images. In: ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1–5 (2020). IEEE
  • Jahanian et al. [2022] Jahanian, A., Puig, X., Tian, Y., Isola, P.: Generative models as a data source for multiview representation learning. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=qhAeZjs7dCL
  • Sohl-Dickstein et al. [2015] Sohl-Dickstein, J., Weiss, E.A., Maheswaranathan, N., Ganguli, S.: Deep Unsupervised Learning using Nonequilibrium Thermodynamics (2015)
  • Ho et al. [2020] Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Advances in neural information processing systems 33, 6840–6851 (2020)
  • Saharia et al. [2022] Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E.L., Ghasemipour, K., Gontijo Lopes, R., Karagol Ayan, B., Salimans, T., et al.: Photorealistic text-to-image diffusion models with deep language understanding. Advances in neural information processing systems 35, 36479–36494 (2022)
  • Rombach et al. [2022] Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10684–10695 (2022)
  • Ramesh et al. [2022] Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., Chen, M.: Hierarchical Text-Conditional Image Generation with CLIP Latents (2022)
  • Sarıyıldız et al. [2023] Sarıyıldız, M.B., Alahari, K., Larlus, D., Kalantidis, Y.: Fake it till you make it: Learning transferable representations from synthetic imagenet clones. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8011–8021 (2023)
  • Zhang et al. [2024] Zhang, Y., Zhou, D., Hooi, B., Wang, K., Feng, J.: Expanding small-scale datasets with guided imagination. Advances in Neural Information Processing Systems 36 (2024)
  • He et al. [2022] He, R., Sun, S., Yu, X., Xue, C., Zhang, W., Torr, P., Bai, S., Qi, X.: Is synthetic data from generative models ready for image recognition? arXiv preprint arXiv:2210.07574 (2022)
  • Shin et al. [2023] Shin, J., Kang, M., Park, J.: Fill-Up: Balancing Long-Tailed Data with Generative Models. arXiv preprint arXiv:2306.07200 (2023)
  • Jodelet et al. [2023] Jodelet, Q., Liu, X., Phua, Y.J., Murata, T.: Class-incremental learning using diffusion model for distillation and replay. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, pp. 3425–3433 (2023)
  • Duan et al. [2023] Duan, R., Liu, Y., Chen, J., Kortylewski, A., Yuille, A.: Prompt-Based Exemplar Super-Compression and Regeneration for Class-Incremental Learning (2023)
  • Du et al. [2024] Du, X., Sun, Y., Zhu, J., Li, Y.: Dream the impossible: Outlier imagination with diffusion models. Advances in Neural Information Processing Systems 36 (2024)
  • Nguyen et al. [2024] Nguyen, Q., Vu, T., Tran, A., Nguyen, K.: Dataset diffusion: Diffusion-based synthetic data generation for pixel-level semantic segmentation. Advances in Neural Information Processing Systems 36 (2024)
  • Tian et al. [2023] Tian, Y., Fan, L., Isola, P., Chang, H., Krishnan, D.: StableRep: Synthetic Images from Text-to-Image Models Make Strong Visual Representation Learners (2023)
  • Lee et al. [2019] Lee, K., Lee, K., Shin, J., Lee, H.: Overcoming catastrophic forgetting with unlabeled data in the wild. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 312–321 (2019)
  • Zhang et al. [2020] Zhang, J., Zhang, J., Ghosh, S., Li, D., Tasci, S., Heck, L., Zhang, H., Kuo, C.-C.J.: Class-incremental learning via deep model consolidation. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 1131–1140 (2020)
  • Lechat et al. [2021] Lechat, A., Herbin, S., Jurie, F.: Pseudo-labeling for class incremental learning. In: BMVC 2021: The British Machine Vision Conference (2021)
  • Bellitto et al. [2022] Bellitto, G., Pennisi, M., Palazzo, S., Bonicelli, L., Boschini, M., Calderara, S.: Effects of auxiliary knowledge on continual learning. In: 2022 26th International Conference on Pattern Recognition (ICPR), pp. 1357–1363 (2022). IEEE
  • Liu et al. [2024] Liu, Y., Li, Y., Schiele, B., Sun, Q.: Wakening past concepts without past data: Class-incremental learning from online placebos. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 2226–2235 (2024)
  • Chen et al. [2023] Chen, J., Wang, Y., Wang, P., Chen, X., Zhang, Z., Lei, Z., Li, Q.: DiffusePast: Diffusion-based Generative Replay for Class Incremental Semantic Segmentation (2023)
  • Kim et al. [2024] Kim, J., Cho, H., Kim, J., Tiruneh, Y.Y., Baek, S.: SDDGR: Stable Diffusion-based Deep Generative Replay for Class Incremental Object Detection (2024)
  • Douillard et al. [2021] Douillard, A., Valle, E., Ollion, C., Robert, T., Cord, M.: Insights from the future for continual learning. In: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE, ??? (2021). https://doi.org/10.1109/cvprw53098.2021.00387 . http://dx.doi.org/10.1109/CVPRW53098.2021.00387
  • Pernici et al. [2021] Pernici, F., Bruni, M., Baecchi, C., Turchini, F., Del Bimbo, A.: Class-incremental learning with pre-allocated fixed classifiers. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 6259–6266 (2021). IEEE
  • Zhou et al. [2022] Zhou, D.-W., Wang, F.-Y., Ye, H.-J., Ma, L., Pu, S., Zhan, D.-C.: Forward compatible few-shot class-incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9046–9056 (2022)
  • Brown et al. [2020] Brown, T.B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D.M., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., McCandlish, S., Radford, A., Sutskever, I., Amodei, D.: Language Models are Few-Shot Learners (2020)
  • Miller [1995] Miller, G.A.: Wordnet: A lexical database for english. Commun. ACM 38(11), 39–41 (1995) https://doi.org/10.1145/219717.219748
  • Ho and Salimans [2021] Ho, J., Salimans, T.: Classifier-free diffusion guidance. In: NeurIPS 2021 Workshop on Deep Generative Models and Downstream Applications (2021). https://openreview.net/forum?id=qw8AKxfYbI
  • Krizhevsky et al. [2009] Krizhevsky, A., et al.: Learning multiple layers of features from tiny images (2009)
  • Russakovsky et al. [2015] Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al.: Imagenet large scale visual recognition challenge. International journal of computer vision 115(3), 211–252 (2015)
  • Cubuk et al. [2019] Cubuk, E.D., Zoph, B., Mane, D., Vasudevan, V., Le, Q.V.: Autoaugment: Learning augmentation strategies from data. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019)
  • Paszke et al. [2019] Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., Chintala, S.: PyTorch: An Imperative Style, High-Performance Deep Learning Library. In: Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems 32, pp. 8024–8035. Curran Associates, Inc., ??? (2019). http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf
  • Zhou et al. [2023] Zhou, D.-W., Wang, F.-Y., Ye, H.-J., Zhan, D.-C.: Pycil: a python toolbox for class-incremental learning. SCIENCE CHINA Information Sciences 66(9), 197101 (2023) https://doi.org/10.1007/s11432-022-3600-y
  • Torralba et al. [2008] Torralba, A., Fergus, R., Freeman, W.T.: 80 million tiny images: A large data set for nonparametric object and scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 30(11), 1958–1970 (2008) https://doi.org/10.1109/TPAMI.2008.128
  • Meng et al. [2024] Meng, Z., Zhang, J., Yang, C., Zhan, Z., Zhao, P., WAng, Y.: Diffclass: Diffusion-based class incremental learning. arXiv preprint arXiv:2403.05016 (2024)