This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Covariance-based Space Regularization for Few-shot Class Incremental Learning

Yijie Hu1, Guanyu Yang1, Zhaorui Tan1, Xiaowei Huang2, Kaizhu Huang3,Qiu-Feng Wang1
Xi’an Jiaotong-Liverpool University1, Liverpool University2, Duke Kunshan University3
{Yijie.Hu20, Zhaorui.Tan21}@student.xjtlu.edu.cn, {Guanyu.Yang02, Qiufeng.Wang}@xjtlu.edu.cn
[email protected], [email protected]
Abstract

Few-shot Class Incremental Learning (FSCIL) presents a challenging yet realistic scenario, which requires the model to continually learn new classes with limited labeled data (i.e., incremental sessions) while retaining knowledge of previously learned base classes (i.e., base sessions). Due to the limited data in incremental sessions, models are prone to overfitting new classes and suffering catastrophic forgetting of base classes. To tackle these issues, recent advancements resort to prototype-based approaches to constrain the base class distribution and learn discriminative representations of new classes. Despite the progress, the limited data issue still induces ill-divided feature space, leading the model to confuse the new class with old classes or fail to facilitate good separation among new classes. In this paper, we aim to mitigate these issues by directly constraining the span of each class distribution from a covariance perspective. In detail, we propose a simple yet effective covariance constraint loss to force the model to learn each class distribution with the same covariance matrix. In addition, we propose a perturbation approach to perturb the few-shot training samples in the feature space, which encourages the samples to be away from the weighted distribution of other classes. Regarding perturbed samples as new class data, the classifier is forced to establish explicit boundaries between each new class and the existing ones. Our approach is easy to integrate into existing FSCIL approaches to boost performance. Experiments on three benchmarks validate the effectiveness of our approach, achieving a new state-of-the-art performance of FSCIL.

1 Introduction

Provided with a substantial amount of stationary data, recent advancements in deep learning have enabled neural networks to excel in classification tasks [17, 9, 35, 30, 31]. However, it is not feasible to directly deploy neural networks into the open world, where the data may emerge in a non-stationary way, such as recognizing new types of diseases or new vehicles in autonomous driving. Simply retraining or fine-tuning the model with new data introduces the well-known catastrophic forgetting problem [24, 7]. To address this challenge, Class Incremental Learning (CIL) has been extensively researched to broaden the application scope of neural networks [52, 33]. By simulating the scenario where disjoint new data appears in incremental sessions, CIL aims to learn new concepts without forgetting old knowledge.

The conventional CIL setting assumes the amount of new data is usually sufficient, which may not be realistic as labeling new data can be expensive. To handle this issue, few-shot class incremental learning (FSCIL) has attracted much attention recently, where only a few training samples from new classes are available during incremental sessions [32, 18, 47].

Refer to caption
Figure 1: (a) Prototype-based models demonstrate compact representations of old classes, conserving space for new classes, though risking confusion due to mixed class distributions. (b) The proposed approach aims to regularize each class distribution within a fixed span by constraining covariance and to enhance class separation through the learning of perturbed new class data.

The framework of FSCIL typically involves two key stages. Initially, the model is trained on a base dataset, where all classes (referred to as base classes or old classes) contain sufficient instances. Subsequently, the model engages in incremental sessions, where it is required to learn new classes with limited samples in each session without access to previous data. After training, the model is evaluated on all previously encountered classes. Due to the scarcity of new data, the model is more vulnerable to overfit new classes and hence suffers catastrophic forgetting of old classes during incremental learning sessions [26, 20, 52, 23].

To overcome both overfitting and catastrophic forgetting issues of FSCIL, recent wisdom resorts to prototype-based models [18, 54, 10, 44]. Specifically, prototype-based models [28, 42] replace the linear classifier with learnable prototypes, aiming to learn the most representative point (e.g., center point) of each class. Recent FSCIL works involve two stages, i.e., base session learning and incremental session learning. In the base learning stage, prototype-based models are utilized to learn compact representations of base classes [25, 10, 44]. During the incremental sessions, the feature extractor is frozen, and new prototypes are computed using the mean of each new class. The frozen feature extractor can effectively alleviate the catastrophic forgetting problem, and the prototype classifier can mitigate the overfitting problem. However, as shown in Fig. 1(a), in the fixed feature space, the estimated new class prototypes may lie very close to base class distributions due to the scarcity of the new data. Thus, class distributions can be ill-located during incremental sessions, i.e., new class distribution may lie close to old classes or mixed with other new class distributions [44, 38]. This dilemma may then lead the model to confuse the class distributions in the feature space. Besides, the computed class prototypes can be affected by data noise, exacerbating the confusion dilemma.

To resolve this dilemma, we argue that each class should occupy the same amount of feature space. Motivated by this, we dynamically relocate each new class during incremental sessions to prevent overlapping. Without loss of generality, we assume each class follows the Gaussian distribution, where the mean controls the location and the covariance controls the scope of each feature distribution. To ensure each class takes up the same amount of feature space, we first regularize the covariance of each class in the base session. However, the direct optimization of the distribution-related statistics during training is often inefficient, as it requires model inferences throughout the entire dataset. Inspired by previous works in variational inference [16, 15, 48], we adopt a similar approach to estimate and regularize the class distributions efficiently. Specifically, we derive an evidence lower bound for classification and distribution learning. Maximizing the lower bound equals to maximize the confidence of prediction and minimize the KL divergence between the estimated distribution and the prior distribution. We set the fixed covariance of the prior distribution to be equal to all classes, which formulates the KL divergence to a covariance constraint loss, encouraging each class distribution to have the same covariance during training. The covariance constraint loss serves as an upper bound of KL divergence, which leads to a tighter evidence lower bound.

Furthermore, to reallocate the feature space for the few-shot new classes during incremental sessions, we propose a perturbation approach to expand the distributions of few-shot new classes by generating perturbed samples for new classes, then pushing these perturbed samples away from semantically similar classes. In detail, we introduce a learnable prior distribution for each few-shot sample based on semantic similarity. We first obtain each training sample’s softmax score towards other classes, then we use the weighted mean and the fixed covariance as a prior distribution. We adopt one linear layer to estimate the mean and variance of the new class distribution. The estimated distribution is supervised by minimizing the KL divergence between the estimated distribution and the prior. Next, we multiply the predicted variance with the extracted features and add the predicted mean to create perturbed samples, as shown in Fig. 1(b). The perturbed samples are treated as new training samples, learning which pushes new classes away from the overlapping distributions. In this manner, the new class distribution is expanded by assigning the fixed variance. Learning semantic-guided perturbed samples also facilitates better separation between classes.

Our approach is easy to implement and can be integrated with other approaches. We choose several recent state-of-the-art FSCIL models as baselines [49, 29] and apply the proposed approach to these methods. Extensive experiments on three benchmarks validate the effectiveness of our approach. Our main contributions can be summarized as follows: (1) We propose a covariance constraint loss (CCL) to regularize the class distributions, constraining the span of each distribution within the same range.(2) We propose a semantic guided perturbation approach to perturb the few-shot new data, aiming to learn extensive and discriminative new class distributions. (3) Our proposed method is easy to incorporate into current FSCIL models to boost their performance. Experimental results on the FSCIL benchmark datasets validate the effectiveness of our approach.

2 Related Works

Few-shot Class Incremental Learning (FSCIL) involves the techniques of both class incremental learning and few-shot learning.

2.1 Class Incremental Learning

Class Incremental Learning (CIL) aims to learn new classes from a sequence of classification tasks without access to previously encountered data. The primary objective of CIL is to effectively learn new classes while minimizing the forgetting of old classes. Recent research in CIL can be broadly categorized into three approaches. The first and most straightforward method is to retain old data or knowledge during the learning process. Recent works [26, 12, 27, 52, 50] propose to mitigate the forgetting issue by rehearsing and generating previously retrained class data. Another common approach involves identifying key model parameters associated with previously learned classes and dynamically updating only the remaining parameters during incremental sessions [14, 41, 45, 20]. The third category focuses on addressing the bias inherent in CIL methods, which tend to favor the most recently learned classes [40, 11, 2].

2.2 Few-Shot Learning

Few-Shot Learning (FSL) aims to develop a classification model with very limited data. To generalize on few-shot classes, metric-based methods focus on learning a similarity metric that can effectively distinguish between classes with minimal examples [37, 28, 34, 46]. Hallucination-based approaches utilize data augmentation techniques, such as geometric transformations, style transfer and statistical augmentations, to increase the amount of training data [39, 43, 3, 8].

2.3 Few-Shot Class Incremental Learning

Few-Shot Class Incremental Learning (FSCIL) aims to learn new classes with limited incoming data in an incremental manner [32, 47, 51, 49, 44]. TOPIC [32] first introduces this setting and employs a neural gas algorithm to preserve the topology in the embedding space. To address the challenge of limited data in incremental sessions, prototype learning [28, 42, 19] has been widely adopted in FSCIL to enhance the model’s generalization to new or unseen data. To mitigate the catastrophic forgetting of old classes, recent studies [47, 10, 44] propose freezing the feature extraction backbone after training the base sessions and computing new class prototypes during incremental sessions. To learn more representative prototypes, Zhu et al. [53] introduces a self-promoted prototype refinement mechanism to develop extensible feature representations in the base session. LDC [21] utilizes a recurrent calibration module to learn new prototypes from sampled data, though this can be inefficient due to its recurrent nature. Unlike previous methods, our approach boosts FSCIL by regularizing the feature space for the few-shot new classes.

3 Methodology

In this section, we first give the problem formulation of FSCIL in Sec. 3.1. We then describe our proposed FSCIL method in detail via two sections, i.e., Covariance Constraint Loss (CCL) in Sec. 3.3 and Semantic Perturbation Learning (SPL) in Sec. 3.4, respectively.

3.1 Preliminaries

3.1.1 Problem Formulation

FSCIL aims to train a classification model with TT sequential sessions {𝒟0,𝒟1,,𝒟T}\{\mathcal{D}^{0},\mathcal{D}^{1},\ldots,\mathcal{D}^{T}\}, where 𝒟t={(𝒙it,yit)}i=1|𝒟t|\mathcal{D}^{t}=\{(\boldsymbol{x}_{i}^{t},y_{i}^{t})\}_{i=1}^{|\mathcal{D}^{t}|} is the training dataset at the tt-th session. 𝒙it\boldsymbol{x}_{i}^{t} is the ii-th input sample and its label yit𝒞ty_{i}^{t}\in\mathcal{C}^{t}. The label space 𝒞t\mathcal{C}^{t} of dataset 𝒟t\mathcal{D}^{t} is disjoint between different sessions, i.e., t1t2,𝒞t1𝒞t2=\forall t_{1}\neq t_{2},\mathcal{C}^{t_{1}}\cap\mathcal{C}^{t_{2}}=\emptyset. The first session 𝒟0\mathcal{D}^{0} is called the base session, which usually contains a sufficient amount of training samples for each old class c𝒞0c\in\mathcal{C}^{0}. In the next incremental session 𝒟t\mathcal{D}^{t}, there are NN new classes with KK training samples (usually 1 or 5 samples) in each class, formulating a NN way KK shot problem, i.e. |𝒟t|=NK|\mathcal{D}^{t}|=N\cdot K. In the tt-th session, previous datasets {𝒟0,𝒟1,,𝒟t1}\{\mathcal{D}^{0},\mathcal{D}^{1},\ldots,\mathcal{D}^{t-1}\} are not available, the model can only access the data in 𝒟t\mathcal{D}^{t}. After training in session tt, the model is evaluated on all seen classes 𝒞~t=𝒞0𝒞1𝒞t\tilde{\mathcal{C}}^{t}=\mathcal{C}^{0}\cup\mathcal{C}^{1}\ldots\cup\mathcal{C}^{t}.

3.1.2 Prototype-based Model

Researchers [28, 54, 49, 29] commonly adopt the prototype-based framework in FSCIL, where classifier weights are treated as prototypes. During the base session, given the feature extractor fψ:𝒙xdf_{\psi}:\boldsymbol{x}\rightarrow x\in\mathbb{R}^{d} and the classifier 𝑾0={𝒘10,𝒘20𝒘|C0|0}\boldsymbol{W}^{0}=\{\boldsymbol{w}_{1}^{0},\boldsymbol{w}_{2}^{0}\cdots\boldsymbol{w}_{|C^{0}|}^{0}\}, the model is trained using the cross-entropy loss:

ce(𝒙i0,yi0)=𝔼(𝒙i0,yi0)𝒟01Nc=1|C0|yi0logp(y=cxi0),\mathcal{L}_{ce}(\boldsymbol{x}_{i}^{0},y_{i}^{0})=\mathbb{E}_{(\boldsymbol{x}_{i}^{0},y_{i}^{0})\sim\mathcal{D}^{0}}\frac{1}{N}\sum_{c=1}^{|C^{0}|}-y_{i}^{0}\log p(y=c\mid{x}_{i}^{0}), (1)

where xi0=fψ(𝒙i0){x}_{i}^{0}=f_{\psi}(\boldsymbol{x}_{i}^{0}) denotes the extracted feature, p(y=cxi0)p(y=c\mid{x}_{i}^{0}) computes the confidence score of each sample using softmax of the cosine similarities:

p(y=c|xi0)=exp(cos(𝒘c,xi0))c=1𝒞0exp(cos(𝒘c,xi0)).p(y=c|{x}_{i}^{0})=\frac{\exp\left(\cos({\boldsymbol{w}_{c}},{x}_{i}^{0})\right)}{\sum_{c=1}^{\mathcal{C}^{0}}\exp\left(\cos({\boldsymbol{w}_{c}},{x}_{i}^{0})\right)}. (2)

For each incremental session t>0t>0, the feature extractor fψf_{\psi} is frozen, and the classifier is updated by adding new class prototypes: 𝑾t={𝒘10,𝒘20𝒘|C0|0}{𝒘1t𝒘𝒞tt}\boldsymbol{W}^{t}=\{\boldsymbol{w}_{1}^{0},\boldsymbol{w}_{2}^{0}\cdots\boldsymbol{w}_{|C^{0}|}^{0}\}\bigcup\{\boldsymbol{w}_{1}^{t}\ldots\boldsymbol{w}_{\mathcal{C}^{t}}^{t}\}, where each new class prototype is computed by averaging samples from its corresponding class:

𝒘ct=1|𝒟t|i=1|𝒟t|fψ(𝒙it).\boldsymbol{w}_{c}^{t}=\frac{1}{|\mathcal{D}^{t}|}\sum_{i=1}^{|\mathcal{D}^{t}|}{f_{\psi}(\boldsymbol{x}_{i}^{t})}. (3)

3.2 Overview of the Framework

Our framework follows the two-stage learning procedure in most of recent works [18, 10, 44], which begins with pretraining the model using the base classes. The pretraining stage is called the base session, where we integrate the proposed covariance constraint loss (CCL) to constrain the learned representations of base classes, as shown in Fig. 2(a). During the incremental sessions, we propose a semantic perturbation learning (SPL) approach, aiming to enlarge the separation between old and new classes by learning the perturbed new data, which is shown in Fig. 2(b).

Refer to caption
Figure 2: (a) Base session training. We deploy the covariance constraint loss to learn class distributions with a fixed span. (b) Semantic perturbation learning for the incremental stages. New data samples are perturbed by multiplying with the predicted distribution. The perturbed samples are trained along with the original data to establish the separation between classes.

3.3 Covariance Constraint Loss

Recent wisdom [55, 5, 29] has demonstrated a good pretrained model benefits the learning of incoming few-shot new classes. The common attempt in recent works is to learn very compact base class representations by pushing data to the learned prototypes via fantasizing new classes or metric-based classification losses [18, 49, 29]. However, simply pushing the data close to the class centroid does not explicitly constrain the span of each class distribution, which may lead to representation confusion when learning new classes (Fig. 1(a)). Hence, we apply a covariance constraint to each base class during training, ensuring distinct means but identical covariance across classes, as shown in Fig. 2(a).

In order to constrain the covariance of each distribution, it is vital to estimate base class distributions p(x,y)p(x,y). However, it is generally computational intractable. To solve this issue, previous works [16, 15, 6] adopt the variational inference, which involves a parametric posterior function qϕ(z|x,y)q_{\phi}({z}|x,y) to approximate the true posterior by maximizing the evidence lower bound (ELBO):

logpθ(x,y)\displaystyle\log p_{\theta}(x,y) (4)
=logpθ(yx)+logpθ(z)+logpθ(xz)pθ(zx)\displaystyle=\log p_{\theta}(y\mid x)+\log p_{\theta}(z)+\log\frac{p_{\theta}(x\mid z)}{p_{\theta}(z\mid x)}
=qϕ(zx,y){logpθ(yx)logqϕ(zx,y)pθ(z)+\displaystyle=\int q_{\phi}(z\mid x,y)\{\log p_{\theta}(y\mid x)-\log\frac{q_{\phi}({z}\mid{x,y})}{p_{\theta}(z)}+
logqϕ(zx,y)pθ(xz)pθ(zx)}dz\displaystyle\log\frac{q_{\phi}(z\mid x,y)p_{\theta}(x\mid z)}{p_{\theta}(z\mid x)}\}dz
𝔼qϕ(zx,y)[logpθ(yx)]DKL[qϕ(zx,y)pθ(z)],\displaystyle\geq\mathbb{E}_{q_{\boldsymbol{\phi}}\left({z}\mid{x,y}\right)}\left[\log p_{\theta}\left({y}\mid{x}\right)\right]-D_{KL}\left[q_{\boldsymbol{\phi}}\left({z}\mid{x,y}\right)\|p_{\theta}({z})\right],

where ϕ\phi and θ\theta are modeled by neural networks. The first term in Eq. 4 aims to maximize the likelihood by improving the confidence of the prediction. When optimizing the first term, the optimizing process can be written as

argmaxqϕ(zx,y)xyp(x,y)zqϕ(zx,y)logpθ(yx)\displaystyle\underset{q_{\phi}(z\mid x,y)}{\arg\max}\int_{x}\sum_{y}p(x,y)\int_{z}q_{\phi}(z\mid x,y)\log p_{\theta}(y\mid x) (5)
=\displaystyle= argmaxqϕ(zx,y)xp(x)zqϕ(zx,y)yp(yx)logpθ(yx).\displaystyle\underset{q_{\phi}(z\mid x,y)}{\arg\max}\int_{x}p(x)\int_{z}q_{\phi}(z\mid x,y)\sum_{y}p(y\mid x)\log p_{\theta}(y\mid x).

As qϕ(zx,y)q_{\phi}({z}\mid{x,y}) is a probability distribution, the integral is upper bounded by maxyp(yx)logpθ(yx)\max\sum_{y}p(y\mid x)\log p_{\theta}(y\mid x), which is equivalent to minimizing the cross entropy loss in Eq. 1. The second KL divergence minimizes the divergence between the variational posterior distribution qϕ(zx)q_{\boldsymbol{\phi}({z\mid x})} and the prior distribution pθ(z)p_{\theta}{(z)}. The third term actually can be calculated by DKL[qϕ(zx,y)pθ(zx)/pθ(xz)]D_{KL}\left[q_{\boldsymbol{\phi}}\left({z}\mid{x,y}\right)\|p_{\theta}({z\mid x})/p_{\theta}({x\mid z})\right], which is a non-negative value and we eliminate this item. In order to explicitly contain the covariance of all distributions, the remaining problem lies in how to formulate the posterior distribution qϕ(zx)q_{\boldsymbol{\phi}({z\mid x})} and the prior distribution pθ(z)p_{\theta}(z).

We first assume that zz satisfies multivariate Gaussian with a mean and diagonal covariance. The pθ(z)p_{\theta}(z) can be formulated as 𝒩(μc,𝐈)\mathcal{N}(\mu_{c},\mathbf{I}) with mean μ𝐜\mathbf{\mu_{c}} as the class center and fixed eye matrix 𝐈\mathbf{I}, respectively. As we aim to constrain the covariance, we adopt the fixed covariance as the prior. In order to learn the latent distribution quickly from a single instance during training, we adopt linear layers Pμ()P_{\mu}(\cdot) and Pσ()P_{\sigma}(\cdot) after the feature extraction network to form a statics prediction pipeline to estimate qϕ(z|x,y)q_{\phi}(z|x,y) following previous works [16]

qϕ(z|x,y)\displaystyle q_{\phi}(z|x,y) =𝒩(z|μc^,σc^2)\displaystyle=\mathcal{N}(z|\hat{\mu_{c}},\hat{\sigma_{c}}^{2}) (6)
=[Pμ(x),Pσ(x)]y=c.\displaystyle=[P_{\mu}(x),P_{\sigma}(x)]_{y=c}.

After obtaining the estimated statistical information, we can then formulate the second KL divergence term into the covariance constraint loss by

DKL[qϕ(zx,y)pθ(z)]\displaystyle-D_{KL}\left[q_{\boldsymbol{\phi}}\left({z}\mid{x,y}\right)\|p_{\theta}({z})\right] (7)
=qϕ(z|x,y)logpθ(z)qϕ(z|x,y)dz\displaystyle=\int q_{\phi}({z}|{x,y})\log\frac{p_{\theta}({z})}{q_{\phi}({z}|{x,y})}d{z}
=𝒩(z;μ^,σ^2)log𝒩(z;μc,𝐈)𝒩(z;μc^,σc^2)dz\displaystyle=\int\mathcal{N}\left({z};\hat{\mu},\hat{\sigma}^{2}\right)\log\frac{\mathcal{N}({z};{\mu_{c}},\mathbf{I})}{\mathcal{N}({z};\hat{\mu_{c}},\hat{\sigma_{c}}^{2})}d{z}
=12c=1𝒞0i=1d(1+logσ^c,i2σ^c,i2(μc^μc)2)\displaystyle=\frac{1}{2}\sum_{c=1}^{\mathcal{C}^{0}}\sum_{i=1}^{d}\left(1+\log\hat{\sigma}_{c,i}^{2}-\hat{\sigma}_{c,i}^{2}-(\hat{\mu_{c}}-\mu_{c})^{2}\right)
12c=1𝒞0i=1d(1+logσ^c,i2σ^c,i2),\displaystyle\leq\frac{1}{2}\sum_{c=1}^{\mathcal{C}^{0}}\sum_{i=1}^{d}\left(1+\log\hat{\sigma}_{c,i}^{2}-\hat{\sigma}_{c,i}^{2}\right),

where dd represents the feature dimension. We drop the final term to formulate the covariance constraint loss as ccl=12c=1𝒞0i=1d(1+logσ^c,i2σ^c,i2)\mathcal{L}_{ccl}=\frac{1}{2}\sum_{c=1}^{\mathcal{C}^{0}}\sum_{i=1}^{d}\left(1+\log\hat{\sigma}_{c,i}^{2}-\hat{\sigma}_{c,i}^{2}\right), which constrains the covariance of each feature distribution. Eq. 7 shows that the covariance constraint loss can be viewed as the upper bound of DKL-D_{KL}. Thus, by deploying covariance constraint loss, we are able to obtain a tighter upper bound than Eq. 4, which can be easier to optimize. The overall learning objective for the base session training is

Base=ce+γccl,\mathcal{L}_{Base}=\mathcal{L}_{ce}+\gamma\mathcal{L}_{ccl}, (8)

where γ\gamma is a positive hyperparameter that controls the degree of constraint.

3.4 Semantic Perturbation Learning

In order to prevent the model from overfitting the few-shot new classes during incremental sessions, recent approaches [18, 10, 29] attempt to freeze the feature extractor and learn the new class prototypes by averaging the few-shot samples. Though the overfitting problem is mitigated by fixing the previously learned feature space, the new class distributions can only account for small parts of the feature space compared to base classes due to the scarcity of data. Under this ill-divided space dilemma, as shown in Fig. 1(b), new class prototypes may lie very close to base class distributions or new class prototypes, which may further lead the model to misclassify new classes.

In this paper, we propose to address this issue from two perspectives. Firstly, we intend to expand the new class feature distributions by assigning the fixed covariance, which is the same as the base distributions and allows the new class distributions to take up more feature space. Secondly, to facilitate better separation among classes, we push the few-shot samples to be away from semantically similar distributions and then retrain the classifier to distinguish the perturbed new class samples from other classes. We achieve these two goals by forming the semantic perturbation learning framework, where we aim to learn a perturbation distribution from which any perturbations can perturb new data to be away from those close distributions but within a fixed range. The perturbation distribution can also be learned by formulating a similar variational inference in Eq. 4 but with a different prior distribution. Specifically, we form the prior distribution as the Gaussian distribution pθ(z~)p_{\theta}({\tilde{z}}) as 𝒩(μ~,𝐈)\mathcal{N}(\tilde{\mu},\mathbf{I}), where μ~\tilde{\mu} is the linear combination of other class prototypes. The weight is calculated by the similarity score over classes. For each incremental session, we first initialize the classifier by computing the new class prototypes by averaging the few-shot samples. Given new data sample 𝒙it\boldsymbol{x}_{i}^{t} with label 𝒚it\boldsymbol{y}_{i}^{t}, we compute the similarity score over other classes:

Si,c=𝕀cyitexp(cos(𝒘c,xit))c=1|𝒞t|1exp(cos(𝒘c,xit)),S_{i,c}=\frac{\mathbb{I}_{c\neq{y}_{i}^{t}}\exp\left(cos({\boldsymbol{w}_{c}},{x}_{i}^{t})\right)}{\sum_{c=1}^{|\mathcal{C}^{t}|-1}\exp\left(cos({\boldsymbol{w}_{c}},{x}_{i}^{t})\right)}, (9)

where SK×(|𝒞t|1)S\in\mathbb{R}^{K\times(|\mathcal{C}^{t}|-1)} and 𝕀()\mathbb{I}(\cdot) is an indicator function. For each sample, we can get the prior mean by multiplying the confidence score with other class prototypes

μ~=cyit|𝒞t|1Si,c𝒘c.\tilde{\mu}=\sum_{c\neq{y}_{i}^{t}}^{|\mathcal{C}^{t}|-1}S_{i,c}\boldsymbol{w}_{c}. (10)

We also adopt a linear layer to predict the mean and covariance of each sample, the same as Eq. 6. To be noted, we do not reuse the linear layer in the base session so that this method can be directly applied to other methods in a plug-and-play manner. After predicting the statistics, the overall objective during the incremental training is:

incre\displaystyle\mathcal{L}_{incre} =ce(xit,yit)+ce(μ^+σ^xit,yit)\displaystyle=\mathcal{L}_{ce}({x}_{i}^{t},y_{i}^{t})+\mathcal{L}_{ce}(\hat{\mu}+\hat{\sigma}\odot{x}_{i}^{t},y_{i}^{t}) (11)
αDKL[qϕ(z~xit)pθ(z~xit)],\displaystyle-\alpha D_{KL}\left[q_{\boldsymbol{\phi}}\left({\tilde{z}}\mid{x}_{i}^{t}\right)\|p_{\theta}(\tilde{z}\mid{x}_{i}^{t})\right],

where \odot represents the element-wise multiplication. The first and second term is the classification loss for the few-shot data and the perturbed few-shot data. The third term aims to learn the perturbation distribution by minimizing the KL divergence between the predicted feature distribution of each sample and the prior distribution. The positive hyperparameter α\alpha controls the strength of the perturbation. The KL term is formulated as:

DKL[qϕ(z~xit)pθ(z~xit)]\displaystyle-D_{KL}\left[q_{\boldsymbol{\phi}}\left({\tilde{z}}\mid{x}_{i}^{t}\right)\|p_{\theta}(\tilde{z}\mid{x}_{i}^{t})\right] (12)
=12i=1d(1+logσ^i2σ^i2(μi^μi~)2).\displaystyle=\frac{1}{2}\sum_{i=1}^{d}\left(1+\log\hat{\sigma}_{i}^{2}-\hat{\sigma}_{i}^{2}-(\hat{\mu_{i}}-\tilde{\mu_{i}})^{2}\right).

4 Experiments

4.1 Implementation Details

Datasets and Experimental Settings. We conduct extensive experiments on three benchmark datasets, including MiniImageNet, CIFAR100 and CUB200. MiniImageNet and CIFAR100 contain 100 classes in total, we set the number of base classes as 60. We set 8 incremental sessions, where each session formulates a 5-way 5-shot problem. For fine-grained data CUB200 that contains 200 classes, we set the number of base classes as 100, followed by 10 incremental sessions, and each session formulates a 10-way 5-shot problem. To make the comparison fair, we use the same base and incremental class data in each dataset when conducting the experiments following previous works [1, 51, 49, 29]. All experiments are conducted using one RTX3090 card111https://github.com/tambourine666/Covariance-Space-Regularize.

Baseline Models. We adopt three baseline models, i.e., the naive CE-trained model, fantasy-based model FACT [49], and state-of-the-art contrastive learning model SAVC [29]. Details of the baseline models can be found in Appendix A.

Model Architectures. Following [18, 51, 29, 38], we use ResNet-18 in experiments on CIFAR100 and MiniImageNet. We use ImageNet pre-trained ResNet-18 for the CUB200 dataset following [51, 29, 38]. We follow the same experimental setting as baseline methods for fair comparisons. The dimension dd is set as 64, 512, 512 on CIFAR100, MiniImageNet and CUB200, respectively, which is the same as baseline models. The hyperparameter analysis is shown in Sec. 4.5.2.

4.2 Benchmark Performance

Table 1: Incremental learning performance on MiniImageNet under 5-way 5-shot setup. “Avg Acc.” represents the average accuracy of all sessions. “Final Improv.” calculates the improvement of our method after learning in the final session. Bold represents best performance. * indicates that we reproduce the results using public open-source code.
Methods Accuracy in each session (%) \uparrow Avg Acc. Final Improv.
0 1 2 3 4 5 6 7 8
iCaRL [26] 61.31 46.32 42.94 37.63 30.49 24.00 20.89 18.80 17.21 33.29 +39.20
NCM [11] 61.31 47.80 39.30 31.90 25.70 21.40 18.70 17.20 14.17 30.83 +42.24
Data-free Replay [22] 71.84 67.12 63.21 59.77 57.01 53.95 51.55 49.52 48.21 58.02 +8.20
CEC [18] 72.00 66.83 62.97 59.43 56.70 53.73 51.19 49.24 47.63 57.75 +8.78
MetaFSCIL [5] 72.04 67.94 63.77 60.29 57.58 55.16 52.90 50.79 49.19 58.85 +7.22
C-FSCIL [10] 76.40 71.14 66.46 63.29 60.42 57.46 54.78 53.11 51.41 61.61 +5.00
LIMIT [51] 73.81 72.09 67.87 63.89 60.70 57.77 55.67 53.52 51.23 61.84 +5.18
CE 75.65 70.45 66.09 62.16 58.96 55.92 53.08 51.05 49.39 60.31 +7.02
CE-Ours 75.77 70.71 66.53 63.33 60.42 57.46 54.61 52.48 50.65 61.33 +5.76
FACT [49] 76.25 70.91 66.41 62.79 59.45 56.22 53.37 51.21 49.48 60.68 +6.93
FACT-Ours 75.65 70.63 66.79 63.20 59.72 56.68 53.78 51.53 50.01 60.89 +6.40
SAVC [29] 80.68 75.89 71.54 67.80 64.85 61.42 58.38 56.43 54.91 65.77 +1.50
SAVC-Ours 80.90 75.89 71.80 68.59 65.86 62.41 59.33 57.71 56.41 66.54
Refer to caption
Figure 3: (a) Comparison of harmonic performance after incremental sessions with baseline models on MiniImageNet. (b) Comparison of performance of new classes with baseline model in each incremental session on MiniImageNet

We conduct our experiments on three FSCIL benchmarks, i.e., CIFAR100, MiniImageNet, and CUB200, and compare our approach with the baseline models and other recent FSCIL methods. We also compare our method with classical CIL methods such as ICARL [26], NCM [11], and FSCIL methods CEC [18], C-FSCIL [10] and meta-learning based approaches [4, 51]. We report the numerical results of MiniImageNet in Tab. 1, and more results on CIFAR100 (Tab. A4) and CUB200 (Tab. A5) can be founded in Appendix C.

For baseline models, adding our proposed loss and semantic perturbation learning boosts the model performance for incremental sessions. After 8-session incremental learning, our method can improve the final model performance by at least 0.53% on MiniImageNet and 0.56% on CIFAR100, respectively. It is worth noticed that our method has a more profound effect on SAVC, which learns representations by pushing the data to the class center using contrastive learning. As SAVC does not consider the span of each class distribution explicitly, constraining the class distribution using our proposed methods boosts the base class performance and obtains better performance when learning new classes.

We further demonstrate the comparisons of harmonic performances of all classes after incremental sessions and comparisons of new class performance in each incremental session in  Fig. 3. As shown in Fig. 3(a), compared with baseline models, our approach boosts the harmonic performance by at least 9.80%. The CE model enjoys an improvement of 11.45% from our approach. The significant boost in harmonic performance results from the improvement of new classes. As illustrated in Fig. 3(b), the performance of new classes demonstrates a substantial improvement at each incremental session. Overall, our approach boosts the new class performance by at least 5.65%. Benefiting from the covariance constraint and distribution expansion strategy, we are able to learn more separable representations of new classes, boosting the performance of new classes.

4.3 Ablation Studies of Each Component

Table 2: Ablation results of proposed components. “Base” represents the base model performance. “Old” is the performance of base classes after incremental sessions. “New” is the performance of all new classes. “Avg” is the average performance of all incremental sessions. “PD” is the drop rate of the base class performance after incremental sessions. “H.” is the harmonic accuracy of base and new classes after incremental learning.
Method
Base
Session
Incremental
Session
CIFAR100
ccl\mathcal{L}_{ccl} SPL Base\uparrow Old\uparrow New\uparrow Avg\uparrow PD\downarrow H. \uparrow
CE 76.87 71.15 20.65 62.12 5.72 32.01
CE+SPL 76.87 69.32 23.38 62.20 7.55 34.96
CE+ccl+\mathcal{L}_{ccl} 78.26 72.67 21.90 63.58 5.59 33.66
CE+ccl+SPL+\mathcal{L}_{ccl}+SPL 78.26 71.42 23.62 63.66 6.84 35.51
FACT 78.38 71.08 21.73 62.17 5.40 32.54
FACT+SPL 78.38 71.07 21.82 62.20 7.31 33.40
FACT+ccl\mathcal{L}_{ccl} 79.12 72.17 22.50 62.80 4.73 34.31
FACT+ccl\mathcal{L}_{ccl}+SPL 79.12 71.20 24.05 62.82 5.70 35.96
SAVC 78.98 72.25 21.00 62.81 6.73 32.56
SAVC+SPL 78.98 72.20 21.25 62.88 6.78 32.83
SAVC+ccl\mathcal{L}_{ccl} 79.00 72.88 22.38 63.35 6.12 34.24
SAVC+ccl+SPL\mathcal{L}_{ccl}+SPL 79.00 72.90 23.50 63.48 6.10 35.51

We conduct the ablation experiments of the proposed CCL and SPL on the CIFAR100 dataset, as shown in Tab. 2. Following previous works [49, 29, 38], we verify the effectiveness of our approach by six metrics, i.e., the performance of the base model, the performance on the base classes after incremental sessions, the performance of all new classes, the average performance of all incremental sessions, the performance drop of the base classes (PD=AccuracybaseAccuracyoldPD=Accuracy_{base}-Accuracy_{old}), and the harmonic accuracy of base and new classes after the incremental learning. When we do not deploy the SPL for incremental sessions, the model adopts the prototype-based incremental learning approach mentioned in Sec. 3.1.2.

As shown in Tab. 2, by deploying our proposed covariance constraint loss LcclL_{ccl}, our model is able to achieve higher base class performance compared to the baseline model. By constraining the base class distributions, the drop in PD rate and the gain of new class performance on three baseline models validate that the covariance constraint loss allows the model to learn the new classes better while obtaining less forgetting of old classes.

We also conduct experiments of SPL on both baseline models and ccl\mathcal{L}_{ccl} trained models. The results demonstrate that directly deploying SPL can boost the performance of new classes and harmonic accuracy by at least 0.25% and 0.27%, respectively. It is worth noticing that simply using SPL on the baseline models leads to a higher PD. As we expand the distribution of new classes during incremental sessions without constraining the base class distributions, it may lead the model to confuse the old classes with new classes. If SPL is deployed on the ccl\mathcal{L}_{ccl} trained model, the model is able to achieve the highest performance on new classes and the highest harmonic accuracy, validating the effectiveness of the combination of proposed components. We also conduct experiments on comparing SPL with other incremental update methods in Appendix B.

4.4 T-SNE Visualizations of Proposed Approaches

We compare the different approaches by visualizing the learned feature embedding on the CIFAR100 dataset. In Fig. 4(a), we compare the base model trained with and without covariance constraint loss. We visualize the feature embedding of 8 different base classes, demonstrating that by adding the constraint on the covariance of the feature distribution, the learned feature distribution becomes more separable and compact, leaving more feature space for the incoming new classes. In Fig. 4(b), we demonstrate the effectiveness of SPL by showing the learned new class features before SPL, and new class features along with all perturbed data after SPL during incremental sessions (t>0)(t>0). As shown in the left side of Fig. 4(b), the new class features (colored triangles) tend to lie close with each other and mix with base classes due to the scarcity of the training data. By deploying SPL during the incremental sessions, our approach is able to push the few-shot training samples away from base classes and facilitate better separation among classes. The generated perturbed data in each epoch (denoted by “×\times” in the figure) expands the original small new class feature distribution, which allows the new class distributions to take more feature space, benefiting the new class prediction.

Refer to caption
Figure 4: Comparisons of T-SNE visualization of the learned feature embedding on CIFAR100. (a) Visualization of eight base classes’ feature embedding in the base session. We compare the base model with and without covariance constraint loss. (b) Visualizations of the feature embedding of two few-shot new classes together with their perturbed samples, and three base classes during incremental learning. We compare the baseline model and the model using the proposed SPL.

4.5 Further Analysis

4.5.1 Effects of Incremental Shots

We further conduct experiments on CUB200 to investigate the impact of varying the number of shots during incremental sessions, as shown in Fig. 5(a). Since our method relies on the data of each class to expand the new class distributions, we vary the number of shots per class to observe its effect in incremental sessions. Keeping the number of classes consistent with the current experimental setting, we vary the shot number KK from {1,3,5,10}\{1,3,5,10\} on CUB200. As depicted in the figure, increasing the number of available instances in each class leads to more accurate distributions and a corresponding improvement on performance. Even with a reduced shot number of 1, the model maintains stable performance during incremental sessions.

4.5.2 Hyperparameter Analysis

There are two hyperparameters in our approach, i.e., the γ\gamma controls the impact of covariance constraint loss in Eq. 8 and the α\alpha controls the perturbation in Eq. 11. We conduct experiments by changing γ\gamma and α\alpha from {0,0.0001,0.001,0.01,0.1}\{0,0.0001,0.001,0.01,0.1\}, and show the accuracy of the last session on CUB200 dataset in Fig. 5 (b). We can see that relatively small γ\gamma and α\alpha yield better performance, and the best performance is yield with γ\gamma and α\alpha set as 0.01.

Refer to caption
Figure 5: (a) Results of different incremental shots on CUB200. (b) Results of different hyperparameters on CUB200.

5 Conclusion

In this paper, we propose a covariance constraint loss and semantic perturbation learning to address the ill-divided feature space problem for few-shot class incremental learning (FSCIL). Our main motivation is to constrain the span of each distribution then reallocate the ill-divided feature space by reestablishing the decision boundaries between classes. Based on this, we attempt to learn the model through two steps: during the base session, we propose a covariant constraint loss (CCL) to explicitly constrain the feature distribution and facilitate better class separation. We derive the CCL from a variational inference framework, which estimates and constrains the feature distribution efficiently. For incremental sessions, we propose to generate semantic-guided perturbed data to aid the learning of few-shot new classes. The generated data expands the few-shot distributions and pushes the few-shot samples away from easily confusing classes. The proposed approach can be integrated into current FSCIL methods in a plug-and-play manner, which is easy to implement. We conduct comprehensive experiments on three benchmark datasets and apply our approach to three baseline models. Experimental results demonstrate the effectiveness of our methods and obtain a new state-of-the-art performance.

6 Acknowledgements

The work was partially supported by the following: National Natural Science Foundation of China under no. 92370119 and No. 62276258, and No.62376113.

References

  • [1] Aishwarya Agarwal, Biplab Banerjee, Fabio Cuzzolin, and Subhasis Chaudhuri. Semantics-driven generative replay for few-shot class incremental learning. In Proceedings of the 30th ACM International Conference on Multimedia, pages 5246–5254, 2022.
  • [2] Francisco M Castro, Manuel J Marín-Jiménez, Nicolás Guil, Cordelia Schmid, and Karteek Alahari. End-to-end incremental learning. In Proceedings of the European conference on computer vision (ECCV), pages 233–248, 2018.
  • [3] Xiaohua Chen, Yucan Zhou, Dayan Wu, Wanqian Zhang, Yu Zhou, Bo Li, and Weiping Wang. Imagine by reasoning: A reasoning-based implicit semantic data augmentation for long-tailed classification. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 356–364, 2022.
  • [4] Meng Cheng, Hanli Wang, and Yu Long. Meta-learning-based incremental few-shot object detection. IEEE Transactions on Circuits and Systems for Video Technology, 32(4):2158–2169, 2021.
  • [5] Zhixiang Chi, Li Gu, Huan Liu, Yang Wang, Yuanhao Yu, and Jin Tang. Metafscil: A meta-learning approach for few-shot class incremental learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 14166–14175, 2022.
  • [6] Shehzaad Dhuliawala, Mrinmaya Sachan, and Carl Allen. Variational classification. TMLR, 2023.
  • [7] Ian J Goodfellow, Mehdi Mirza, Da Xiao, Aaron Courville, and Yoshua Bengio. An empirical investigation of catastrophic forgetting in gradient-based neural networks. arXiv preprint arXiv:1312.6211, 2013.
  • [8] Dandan Guo, Long Tian, He Zhao, Mingyuan Zhou, and Hongyuan Zha. Adaptive distribution calibration for few-shot learning with hierarchical optimal transport. Advances in Neural Information Processing Systems, 35:6996–7010, 2022.
  • [9] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  • [10] Michael Hersche, Geethan Karunaratne, Giovanni Cherubini, Luca Benini, Abu Sebastian, and Abbas Rahimi. Constrained few-shot class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9057–9067, 2022.
  • [11] Saihui Hou, Xinyu Pan, Chen Change Loy, Zilei Wang, and Dahua Lin. Learning a unified classifier incrementally via rebalancing. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 831–839, 2019.
  • [12] Minsoo Kang, Jaeyoo Park, and Bohyung Han. Class-incremental learning by knowledge distillation with adaptive feature consolidation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16071–16080, 2022.
  • [13] Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. Supervised contrastive learning. Advances in neural information processing systems, 33:18661–18673, 2020.
  • [14] Do-Yeon Kim, Dong-Jun Han, Jun Seo, and Jaekyun Moon. Warping the space: Weight space rotation for class-incremental few-shot learning. In The Eleventh International Conference on Learning Representations, 2022.
  • [15] Durk P Kingma, Shakir Mohamed, Danilo Jimenez Rezende, and Max Welling. Semi-supervised learning with deep generative models. Advances in neural information processing systems, 27, 2014.
  • [16] Diederik P Kingma and Max Welling. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
  • [17] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25, 2012.
  • [18] Timothée Lesort, Vincenzo Lomonaco, Andrei Stoian, Davide Maltoni, David Filliat, and Natalia Díaz-Rodríguez. Continual learning for robotics: Definition, framework, learning strategies, opportunities and challenges. Information fusion, 58:52–68, 2020.
  • [19] Yushu Li, Xun Xu, Yongyi Su, and Kui Jia. On the robustness of open-world test-time training: Self-training with dynamic prototype expansion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 11836–11846, 2023.
  • [20] Zhizhong Li and Derek Hoiem. Learning without forgetting. IEEE transactions on pattern analysis and machine intelligence, 40(12):2935–2947, 2017.
  • [21] Binghao Liu, Boyu Yang, Lingxi Xie, Ren Wang, Qi Tian, and Qixiang Ye. Learnable distribution calibration for few-shot class-incremental learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  • [22] Huan Liu, Li Gu, Zhixiang Chi, Yang Wang, Yuanhao Yu, Jun Chen, and Jin Tang. Few-shot class-incremental learning via entropy-regularized data-free replay. In European Conference on Computer Vision, pages 146–162. Springer, 2022.
  • [23] Marc Masana, Xialei Liu, Bartłomiej Twardowski, Mikel Menta, Andrew D Bagdanov, and Joost Van De Weijer. Class-incremental learning: survey and performance evaluation on image classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(5):5513–5533, 2022.
  • [24] Michael McCloskey and Neal J Cohen. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of learning and motivation, volume 24, pages 109–165. Elsevier, 1989.
  • [25] Can Peng, Kun Zhao, Tianren Wang, Meng Li, and Brian C Lovell. Few-shot class-incremental learning from an open-set perspective. In European Conference on Computer Vision, pages 382–397. Springer, 2022.
  • [26] Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, and Christoph H Lampert. icarl: Incremental classifier and representation learning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pages 2001–2010, 2017.
  • [27] David Rolnick, Arun Ahuja, Jonathan Schwarz, Timothy Lillicrap, and Gregory Wayne. Experience replay for continual learning. Advances in Neural Information Processing Systems, 32, 2019.
  • [28] Jake Snell, Kevin Swersky, and Richard Zemel. Prototypical networks for few-shot learning. Advances in neural information processing systems, 30, 2017.
  • [29] Zeyin Song, Yifan Zhao, Yujun Shi, Peixi Peng, Li Yuan, and Yonghong Tian. Learning with fantasy: Semantic-aware virtual contrastive constraint for few-shot class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 24183–24192, 2023.
  • [30] Zhaorui Tan, Xi Yang, and Kaizhu Huang. Rethinking multi-domain generalization with a general learning objective. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 23512–23522, 2024.
  • [31] Zhaorui Tan, Xi Yang, Qiufeng Wang, Anh Nguyen, and Kaizhu Huang. Interpret your decision: Logical reasoning regularization for generalization in visual classification. arXiv preprint arXiv:2410.04492, 2024.
  • [32] Xiaoyu Tao, Xiaopeng Hong, Xinyuan Chang, Songlin Dong, Xing Wei, and Yihong Gong. Few-shot class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12183–12192, 2020.
  • [33] Songsong Tian, Lusi Li, Weijun Li, Hang Ran, Xin Ning, and Prayag Tiwari. A survey on few-shot class-incremental learning. arXiv preprint arXiv:2304.08130, 2023.
  • [34] Yonglong Tian, Yue Wang, Dilip Krishnan, Joshua B Tenenbaum, and Phillip Isola. Rethinking few-shot image classification: a good embedding is all you need? In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIV 16, pages 266–282. Springer, 2020.
  • [35] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  • [36] Vikas Verma, Alex Lamb, Christopher Beckham, Amir Najafi, Ioannis Mitliagkas, David Lopez-Paz, and Yoshua Bengio. Manifold mixup: Better representations by interpolating hidden states. In International conference on machine learning, pages 6438–6447. PMLR, 2019.
  • [37] Oriol Vinyals, Charles Blundell, Timothy Lillicrap, Daan Wierstra, et al. Matching networks for one shot learning. Advances in neural information processing systems, 29, 2016.
  • [38] Qi-Wei Wang, Da-Wei Zhou, Yi-Kai Zhang, De-Chuan Zhan, and Han-Jia Ye. Few-shot class-incremental learning via training-free prototype calibration. Advances in Neural Information Processing Systems, 36, 2024.
  • [39] Yulin Wang, Xuran Pan, Shiji Song, Hong Zhang, Gao Huang, and Cheng Wu. Implicit semantic data augmentation for deep networks. Advances in Neural Information Processing Systems, 32, 2019.
  • [40] Yue Wu, Yinpeng Chen, Lijuan Wang, Yuancheng Ye, Zicheng Liu, Yandong Guo, and Yun Fu. Large scale incremental learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 374–382, 2019.
  • [41] Shipeng Yan, Jiangwei Xie, and Xuming He. Der: Dynamically expandable representation for class incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3014–3023, 2021.
  • [42] Hong-Ming Yang, Xu-Yao Zhang, Fei Yin, Qing Yang, and Cheng-Lin Liu. Convolutional prototype network for open set recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(5):2358–2370, 2020.
  • [43] Shuo Yang, Songhua Wu, Tongliang Liu, and Min Xu. Bridging the gap between few-shot and many-shot learning via distribution calibration. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(12):9830–9843, 2021.
  • [44] Yibo Yang, Haobo Yuan, Xiangtai Li, Zhouchen Lin, Philip Torr, and Dacheng Tao. Neural collapse inspired feature-classifier alignment for few-shot class incremental learning. arXiv preprint arXiv:2302.03004, 2023.
  • [45] Jaehong Yoon, Sultan Madjid, Sung Ju Hwang, Chang-Dong Yoo, et al. On the soft-subnetwork for few-shot class incremental learning. In International Conference on Learning Representations (ICLR) 2023. International Conference on Learning Representations, 2023.
  • [46] Tianyuan Yu, Sen He, Yi-Zhe Song, and Tao Xiang. Hybrid graph neural networks for few-shot learning. In Proceedings of the AAAI conference on artificial intelligence, volume 36, pages 3179–3187, 2022.
  • [47] Chi Zhang, Nan Song, Guosheng Lin, Yun Zheng, Pan Pan, and Yinghui Xu. Few-shot incremental learning with continually evolved classifiers. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 12455–12464, 2021.
  • [48] Jian Zhang, Chenglong Zhao, Bingbing Ni, Minghao Xu, and Xiaokang Yang. Variational few-shot learning. In Proceedings of the IEEE/CVF International conference on computer vision, pages 1685–1694, 2019.
  • [49] Da-Wei Zhou, Fu-Yun Wang, Han-Jia Ye, Liang Ma, Shiliang Pu, and De-Chuan Zhan. Forward compatible few-shot class-incremental learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9046–9056, 2022.
  • [50] Da-Wei Zhou, Qi-Wei Wang, Han-Jia Ye, and De-Chuan Zhan. A model or 603 exemplars: Towards memory-efficient class-incremental learning. In The Eleventh International Conference on Learning Representations, 2022.
  • [51] Da-Wei Zhou, Han-Jia Ye, Liang Ma, Di Xie, Shiliang Pu, and De-Chuan Zhan. Few-shot class-incremental learning by sampling multi-phase tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  • [52] Fei Zhu, Zhen Cheng, Xu-yao Zhang, and Cheng-lin Liu. Class-incremental learning via dual augmentation. Advances in Neural Information Processing Systems, 34:14306–14318, 2021.
  • [53] Fei Zhu, Xu-Yao Zhang, Chuang Wang, Fei Yin, and Cheng-Lin Liu. Prototype augmentation and self-supervision for incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5871–5880, 2021.
  • [54] Kai Zhu, Yang Cao, Wei Zhai, Jie Cheng, and Zheng-Jun Zha. Self-promoted prototype refinement for few-shot class-incremental learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 6801–6810, 2021.
  • [55] Yixiong Zou, Shanghang Zhang, Yuhua Li, and Ruixuan Li. Margin-based few-shot class-incremental learning with class-level overfitting mitigation. Advances in neural information processing systems, 35:27267–27279, 2022.

This Appendix first provides details of the three baseline models in Appendix A. Comparisons of different incremental learning approaches with SPL are provided in Appendix B. In Appendix C, we provide additional experimental results on CIFAR100 and CUB200.

Appendix A Details of Baseline Models

We choose three prevalent baseline models, i.e., the naive CE-trained model, fantasy-based model FACT [49], and SAVC as the baseline models [29].

  • CE: The base model is trained simply using cross-entropy loss in the base session. For the incremental sessions, the feature extractor is frozen, and only the classifier is updated.

  • FACT [49]: During the base session, virtual new classes are synthesized by manifold mixup [36] to assist the base training, intending to save feature space for new classes. For incremental sessions, the model is updated by adding new prototypes to the classifier.

  • SAVC [29]: Contrastive learning [13] is adopted in base session to learn compact representations. During the incremental sessions, multiple prototypes from each new class are ensembled as new classifier parameters.

Appendix B Comparison of Different Incremental Learning Approaches

We conduct experiments on comparing different incremental learning approaches on fine-grained dataset CUB200 to verify the effectiveness when learning new classes. As shown in Tab. A3, we compare the proposed SPL with two commonly used incremental learning approaches, i.e., prototype-based update in Sec. 3.1.2, fine-tuning the last layer of the model by CE using few-shot data [44, 29]. We use the same base model, followed by 10 incremental sessions. The prototype-based model obtains the lowest new class performance, average performance of ten sessions, and harmonic accuracy since it cannot separate the new classes from other classes efficiently. The prototype-based model does obtain the highest old class performance and the least drop in PD, as it does not involve any update of feature space. The finetune approach boosts the new class performance by updating the feature space, hence obtaining higher average performance and harmonic accuracy. Compared to the finetuning approach, our SPL expands the new class feature distributions and facilitates a wider margin between classes. Therefore, SPL can retrain higher base class performance while learning new classes more effectively, achieving the highest performance of new classes, average performance and harmonic accuracy.

Table A3: Comparison of different incremental Methods. “Prototype-based” refers to the approach that simply updates the new class prototypes during incremental learning. “Finetune by CE” denotes using CE to finetune the last layer of the model with few-shot data.
Incremental Method CUB200
Base\uparrow Old\uparrow New\uparrow Avg\uparrow PD\downarrow H. \uparrow
Prototype-based [49] 81.31 76.96 47.00 68.88 4.35 58.35
Finetune by CE [29] 81.31 76.54 47.71 68.93 4.77 58.78
SPL 81.31 76.68 47.85 69.12 4.63 58.93

Appendix C More Benchmark Results

Table A4: Incremental learning performance on CIFAR100 under 5-way 5-shot setup. “Avg Acc.” represents the average accuracy of all sessions. “Final Improv.” calculates the improvement of our method after learning in the final session. Bold represents best performance. * indicates that we reproduce the results using public open-source code
Methods Accuracy in each session (%) \uparrow Avg Acc. Final Improv.
0 1 2 3 4 5 6 7 8
iCaRL [26] 64.10 53.28 41.69 34.13 27.93 25.06 20.41 15.48 13.73 32.87 +39.41
NCM [11] 64.10 53.05 43.96 36.97 31.61 26.73 21.23 16.78 13.54 34.22 +39.60
Data-free Replay [22] 74.40 70.20 66.54 62.51 59.71 56.58 54.52 52.39 50.14 60.78 +3.00
Self-promoted [54] 64.10 65.86 61.36 57.45 53.69 50.75 48.58 45.66 43.25 54.52 +9.89
CEC [18] 73.07 68.88 65.26 61.19 58.09 55.57 53.22 51.34 49.14 59.53 +4.00
MetaFSCIL [5] 74.50 70.10 66.84 62.77 59.48 56.52 54.36 52.56 49.97 60.79 +8.05
C-FSCIL [10] 77.47 72.40 67.47 63.25 59.84 56.95 54.42 52.47 50.47 61.64 +3.17
LIMIT [51] 72.32 68.47 64.30 60.78 57.95 55.07 52.70 50.72 49.19 59.06 +3.95
CE 76.87 72.38 68.06 63.83 60.52 57.76 55.47 53.25 50.94 62.12 +2.20
CE-Ours 78.27 73.80 69.69 65.53 62.07 59.33 57.22 54.75 52.30 62.21 +0.84
FACT [49] 78.38 71.86 67.87 64.10 60.70 57.75 55.83 53.6 51.34 62.17 +1.80
FACT-Ours 79.12 72.62 68.49 64.31 61.51 58.64 56.38 54.22 52.34 62.82 +0.80
SAVC [29] 78.98 73.02 68.69 64.49 60.91 58.08 55.79 53.61 51.75 62.81 +1.39
SAVC-Ours 79.00 73.29 68.84 64.75 61.60 58.74 56.84 55.12 53.14 63.48
Table A5: Performance of FSCIL in each session on CUB200 under 10-way 5-shot setup and comparison with other studies. “Average Acc.” is the average accuracy of all sessions. “Final Improv.” calculates the improvement of our method in the last session. Bold represents best performance. * indicates that we reproduce the results using public open-source code.
Methods Accuracy in each session (%) \uparrow Avg Acc. Final Improv.
0 1 2 3 4 5 6 7 8 9 10
iCaRL [26] 68.68 52.65 48.61 44.16 36.62 29.52 27.83 26.26 24.01 23.89 21.16 36.67 +41.54
Data-free Replay [22] 75.90 72.14 68.64 63.76 62.58 59.11 57.82 55.89 54.92 53.58 52.39 61.52 +10.31
LDC [21] 77.89 76.93 74.64 70.06 68.88 67.15 64.83 64.16 63.03 62.39 61.58 68.32 +1.12
CEC [18] 75.85 71.94 68.50 63.50 62.43 58.27 57.73 55.81 54.83 53.52 52.28 61.33 +10.42
LIMIT [51] 76.32 74.18 72.68 69.19 68.79 65.64 63.57 62.69 61.47 60.44 58.45 66.67 +4.25
MetaFSCIL [5] 75.90 72.41 68.78 64.78 62.96 59.99 58.3 56.85 54.78 53.82 52.64 61.93 +10.06
CE 79.32 75.67 72.56 67.42 66.46 62.00 60.85 59.31 57.78 56.88 55.73 64.91 +6.97
CE-Ours 79.59 75.32 72.31 67.46 66.68 63.61 62.68 61.07 59.09 59.20 58.34 65.71 +4.36
FACT 77.28 73.67 70.19 65.59 64.77 61.60 60.68 58.89 57.38 57.26 56.11 63.87 +6.59
FACT-Ours 77.78 74.23 70.42 65.97 65.31 61.58 61.42 59.61 57.42 57.26 56.49 65.15 +6.21
SAVC 81.31 77.35 74.49 69.65 69.78 67.10 66.48 64.09 63.16 62.48 61.81 68.88 +0.89
SAVC-Ours 82.67 78.58 75.66 70.83 70.37 67.30 66.80 65.57 64.01 63.45 62.70 69.81

We also present the performance of our method on the CIFAR100 and CUB200 datasets, as shown in Tab. A4 and Tab. A5, respectively. On CIFAR100, our approach boosts the performance of baseline methods in all sessions. Our method improves the final performance of three baselines by at least 0.65% and boosts the average performance on all incremental sessions. The improvement is attributed to the covariance constraint loss and semantic perturbation learning, which promote effective class separation and few-shot new class learning. On the fine-grained dataset CUB200, which includes 200 classes, our method achieves a final performance of 62.70%, demonstrating the effectiveness of our approaches. We obtain an improvement in final accuracy of 2.61% by applying our approach to the CE baseline model. In session 1 and session 2, our method yields lower performance on the CE model due to the imbalance of base class and new class in the testing data, but in the following incremental sessions, our approach is able to boost the overall performance.