This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Learning a Condensed Frame for Memory-Efficient Video Class-Incremental Learning

Yixuan Pei1∗  Zhiwu Qing2  Jun Cen3  Xiang Wang2  Shiwei Zhang4
Yaxiong Wang1Mingqian Tang4Nong Sang2Xueming Qian1

Xi’an Jiaotong University1
Huazhong University of Science and Technology2
The Hong Kong University of Science and Technology3
Alibaba Group4
{peiyixuan@stu, wangyx15@stu, qianxm@mail}.xjtu.edu.cn,
{qzw,wxiang,nsang}@hust.edu.cn,

[email protected], {zhangjin.zsw, mingqian.tmq}@alibab-inc.com
equal contribution
Abstract

Recent incremental learning for action recognition usually stores representative videos to mitigate catastrophic forgetting. However, only a few bulky videos can be stored due to the limited memory. To address this problem, we propose FrameMaker, a memory-efficient video class-incremental learning approach that learns to produce a condensed frame for each selected video. Specifically, FrameMaker is mainly composed of two crucial components: Frame Condensing and Instance-Specific Prompt. The former is to reduce the memory cost by preserving only one condensed frame instead of the whole video, while the latter aims to compensate the lost spatio-temporal details in the Frame Condensing stage. By this means, FrameMaker enables a remarkable reduction in memory but keep enough information that can be applied to following incremental tasks. Experimental results on multiple challenging benchmarks, i.e., HMDB51, UCF101 and Something-Something V2, demonstrate that FrameMaker can achieve better performance to recent advanced methods while consuming only 20% memory. Additionally, under the same memory consumption conditions, FrameMaker significantly outperforms existing state-of-the-arts by a convincing margin.

1 Introduction

Training video action recognition models with all classes in a single fine-tuning stage have been widely studied [6, 34, 51, 62, 52, 30, 13, 37] in recent years. However, in many realistic scenarios, limited by privacy matters or technologies, the different classes can only be presented in sequence, where the previously trained classes are either unavailable or partially available in limited memory. Naively training models will overfit currently available data and cause performance deterioration on previously seen classes, which is named catastrophic forgetting [39]. Class-incremental learning is a machine learning paradigm to fight this challenge when fine-tuning a single model in a sequence of independent classes.

A branch of class-incremental learning methods [7, 9, 12, 21, 45, 33, 60, 5] has achieved remarkable performances in the image domain by re-training a portion of past examples. Meanwhile, some existing video incremental learning methods [41, 54] have demonstrated that storing more past examples can improve the ability to mitigate forgetting. Nevertheless, videos are information-redundant and require a high memory load to be saved, hence it is impractical to store a good deal of training videos for each class. Therefore, herding strategy [45] is adopted to select a small set of representative videos for the exemplar memory [41, 54]. vCLIMB [54] tries to down-sample frames for each video to reduce the memory consumption, and imposes a temporal-consistency regularization for better performance. Despite the remarkable performance, these methods still demand to conserve multiple (i.e., 8-16) frames for each representative video, leading to non-negligible memory overhead, which limits their further real-world applications.

Refer to caption
Figure 1: An intuitive comparison between FrameMaker and TCD [41]. (a) Compared with TCD, FrameMaker stores only one condensed frame for each video exemplar. (b) Illustration of memory consumption and accuracy curve. FrameMaker costs 1.2Mb memory, which slightly exceeds the performance of TCD with 6.0Mb memory.

To remedy the above weaknesses, we present FrameMaker, a memory-efficient video class-incremental learning approach. Concretely, the proposed FrameMaker contains two components, Frame Condensing and Instance-Specific Prompt. In Frame Condensing, FrameMaker assigns a learnable weight to each frame, and then these weights are adjusted to make the weighted sum frame i.e., the condensed frame, share the same embedding features with the original video clip. Obviously, the collapsed temporal dimension and mixed spatial cues by Frame Condensing may degrade the accuracy of action recognition. To compensate the missing spatio-temporal details, inspired by Task-Specific Prompt [36], we further propose Instance-Specific Prompt, which learns a set of parameters for each condensed frame to perform fine-grained pixel-wise modifications. In this way, the memory required to store a single video can be remarkably compressed. An intuitive comparison is given in Figure 1(a). On the premise of comparable performance with TCD [41], FrameMaker significantly reduces memory overhead by 80%. Moreover, the comparison curve displayed in Figure 1(b) further indicates that FrameMaker achieves a consistent performance improvement over recent TCD with the same memory consumption.

In summary, we make the following contributions: 1) We present a memory-efficient video class-incremental learning method FrameMaker, which significantly reduces the memory consumption of example videos by integrating multiple frames into one condensed frame; 2) For the first time, we show a novel perspective of Prompt, i.e., Instance-Specific Prompt, and demonstrate that it can embed effective spatio-temporal interactions into learned condensed frames for video class-incremental learning; 3) Experimental results reveal that FrameMaker consistently outperforms recent state-of-the-art methods with the same memory overhead on three challenging benchmarks (HMDB51, UCF101, and Something-Something V2).

2 Related Works

2.1 Class-Incremental Learning

The class-incremental learning has been well-studied in image domain [1, 32, 26, 63], existing approaches could be grouped into knowledge distillation-based methods  [20], memory-based methods [40, 47, 45], and architecture-based methods. In video domain, class-incremental learning for action recognition is still an underexplored area. The existing solutions are designed for better temporal constraints, such as decomposing the spatio-temporal features [65], exploiting time-channel importance maps [41], and applying temporal consistency regularization [54]. Notably, both  [41] and  [54] have shown that more stored examples in memory can effectively encourage the performance. However, FrameMaker attempts to minimize the stored frames for each video by a pretty maneuver, which can provide decent performance with slight storage requirements.

2.2 Prompt-Based Learning

Prompting is an emerging transfer learning technique in natural language processing [36], which adapts the language model to downstream tasks by a fixed function. Since the design of prompting functions is challenging, the learnable prompts [29, 31] are proposed to learn task-specific knowledge parameters for the input. Recently, in the visual domain, some works have also emerged to adapt the pre-trained models to new tasks by adding prompt tokens [24] or pixels [3] in the data space. Besides, DualPrompt [58] and L2P [57] get rid of the rehearsal buffer in continual learning by learning a set of task-specific parameters. In this paper, we draw inspiration from prompting, and present an Instance-Specific Prompt to replenish the collapsed spatio-temporal cues for condensed frames.

2.3 Action Recognition

The approaches designed for action recognition task can be grouped into three categories: 2D CNN, 3D CNN and Transformer-based methods. 2D CNN based methods is more efficient [19, 49], while 3D CNN [16, 6, 15, 51, 53, 10, 62, 43, 56, 52] achieves better accuracy in action recognition at the cost of more computation cost. With the rapid development of vision transformer [35, 22, 38], researchers also apply it into the video recognition [2, 4, 37, 13], which is proven to be remarkable effective. In this work, we follow the settings in  [41, 65], and employ the Temporal Shift Module (TSM) [34] based on 2D CNN for efficient experiments.

The video summarization and efficient action recognition tasks have provided some methods for key-frame selection and generation, which are related to this work. The video summarization [67, 55, 64] task selects the most representative frames for efficient human understanding, which has a different paradigm with action recognition task, hence we will mainly discuss efficient The efficient video classification task aims to learn a strategy to improve classification accuracy with fewer discriminative frames as input. The approaches designed for this task can be grouped into two categories: rule-based and model-based methods. The former [27, 66] select frames via some pre-defined rules such as motion distribution [66] or channel sampling [25], and the later design a submodel to select or generate key frames [44, 23, 61, 59, 14, 17]. In this paper, we propose model-free frame condensing method for video class-incremental learning, which only learns a few condensing weights and prompting parameters for each instance.

Refer to caption
Figure 2: An overview of the proposed FrameMaker, which mainly contains Frame Condensing and Instance-Specific Prompt. For each selected representative video VikV_{i}^{k}, we first uniformly sample TT frames and learn a condensing weight wikw_{i}^{k} for each frame to perform Frame Condensing. Then a learnable prompt PikP_{i}^{k} is introduced for the condensed frame to compensate the lost spatio-temporal information. Finally, distillation loss and cross entropy loss are both used to guide the optimization of condensing weights and prompt.

3 Method

This section presents the general framework of FrameMaker as shown in Figure 2, which mainly consists of two crucial components: Frame Condensing and Instance-Specific Prompt. Firstly, Frame Condensing is designed to integrate multiple frames into one frame for efficient storage. Then we introduce Instance-Specific Prompt to compensate the collapsed temporal information and the mixed spatial knowledge. Finally, the specific class-incremental training details for subsequent tasks using Condensed Frames are given.

3.1 Problem formulation

The purpose of video class-incremental learning is to train a model F(;Θ)=h(f(;θ);ξ)F(;\Theta)=h(f(;\theta);\xi) parameterized by Θ\Theta as the tasks {𝒯1,𝒯2,,𝒯K}\{\mathcal{T}^{1},\mathcal{T}^{2},\cdots,\mathcal{T}^{K}\} arrive in a sequence, where f(;θ)f(;\theta) is the feature extractor and h(;ξ)h(;\xi) is the classification head. Each task 𝒯K\mathcal{T}^{K} has its specific dataset Dk={(dik,yik),yikLk}D^{k}=\{(d_{i}^{k},y_{i}^{k}),y_{i}^{k}\in{L^{k}}\} whose labels belong to a predefined label set LkL^{k} and have not appeared in previous tasks. The memory-based methods [45, 54, 41] have been shown to be effective in preventing catastrophic forgetting. Specifically, after the incremental training step kk, a memory bank MkM^{k} will be established for the dataset DkD^{k} in this task to store the representative exemplars or features. Exemplars stored in previous datasets M1:k=M1M2MkM^{1:k}=M^{1}\bigcup{M^{2}}\bigcup{}\cdots\bigcup{M^{k}} will be used in subsequent incremental tasks to mitigate forgetting. At the incremental step kk, the model F(;Θk)F(;\Theta_{k}) is trained from F(;Θk1)F(;\Theta_{k-1}) with Dk=DkM1:(k1){D^{\prime}}^{k}=D^{k}\bigcup{M^{1:(k-1)}} and evaluated on all seen classes.

3.2 FrameMaker

Frame Condensing.

We follow the standard protocol of video class-incremental methods [54, 41], which are based on memory-replay strategy and knowledge distillation. Differently, we devotes to reducing the number of stored frames for representative videos to achieve memory-efficient video class-incremental learning. Formally, we first select a subset of video instances VkV^{k} by herding strategy [45] from DkD^{k} after the incremental step kk. Given a video VikVkV_{i}^{k}\in{V^{k}}, we then sample TT frames {Ii1k,Ii1k,,IiTk}\{I_{i1}^{k},I_{i1}^{k},\cdots,I_{iT}^{k}\} uniformly and integrate them to condense a more representative one. To this end, we define learnable weights Wik={wi1k,wi2k,,wiTk}W_{i}^{k}=\{w_{i1}^{k},w_{i2}^{k},\cdots,w_{iT}^{k}\} for the ii-th video instance VikV_{i}^{k} in exemplar set VkV^{k}. And the condensed frame can be calculated as follows:

Iik=t=1Tewitkt=1TewitkIitkC×H×W,I_{i}^{k}=\sum_{t=1}^{T}\frac{e^{w_{it}^{k}}}{\sum_{t=1}^{T}e^{w_{it}^{k}}}{I_{it}^{k}}\in{\mathbb{R}^{{C}\times{{H}}\times{{W}}}}\ , (1)

where CC, HH and WW are the size of channel, height and width of input frames, respectively. Next, the condensed frame IikI_{i}^{k} is excepted to empower the same or much similar expressive ability as the original video clip. Hence, the embedding features of the condensed frame extracted from current model should be consistent with the features of the original video clip:

Lfc=f(Iik;θk)f(Vik;θk)2,L^{\text{c}}_{\text{f}}={\mid\mid{f(I_{i}^{k};\theta_{k})-f(V_{i}^{k};\theta_{k})}\mid\mid}^{2}\ , (2)

where f(;θk)f(;\theta_{k}) is the feature extractor of current model F(;Θk)F(;\Theta_{k}). This consistency regularization can foster the condensing weights to exploit memory-efficient expressions of video clips. To further improve the adaptability of the condensed frames to the current model, we employ cross entropy loss to supervise the classification confidence from the condensed frame:

Lcec=CrossEntropy(F(Iik;Θk),yik).L^{\text{c}}_{\text{ce}}={\text{CrossEntropy}(F(I_{i}^{k};\Theta_{k}),y_{i}^{k})}\ . (3)

The full objective function for condensing weights WikW_{i}^{k} is given by:

Lc=Lfc+Lcec.L_{\text{c}}={{L^{\text{c}}_{\text{f}}}+L^{\text{c}}_{\text{ce}}}\ . (4)

With the methods described above, we can learn an effective representation, i.e., the condensed frame for each video based on current model.

Instance-Specific Prompt.

However, condensing a video into one frame would inevitably lose the the temporal dynamics and complete spatial information of the original video to a certain extent. Therefore, an intriguing perspective of prompt, termed as Instance-Specific Prompt, is proposed to replenish the missing spatio-temporal cues for the condensed frame.

Specifically, we first construct a learnable prompt PikC×H×WP_{i}^{k}\in{\mathbb{R}^{{C}\times{{H}}\times{{W}}}} for each exemplar video VikV_{i}^{k}, which shares the same spatial resolution as its original clip. The prompt PikP_{i}^{k} is then pixel-wise summed with the condensed frame IikI_{i}^{k}, and learns to pull the embedding feature of the summed frame (Iik+Pik)(I_{i}^{k}+P_{i}^{k}) and the video clip VikV_{i}^{k} together, which is similar to Eq. 2:

Lfp=f((Iik+Pik);θk)f(Vik;θk)2.L^{\text{p}}_{\text{f}}={\mid\mid{f((I_{i}^{k}+P_{i}^{k});\theta_{k})-f(V_{i}^{k};\theta_{k})}\mid\mid}^{2}\ . (5)

Notably, the Eq. 2 for condensing weights is difficult to learn the satisfactory spatio-temporal features due to the collapse of the temporal dimension. Nevertheless, the Eq. 5 can enrich the representations of condensing frames by introducing more flexible learnable parameters in the input space. Besides, the Cross-Entropy loss is also utilized to enhance the semantic perception:

Lcep=CrossEntropy(F((Iik+Pik);Θk),yik).L^{\text{p}}_{\text{ce}}=\text{CrossEntropy}(F((I_{i}^{k}+P_{i}^{k});\Theta_{k}),y_{i}^{k})\ . (6)

Theoretically, the condensing weights WikW_{i}^{k} and the prompting parameters PikP_{i}^{k} can be jointly updated by Lp=Lfp+LcepL_{\text{p}}={L^{\text{p}}_{\text{f}}+L^{\text{p}}_{\text{ce}}}. Interestingly, we observed in our practice that the flexible prompt PikP_{i}^{k} leads to the condensing weights WikW_{i}^{k} being under optimized. Hence, we give the final training objective function for frame condensing as:

Lfc=αLfc+βLcec+γLfp+ηLcep,L_{\text{fc}}={\alpha}L^{\text{c}}_{{\text{f}}}+{\beta}L^{\text{c}}_{{\text{ce}}}+{\gamma}L^{\text{p}}_{{\text{f}}}+{\eta}L^{\text{p}}_{\text{ce}}\ , (7)

where additional LfcL^{\text{c}}_{\text{f}} and LcecL^{\text{c}}_{\text{ce}} are added to achieve stronger constraint on WikW_{i}^{k} for its effective training. α{\alpha}, β{\beta}, γ{\gamma} and η{\eta} are the balance weights for each term, and they are empirically set to 1.0 unless otherwise specified.

After the optimization of condensing weights and prompting parameters, the prompt is added directly to the condensed frame and stored together. And the learned prompts will be frozen when the corresponding task is completed. Therefore, there is no extra memory cost for prompts.

3.3 Training

It is worth noting that FrameMaker aims to condense the stored frames for representative videos, and other class-incremental learning steps still follow the standard pipeline. Specifically, when training the incremental step kk, we use the dataset Dk=DkMk1{D^{\prime}}^{k}=D^{k}\bigcup{M^{k-1}} to update the model from F(;Θk1)F(;\Theta_{k-1}), where DkD^{k} is the dataset of the task kk which consists of videos belonging to Lk{L^{k}}, and Mk1{M^{k-1}} is memory bank which contains condensed frames generated by FrameMaker. The training samples from DkD^{k} and Mk1{M^{k-1}} are alternately fed into the current model according to the proportion of their sample number. To further prevent the catastrophic forgetting, we also employ the knowledge distillation method proposed in PODNet [12] to transfer the knowledge from the previous model F(;Θk1)F(;\Theta_{k-1}) to current model F(;Θk)F(;\Theta_{k}). The overall objective function for task kk reads:

Lcil=Lced+Lcem+Ldistm,L_{\text{cil}}={L^{\text{d}}_{{\text{ce}}}+L^{\text{m}}_{{\text{ce}}}+L^{\text{m}}_{{\text{dist}}}}\ , (8)

where LcedL^{\text{d}}_{{\text{ce}}} and LcemL^{\text{m}}_{{\text{ce}}} are the Cross Entropy losses for new task data DkD^{k} and condensed frame exemplars Mk1{M^{k-1}}, respectively. And LdistmL^{\text{m}}_{{\text{dist}}} is the knowledge distillation loss [12] for condensed frame exemplars.

4 Experiments

4.1 Experimental Setup

Datasets and evaluation metrics.

The proposed FrameMaker is evaluated on three standard action recognition datasets, UCF101 [50], HMDB51 [28] and Something-Something V2 [18]. HMDB51 dataset consists of 6.8K videos belonging to 51 classes from YouTube or other websites. UCF101 dataset contains 13.3K videos from 101 classes. Something-Something V2 is a crowd-sourced dataset that includes 220K videos from 174 classes.

For UCF101, the model is trained on 51 classes first, and the remaining 50 classes are divided into 5, 10 and 25 tasks. For HMDB51, we train the base model using videos from 26 classes, and the remaining 25 classes are separated into 5 or 25 groups. For Something-Something V2, we first train 84 classes in the initial stage, and generate the groups of 10 and 5 classes.

To evaluate the performance of class-incremental learning, we infer the test videos from all seen categories after each task, and finally, report the average accuracy of all tasks. Following  [41], two different metrics are reported, i.e., CNN and NME. CNN refers to training a fully-connected layer for extracted features, which is a standard classification protocol. NME is proposed by iCaRL [45], which assigns the labels for test data by comparing the feature embeddings with the mean-of-exemplars.

Implementation details.

TSM [34] is employed as our backbone, and we follow the data preprocessing procedure of TSM. For UCF101, we train a ResNet-34 TSM for 50 epochs with a batch size 256 from an initial learning rate 0.04. For HMDB51 and Something-Something V2, we train a ResNet-50 TSM for 50 epochs with a batch size of 128 from an initial learning rate of 1e-3 and 0.04, respectively. All used networks are first pre-trained on ImageNet [8] for initialization. These settings are consistent with TCD [41]. We train all models on eight NVIDIA V100 GPUs and use PyTorch [42] for all our experiments.

Table 1: Comparison with the state-of-the-art approaches over class-incremental action recognition performance on UCF101 and HMDB51. Our FrameMaker achieves the best performance under all experimental settings.
UCF101 HMDB51
Num. of Classes 10×510\times 5 stages 5×105\times 10 stages 2×252\times 25 stages 5×55\times 5 stages 1×251\times 25 stages
Classifier CNN NME CNN NME CNN NME CNN NME CNN NME
Finetuning 24.97 - 13.45 - 5.78 - 16.82 - 4.83 -
LwFMC [33] 42.14 - 25.59 - 11.68 - 26.82 - 16.49 -
LwM [9] 43.39 - 26.07 - 12.08 - 26.97 - 16.50 -
iCaRL [45] - 65.34 - 64.51 - 58.73 - 40.09 - 33.77
UCIR [21] 74.31 74.09 70.42 70.50 63.22 64.00 44.90 46.53 37.04 37.15
PODNet [12] 73.26 74.37 71.58 73.75 70.28 71.87 44.32 48.78 38.76 46.62
TCD [41] 74.89 77.16 73.43 75.35 72.19 74.01 45.34 50.36 40.47 46.66
FrameMaker 78.13 78.64 76.38 78.14 75.77 77.49 47.54 51.12 42.65 47.37
Table 2: Comparison with the top approaches over class-incremental action recognition performance on Something-Something V2.
Num. of Classes 10×910\times 9 stages 5×185\times 18 stages
Classifier CNN NME CNN NME
UCIR [21] 26.84 17.98 20.69 12.57
PODNet [12] 34.94 27.33 26.95 17.49
TCD [41] 35.78 28.88 29.60 21.63
FrameMaker 37.25 29.92 30.98 22.84
Table 3: Ablations for Frame Condensing (FC) and Instance-Specific Prompting (ISP) on UCF101 with 10 steps and HMDB51 with 5 steps.
UCF101 HMDB51
Frames FC ISP CNN NME CNN NME
All - - 72.09 75.70 43.38 47.00
Random 68.64 73.96 39.59 43.48
Random 70.71 75.04 39.81 43.74
Average 70.82 75.45 41.84 45.45
Average 71.51 76.23 42.59 46.46
Condensed 72.29 76.42 42.18 46.27
Condensed 72.93 76.64 43.39 46.88

4.2 Comparison with State-of-the-art Results

This section presents the qualitative comparison of our proposed FrameMaker with the existing class incremental learning approaches under multiple challenging settings on three datasets: UCF101 [50], HMDB51 [28] and Something-Something V2 [18]. For a fair comparison, we use the same exemplar memory size per class, model structure and pre-training initialization as the existing methods [41].

Table 1 summarizes the results on UCF101 and HMDB51, which shows that FrameMaker outperforms other methods consistently under different configurations in terms of both CNN and NME scores. The average accuracy of our FrameMaker surpasses TCD by around 3.0% on UCF101 and 2.0% on HMDB51 in terms of CNN, respectively. These results demonstrate that one condensed frame can also be equipped with an effective spatio-temporal representation.

FrameMaker is compared with recent advanced methods on Something-Something V2 in Table 3. FrameMaker sets new state-of-the-art performance under multiple challenging settings. Although FrameMaker only uses one condensed frame for future replay, it still achieves decent performance on the large-scale motion-sensitive dataset. We speculate that these gains mainly come from two aspects: (1) The proposed Instance-Specific Prompt effectively rescues the lost spatio-temporal details, which is also discussed in Section 4.3. (2) Memory-efficient FrameMaker allows us to store more exemplar videos with less memory consumption, which greatly alleviates catastrophic forgetting.

4.3 Ablation Study

In this section, we present ablation studies to analyze the properties and the effectiveness of FrameMaker. If not specified, the ablation studies are performed on UCF101 with 10 steps and HMDB51 with 5 steps. For comparison with the case of storing all frames, we select 5 exemplar videos for each class.

Table 4: Ablations for objective functions in Frame Condensing on UCF101 with 10 steps.
Lfc{L^{\text{c}}_{\text{f}}} LcecL^{\text{c}}_{\text{ce}} CNN NME
70.82 75.45
71.35 75.76
71.96 76.07
72.29 76.42
Table 5: Ablations for objective functions in Instance-Specific Prompt on UCF101 with 10 steps.
L*cL^{\text{c}}_{\text{*}} LfpL^{\text{p}}_{\text{f}} LcepL^{\text{p}}_{{}^{\text{ce}}} CNN NME
71.83 75.94
72.35 76.48
72.58 76.57
72.93 76.64
Table 6: The number of frames for Frame Condensing on UCF101 with 10 steps.
TT CNN NME
2 71.62 75.89
4 71.70 76.12
8 72.93 76.64
16 72.42 76.11
Table 7: Ablations for different backbones for FrameMaker on HMDB51 with 5 steps.
Backbone Frames CNN NME Mem.
TSM All 43.38 47.00 6.00Mb
TSM FC+ISP 43.39 46.88 0.75Mb
R3D50 All 39.85 45.64 6.00Mb
R3D50 FC+ISP 39.88 45.08 0.75Mb
ViT All 35.34 39.46 6.00Mb
ViT FC+ISP 35.25 39.58 0.75Mb
Table 8: Ablations for alternative prompting strategies for FrameMaker on HMDB51 with 5 steps.
Position Type CNN NME Mem.
Feature T.-Spec. 41.67 45.71 77.07Mb
Feature C.-Spec. 41.72 45.93 385.35Mb
Feature I.-Spec. 42.19 46.53 1926.75Mb
Frame T.-Spec. 41.16 45.34 0.75Mb
Frame C.-Spec. 42.01 45.47 0.75Mb
Frame I.-Spec. 43.39 46.88 0.75Mb

Frame Condensing and Instance-Specific Prompt.

To show the effectiveness of Frame Condensing and Instance-Specific Prompt, we compare them with the following alternative methods to produce stored frames for future replay: (1) All Frames, which simply saves the whole video. (2) One Random Frame, which randomly selects one frame for each exemplar video. (3) One Average Frame, which averages the input frames for each exemplar video. (4) One Condensed Frame, which is generated by our Frame Condensing procedure. The results are reported in Table 3. From the table, in terms of CNN, we observe that Instance-Specific Prompt yields an improvement of 2.07% and 0.22% with randomly sampled frames, 0.69% and 0.75% with averaged frames on UCF101 and HMDB51, respectively, which implies that prompts can always replenish the ignored spatio-temporal cues in video frames. Further, Frame Condensing with Instance-Specific Prompt leads to at least on par or better performance than preserving all frames while using just one condensed frame under all settings on both datasets.

Loss terms.

To preserve the best spatio-temporal information in the condensed frames, the distillation loss Lf*L^{\text{*}}_{\text{f}} and the cross entropy loss Lce*L^{\text{*}}_{\text{ce}} are both applied for Frame Condensing and Instance-Specific Prompt. Table 6 presents the results with Frame Condensing from several different combinations of loss terms. LcecL^{\text{c}}_{\text{ce}} provides more explicit semantic supervision, making the accuracy to step further. Similar experiments for Instance-Specific Prompt are shown in Table 6. One noticeable thing is that once introducing prompt, we only need to calculate the final LfpL^{\text{p}}_{\text{f}} and LcepL^{\text{p}}_{\text{ce}} to realize the joint update of condensing weights WikW^{k}_{i} and prompt PikP^{k}_{i}. However, we observe that the learned weights, in this case, tend to be average, which indicates that WikW^{k}_{i} is under-optimized. We speculate that this is due to the strong plasticity of prompt, which leads to the insufficient optimization of WikW^{k}_{i}. To this end, we further employ LfcL^{\text{c}}_{\text{f}} and LcecL^{\text{c}}_{\text{ce}} for WikW^{k}_{i} as the guidance. The improvement in Table 6 and the visualizations in Figure 3 show that WikW^{k}_{i} receives effective learning.

The number of frames used for Frame Condensing.

As discussed in Section 3.2, the TT frames are uniformly sampled from a representative video for Frame Condensing. In Table 6, we show the effect of TT in video class-incremental learning. We observe that the performance increases as TT increases for both CNN and NME. However, the performance is saturated when using 16 frames, which is in line with the conclusion of existing methods [54, 41], i.e., storing more frames in a video does not necessarily deliver performance improvement. This phenomenon reveals that FrameMaker can empower one Condensed Frame to absorb plenty of spatio-temporal features from original clips.

Different backbones.

To evaluate the applicability of proposed FrameMaker, we replace the backbone with 3D CNN-based method R3D50 [16] and video transformer ViT[11], as shown in Table 8. In these experiments, we simply replicate the condensed frames multiple times along the the temporal dimension to meet the 3D models and Transformers. From the results, we can find that our FrameMaker can consistently achieve comparable performance on the three different backbones with only 12% memory cost. These results demonstrate the better generalization of FrameMaker.

Alternative prompting strategies.

We conduct experiments to explore the types and positions of the prompts in Table 8. (i) We compare our Instance-Specific (I.-Spec.) Prompt (ISP) with the Task- (T.-Spec.) and Class-Specific (C.-Spec.) prompts and observe that our ISP can achieve the best performance when operating on both features and frames. Since the actions in videos have great intra-class variance, simply sharing task and class-specific prompts is insufficient to reserve the important spatia-temporal cues for each instance; (ii) Positions of prompt: For the feature settings, we add the prompt on the features of the 4 stages of the ResNet. The feature-based prompt is worse than our frame-based prompt, which may be caused by the mismatch between the running model and fixed prompts during incremental procedure. Moreover, additional features require additional storage space because the corresponding prompt cannot be directly added on the changed feature.

Table 9: Analysis for memory budget on UCF101 with 10 steps. Here ‘xF×yVx\text{F}\times y\text{V}=zzMb’ indicates that xx sampled frames from yy different videos are stored for each class, and xx frames are sampled from each video. We assume that the spatial resolution of frames is 224×\times224, and the total memory consumption is zzMbytes.
Memory Per Class 8F×1V8\text{F}\times 1\text{V}=1.2Mb 8F×2V8\text{F}\times 2\text{V}=2.4Mb 8F×5V8\text{F}\times 5\text{V}=6.0Mb
Classifier CNN NME CNN NME CNN NME
iCaRL [45] - 58.05 - 60.50 - 64.51
UCIR [21] 61.92 65.52 66.43 67.58 70.42 70.50
PODNet [12] 63.18 70.96 65.93 72.78 71.58 73.75
TCD [41] 64.52 71.96 68.40 73.30 73.43 75.35
Memory Per Class 1F×1V1\text{F}\times 1\text{V}=0.15Mb 1F×2V1\text{F}\times 2\text{V}=0.3Mb 1F×5V1\text{F}\times 5\text{V}=0.75Mb
FrameMaker 49.37 70.78 62.06 74.18 72.93 76.64
Memory Per Class 1F×8V1\text{F}\times 8\text{V}=1.2Mb 1F×16V1\text{F}\times 16\text{V}=2.4Mb 1F×40V1\text{F}\times 40\text{V}=6.0Mb
FrameMaker 73.64 76.98 75.19 77.43 76.38 78.14

Memory budget comparison.

We now compare the memory budget of FramMaker with existing approaches. To make a direct comparison, we use the definition of the working memory size in terms of stored frames, following [54]. The results are summarized in Table 9. We can see that our FrameMaker only costs 1.2Mb memory, which surpasses TCD [41] using 6.0Mb by 0.21% and 1.63% in terms of CNN and NME. FrameMaker saves 80% of memory space with the comparable performance. Meanwhile, we further increase the stored videos but keep the same memory with existing methods. FrameMaker can still effectively promote performance and shows superior abilities to fight catastrophic forgetting. Our results are about 3% ahead of TCD in both indicators. Besides, we also attempt to select the same number of videos as the existing methods. However, the accuracy of the trained linear classifier, i.e., CNN, is weaker with a small number of video instances. Interestingly, NME evaluated by the exemplar class centre yielded fairly or even better results, implying that the model suffers only slight forgetting. We hypothesize that this is caused by the poor feature diversity that a few condensed frames are provided. Therefore, the classifier tends to overfit fixed sample features, resulting in poor generalization.

Refer to caption
Figure 3: Visualization of Condensing Weights and Instance-Specific Prompts on HMDB51.

Visualization of condensing weights and instance-specific prompts.

To intuitively understand our FrameMaker, we provide visualization of condensing weights, Instance-Specific Prompts, condensed frames and learned prompts in Figure 3. To save space, we uniformly select the most typical 4 frames from TT frames for display. Based on this, we have the following fascinating observations: (1) For frames with strong scene bias, the learned weights are more uniform, while those with motion bias are not. For example, the action classification for (a) and (b) in Figure 3 does not depend on a specific frame. Hence, the condensing weight for each frame is roughly the same. In contrast, stair climbing and hugging in (c) and (d) only occur in some specific frames, and those keyframes are highlighted by condensing weights. This demonstrates that it is reasonable to condense multiple frames into one frame since our proposed Frame Condensing can assist the learning of which frame is meaningful for the action recognition task. (2) It is difficult for humans to absorb useful knowledge from the learned prompts shown in Figure 3, and even they seem to be the same. It should be emphasized that the prompts are different. We try to share the same prompt for all videos to verify this. The experiment is conducted on UCF101 with 10 steps, and the performances decrease by 0.83% and 0.78% in terms of CNN and NME, respectively. This fully validates that our Instance-Specific Prompt is to replenish the specific information for different instances.

Refer to caption
Figure 4: Per-category and per-task accuracy on UCF101 and HMDB51. ‘FC’ and ‘ISP’ refer to Frame Condensing and Instance-Specific Prompt, respectively. The plots in (a) and (b) indicate that the introduction of prompt improves accuracy in almost all tasks. The results in (c) and (d) show that our Instance-Specific Prompt significantly improves performance from Frame Condensing in all categories on both datasets.

Per-category and per-task analysis.

We depict the task-wise and category-wise accuracy with different forms of stored frames in Figure 4, i.e., "All Frames", "Frame Condensing without Instance-Specific Prompt" and "Frame Condensing with Instance-Specific Prompt". As expected, integrating multiple frames into one condensed frame does undermine performance in almost all cases unless Instance-Specific Prompt is introduced. As shown in Figure 4(a) and (b), the gains on incremental tasks yielded by Instance-Specific Prompt mainly fall in the later tasks, which implies that the spatio-temporal knowledge absorbed by prompt can effectively alleviate forgetting. Meanwhile, from Figure 4(c) and (d), we also observe that Frame Condensing can lead to the performance degradation of actions with large temporal and spatial variations or short duration, such as "CliffDiving", "Ride horse" and "Hug", etc. Therefore, Frame Condensing with only a few parameters is difficult to capture these complex spatio-temporal interactions. However, Instance-Specific Prompt helps alleviate this issue since it recovers the lost spatio-temporal details in Frame Condensing for each instance.

5 Discussions

Limitations.

Compared with the previous video incremental learning approaches, FrameMaker provides a simple framework that significantly reduces the amount of memory required for each exemplar video. Nevertheless, FrameMaker may fail when the memory budget is highly constrained, which needs to be further explored in future work.

Societal Impacts.

Although FrameMaker only stores condensed frames in memory buffer for experience replay, which still retains some information about the original video. Therefore, FrameMaker may be used in applications with privacy concerns [48].

Conclusion.

This paper proposes FrameMaker, a memory-efficient approach for video class-incremental learning. It explores how to learn an effective condensed frame with less memory for each video that can be applied to the future incremental task. FrameMaker mainly consists of two key components, i.e., Frame Condensing and Instance-Specific Prompt. The former learns a group of condensed weights for each video to integrate multiple frames into one condensed frame, while the latter learns to retrieve the collapsed spatio-temporal structure for the condensed frame. In this way, FrameMaker offers better results with only 20% memory consumption compared to recent advanced methods and sets a new state-of-the-art class-incremental learning performance on multiple challenging datasets. Memory-based class-incremental learning is one of the effective methods to fight forgetting. The efficient utilization of limited space is worth exploring in future work.

Acknowledgments.

This work is supported by the National Key R&D Program of China under Grant No. 2018AAA0101501, the National Natural Science Foundation of China undergrant 61871435 and 62272380, Fundamental Research Funds for the Central Universities no.2019kfyXKJC024, the Science and Technology Program of Xi’an, China under Grant 21RGZN0017, and by Alibaba Group through Alibaba Innovative Research Program.

References

  • Aljundi et al. [2018] Rahaf Aljundi, Francesca Babiloni, Mohamed Elhoseiny, Marcus Rohrbach, and Tinne Tuytelaars. Memory aware synapses: Learning what (not) to forget. In Proceedings of the European Conference on Computer Vision, pages 139–154, 2018.
  • Arnab et al. [2021] Anurag Arnab, Mostafa Dehghani, Georg Heigold, Chen Sun, Mario Lučić, and Cordelia Schmid. Vivit: A video vision transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6836–6846, 2021.
  • Bahng et al. [2022] Hyojin Bahng, Ali Jahanian, Swami Sankaranarayanan, and Phillip Isola. Visual prompting: Modifying pixel space to adapt pre-trained models. arXiv preprint arXiv:2203.17274, 2022.
  • Bertasius et al. [2021] Gedas Bertasius, Heng Wang, and Lorenzo Torresani. Is space-time attention all you need for video understanding. arXiv preprint arXiv:2102.05095, 2(3):4, 2021.
  • Buzzega et al. [2020] Pietro Buzzega, Matteo Boschini, Angelo Porrello, Davide Abati, and Simone Calderara. Dark experience for general continual learning: a strong, simple baseline. Advances in neural information processing systems, 33:15920–15930, 2020.
  • Carreira and Zisserman [2017] Joao Carreira and Andrew Zisserman. Quo vadis, action recognition? a new model and the kinetics dataset. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 6299–6308, 2017.
  • Castro et al. [2018] Francisco M Castro, Manuel J Marín-Jiménez, Nicolás Guil, Cordelia Schmid, and Karteek Alahari. End-to-end incremental learning. In Proceedings of the European conference on computer vision, pages 233–248, 2018.
  • Deng et al. [2009] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
  • Dhar et al. [2019] Prithviraj Dhar, Rajat Vikram Singh, Kuan-Chuan Peng, Ziyan Wu, and Rama Chellappa. Learning without memorizing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5138–5146, 2019.
  • Diba et al. [2018] Ali Diba, Mohsen Fayyaz, Vivek Sharma, M Mahdi Arzani, Rahman Yousefzadeh, Juergen Gall, and Luc Van Gool. Spatio-temporal channel correlation networks for action classification. In Proceedings of the European Conference on Computer Vision, pages 284–299, 2018.
  • Dosovitskiy et al. [2020] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
  • Douillard et al. [2020] Arthur Douillard, Matthieu Cord, Charles Ollion, Thomas Robert, and Eduardo Valle. Podnet: Pooled outputs distillation for small-tasks incremental learning. In European Conference on Computer Vision, pages 86–102. Springer, 2020.
  • Fan et al. [2021] Haoqi Fan, Bo Xiong, Karttikeya Mangalam, Yanghao Li, Zhicheng Yan, Jitendra Malik, and Christoph Feichtenhofer. Multiscale vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6824–6835, 2021.
  • Fan et al. [2018] Hehe Fan, Zhongwen Xu, Linchao Zhu, Chenggang Yan, Jianjun Ge, and Yi Yang. Watching a small portion could be as good as watching all: Towards efficient video classification. In IJCAI International Joint Conference on Artificial Intelligence, 2018.
  • Feichtenhofer [2020] Christoph Feichtenhofer. X3d: Expanding architectures for efficient video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 203–213, 2020.
  • Feichtenhofer et al. [2019] Christoph Feichtenhofer, Haoqi Fan, Jitendra Malik, and Kaiming He. Slowfast networks for video recognition. In Proceedings of the IEEE/CVF international conference on computer vision, pages 6202–6211, 2019.
  • Gowda et al. [2021] Shreyank N Gowda, Marcus Rohrbach, and Laura Sevilla-Lara. Smart frame selection for action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 1451–1459, 2021.
  • Goyal et al. [2017] Raghav Goyal, Samira Ebrahimi Kahou, Vincent Michalski, Joanna Materzynska, Susanne Westphal, Heuna Kim, Valentin Haenel, Ingo Fruend, Peter Yianilos, Moritz Mueller-Freitag, et al. The" something something" video database for learning and evaluating visual common sense. In Proceedings of the IEEE international conference on computer vision, pages 5842–5850, 2017.
  • He et al. [2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  • Hinton et al. [2015] Geoffrey Hinton, Oriol Vinyals, Jeff Dean, et al. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2(7), 2015.
  • Hou et al. [2019] Saihui Hou, Xinyu Pan, Chen Change Loy, Zilei Wang, and Dahua Lin. Learning a unified classifier incrementally via rebalancing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 831–839, 2019.
  • Hu et al. [2019] Han Hu, Zheng Zhang, Zhenda Xie, and Stephen Lin. Local relation networks for image recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.
  • Huang et al. [2018] De-An Huang, Vignesh Ramanathan, Dhruv Mahajan, Lorenzo Torresani, Manohar Paluri, Li Fei-Fei, and Juan Carlos Niebles. What makes a video a video: Analyzing temporal information in video understanding models and datasets. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 7366–7375, 2018.
  • Jia et al. [2022] Menglin Jia, Luming Tang, Bor-Chun Chen, Claire Cardie, Serge Belongie, Bharath Hariharan, and Ser-Nam Lim. Visual prompt tuning. arXiv preprint arXiv:2203.12119, 2022.
  • Kim et al. [2022] Kiyoon Kim, Shreyank N Gowda, Oisin Mac Aodha, and Laura Sevilla-Lara. Capturing temporal information in a single frame: Channel sampling strategies for action recognition. arXiv preprint arXiv:2201.10394, 2022.
  • Kirkpatrick et al. [2017] James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, et al. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13):3521–3526, 2017.
  • Korbar et al. [2019] Bruno Korbar, Du Tran, and Lorenzo Torresani. Scsampler: Sampling salient clips from video for efficient action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6232–6242, 2019.
  • Kuehne et al. [2011] Hildegard Kuehne, Hueihan Jhuang, Estíbaliz Garrote, Tomaso Poggio, and Thomas Serre. Hmdb: a large video database for human motion recognition. In 2011 International conference on computer vision, pages 2556–2563. IEEE, 2011.
  • Lester et al. [2021] Brian Lester, Rami Al-Rfou, and Noah Constant. The power of scale for parameter-efficient prompt tuning. arXiv preprint arXiv:2104.08691, 2021.
  • Li et al. [2022] Kunchang Li, Yali Wang, Peng Gao, Guanglu Song, Yu Liu, Hongsheng Li, and Yu Qiao. Uniformer: Unified transformer for efficient spatiotemporal representation learning. arXiv preprint arXiv:2201.04676, 2022.
  • Li and Liang [2021] Xiang Lisa Li and Percy Liang. Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190, 2021.
  • Li and Hoiem [2017a] Zhizhong Li and Derek Hoiem. Learning without forgetting. IEEE transactions on pattern analysis and machine intelligence, 40(12):2935–2947, 2017a.
  • Li and Hoiem [2017b] Zhizhong Li and Derek Hoiem. Learning without forgetting. IEEE transactions on pattern analysis and machine intelligence, 40(12):2935–2947, 2017b.
  • Lin et al. [2019] Ji Lin, Chuang Gan, and Song Han. Tsm: Temporal shift module for efficient video understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7083–7093, 2019.
  • Liu et al. [2022a] Chengxu Liu, Huan Yang, Jianlong Fu, and Xueming Qian. Learning trajectory-aware transformer for video super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022a.
  • Liu et al. [2021a] Pengfei Liu, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. arXiv preprint arXiv:2107.13586, 2021a.
  • Liu et al. [2021b] Ze Liu, Jia Ning, Yue Cao, Yixuan Wei, Zheng Zhang, Stephen Lin, and Han Hu. Video swin transformer. arXiv preprint arXiv:2106.13230, 2021b.
  • Liu et al. [2022b] Ze Liu, Han Hu, Yutong Lin, Zhuliang Yao, Zhenda Xie, Yixuan Wei, Jia Ning, Yue Cao, Zheng Zhang, Li Dong, Furu Wei, and Baining Guo. Swin transformer v2: Scaling up capacity and resolution. In International Conference on Computer Vision and Pattern Recognition, 2022b.
  • McCloskey and Cohen [1989] Michael McCloskey and Neal J Cohen. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of learning and motivation, volume 24, pages 109–165. Elsevier, 1989.
  • Ostapenko et al. [2019] Oleksiy Ostapenko, Mihai Puscas, Tassilo Klein, Patrick Jahnichen, and Moin Nabi. Learning to remember: A synaptic plasticity driven framework for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11321–11329, 2019.
  • Park et al. [2021] Jaeyoo Park, Minsoo Kang, and Bohyung Han. Class-incremental learning for action recognition in videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 13698–13707, 2021.
  • Paszke et al. [2019] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
  • Qiu et al. [2017] Zhaofan Qiu, Ting Yao, and Tao Mei. Learning spatio-temporal representation with pseudo-3d residual networks. In proceedings of the IEEE International Conference on Computer Vision, pages 5533–5541, 2017.
  • Qiu et al. [2021] Zhaofan Qiu, Ting Yao, Yan Shu, Chong-Wah Ngo, and Tao Mei. Condensing a sequence to one informative frame for video recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 16311–16320, 2021.
  • Rebuffi et al. [2017] Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, and Christoph H Lampert. icarl: Incremental classifier and representation learning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pages 2001–2010, 2017.
  • Reddy and Shah [2013] Kishore K Reddy and Mubarak Shah. Recognizing 50 human action categories of web videos. Machine vision and applications, 24(5):971–981, 2013.
  • Shin et al. [2017] Hanul Shin, Jung Kwon Lee, Jaehong Kim, and Jiwon Kim. Continual learning with deep generative replay. Advances in neural information processing systems, 30, 2017.
  • Shokri and Shmatikov [2015] Reza Shokri and Vitaly Shmatikov. Privacy-preserving deep learning. In Proceedings of the 22nd ACM SIGSAC conference on computer and communications security, pages 1310–1321, 2015.
  • Simonyan and Zisserman [2014] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
  • Soomro et al. [2012] Khurram Soomro, Amir Roshan Zamir, and Mubarak Shah. Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402, 2012.
  • Tran et al. [2015] Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE international conference on computer vision, pages 4489–4497, 2015.
  • Tran et al. [2018] Du Tran, Heng Wang, Lorenzo Torresani, Jamie Ray, Yann LeCun, and Manohar Paluri. A closer look at spatiotemporal convolutions for action recognition. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pages 6450–6459, 2018.
  • Tran et al. [2019] Du Tran, Heng Wang, Lorenzo Torresani, and Matt Feiszli. Video classification with channel-separated convolutional networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5552–5561, 2019.
  • Villa et al. [2022] Andrés Villa, Kumail Alhamoud, Juan León Alcázar, Fabian Caba Heilbron, Victor Escorcia, and Bernard Ghanem. vclimb: A novel video class incremental learning benchmark. arXiv preprint arXiv:2201.09381, 2022.
  • Wang et al. [2019] Junbo Wang, Wei Wang, Zhiyong Wang, Liang Wang, Dagan Feng, and Tieniu Tan. Stacked memory network for video summarization. In Proceedings of the 27th ACM International Conference on Multimedia, pages 836–844, 2019.
  • Wang et al. [2018] Limin Wang, Wei Li, Wen Li, and Luc Van Gool. Appearance-and-relation networks for video classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1430–1439, 2018.
  • Wang et al. [2021] Zifeng Wang, Zizhao Zhang, Chen-Yu Lee, Han Zhang, Ruoxi Sun, Xiaoqi Ren, Guolong Su, Vincent Perot, Jennifer Dy, and Tomas Pfister. Learning to prompt for continual learning. arXiv preprint arXiv:2112.08654, 2021.
  • Wang et al. [2022] Zifeng Wang, Zizhao Zhang, Sayna Ebrahimi, Ruoxi Sun, Han Zhang, Chen-Yu Lee, Xiaoqi Ren, Guolong Su, Vincent Perot, Jennifer Dy, et al. Dualprompt: Complementary prompting for rehearsal-free continual learning. arXiv preprint arXiv:2204.04799, 2022.
  • Wu et al. [2019a] Wenhao Wu, Dongliang He, Xiao Tan, Shifeng Chen, and Shilei Wen. Multi-agent reinforcement learning based frame sampling for effective untrimmed video recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6222–6231, 2019a.
  • Wu et al. [2019b] Yue Wu, Yinpeng Chen, Lijuan Wang, Yuancheng Ye, Zicheng Liu, Yandong Guo, and Yun Fu. Large scale incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 374–382, 2019b.
  • Wu et al. [2019c] Zuxuan Wu, Caiming Xiong, Chih-Yao Ma, Richard Socher, and Larry S Davis. Adaframe: Adaptive frame selection for fast video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1278–1287, 2019c.
  • Xie et al. [2017] Saining Xie, Chen Sun, Jonathan Huang, Zhuowen Tu, and Kevin Murphy. Rethinking spatiotemporal feature learning for video understanding. arXiv preprint arXiv:1712.04851, 1(2):5, 2017.
  • Zenke et al. [2017] Friedemann Zenke, Ben Poole, and Surya Ganguli. Continual learning through synaptic intelligence. In International Conference on Machine Learning, pages 3987–3995. PMLR, 2017.
  • Zhao et al. [2021a] Bin Zhao, Haopeng Li, Xiaoqiang Lu, and Xuelong Li. Reconstructive sequence-graph network for video summarization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(5):2793–2801, 2021a.
  • Zhao et al. [2021b] Hanbin Zhao, Xin Qin, Shihao Su, Yongjian Fu, Zibo Lin, and Xi Li. When video classification meets incremental classes. In Proceedings of the 29th ACM International Conference on Multimedia, pages 880–889, 2021b.
  • Zhi et al. [2021] Yuan Zhi, Zhan Tong, Limin Wang, and Gangshan Wu. Mgsampler: An explainable sampling strategy for video action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1513–1522, 2021.
  • Zhu et al. [2020] Wencheng Zhu, Jiwen Lu, Jiahao Li, and Jie Zhou. Dsnet: A flexible detect-to-summarize network for video summarization. IEEE Transactions on Image Processing, 30:948–962, 2020.

Checklist

The checklist follows the references. Please read the checklist guidelines carefully for information on how to answer these questions. For each question, change the default [TODO] to [Yes] , [No] , or [N/A] . You are strongly encouraged to include a justification to your answer, either by referencing the appropriate section of your paper or providing a brief inline description. For example:

  • Did you include the license to the code and datasets? [N/A]

Please do not modify the questions and only use the provided macros for your answers. Note that the Checklist section does not count towards the page limit. In your paper, please delete this instructions block and only keep the Checklist section heading above along with the questions/answers below.

  1. 1.

    For all authors…

    1. (a)

      Do the main claims made in the abstract and introduction accurately reflect the paper’s contributions and scope? [Yes]

    2. (b)

      Did you describe the limitations of your work? [Yes]

    3. (c)

      Did you discuss any potential negative societal impacts of your work? [Yes]

    4. (d)

      Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes]

  2. 2.

    If you are including theoretical results…

    1. (a)

      Did you state the full set of assumptions of all theoretical results? [N/A]

    2. (b)

      Did you include complete proofs of all theoretical results? [N/A]

  3. 3.

    If you ran experiments…

    1. (a)

      Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [No]

    2. (b)

      Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes]

    3. (c)

      Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [No]

    4. (d)

      Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes]

  4. 4.

    If you are using existing assets (e.g., code, data, models) or curating/releasing new assets…

    1. (a)

      If your work uses existing assets, did you cite the creators? [Yes]

    2. (b)

      Did you mention the license of the assets? [No]

    3. (c)

      Did you include any new assets either in the supplemental material or as a URL? [No]

    4. (d)

      Did you discuss whether and how consent was obtained from people whose data you’re using/curating? [No]

    5. (e)

      Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [No]

  5. 5.

    If you used crowdsourcing or conducted research with human subjects…

    1. (a)

      Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A]

    2. (b)

      Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A]

    3. (c)

      Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]

Appendix A Additional Experiments for Different Memory Budgets

This is supplementary to Section 4.2 "Comparison with State-of-the-art Results". Limited by space, we only summarize the results of our approach using the same exemplar memory size per class as others and the experiment results of "Memory budget comparison" on UCF101 with 10 steps in the main body. Table 10 supplements the average accuracy of FrameMaker with different memory budgets under all benchmarks of UCF101 and HMDB51 datasets, which is compared with the best performance of TCD [41].

Table 10: Comparison with the state-of-the-art approaches over class-incremental action recognition performance on UCF101 and HMDB51 with different memory budgets.
UCF101 HMDB51
Num. of Classes 10×510\times 5 stages 5×105\times 10 stages 2×252\times 25 stages 5×55\times 5 stages 1×251\times 25 stages
Memory Per Class CNN NME CNN NME CNN NME CNN NME CNN NME
8F×5V8\text{F}\times 5\text{V}=6.0Mb [41] 74.89 77.16 73.43 75.35 72.19 74.01 45.34 50.36 40.47 46.66
1F×1V1\text{F}\times 1\text{V}=0.15Mb 58.75 73.97 49.37 70.78 48.31 65.86 36.08 43.82 32.53 39.00
1F×2V1\text{F}\times 2\text{V}=0.3Mb 67.59 76.18 62.06 74.18 55.84 72.88 39.33 46.07 34.05 43.18
1F×5V1\text{F}\times 5\text{V}=0.75Mb 73.05 76.70 72.93 76.64 68.79 75.36 43.39 46.88 38.95 44.18
1F×8V1\text{F}\times 8\text{V}=1.2Mb 76.09 77.88 73.64 76.98 73.57 76.76 46.25 49.02 40.80 46.97
1F×16V1\text{F}\times 16\text{V}=2.4Mb 77.24 78.06 75.19 77.43 74.92 76.92 46.97 50.42 41.96 47.03
1F×40V1\text{F}\times 40\text{V}=6.0Mb 78.13 78.64 76.38 78.14 75.77 77.49 47.54 51.12 42.65 47.37

As can be seen from Table 10, under other settings, the results of our method are similar to that on UCF101 with 10 steps. When using about 20% of the storage space, our method can achieve better performance than TCD, and the accuracy is further improved when the number of saved exemplars is increased, which also proves the robustness of our approach.

Appendix B Additional Analysis of the Balance Weights

This is supplementary to Section 4.3 "Ablation Study". We further discuss the sensitivity to the balance weight for each term in LfcL_{\text{fc}} on UCF101 with 10 steps.

We first discuss the sensitivity of α{\alpha} and β{\beta}, the balance weights for Lfc{L^{\text{c}}_{\text{f}}} and LcecL^{\text{c}}_{\text{ce}} which are used to achieve stronger constraint on WikW_{i}^{k} for its effective training. Table 11 shows the performance of frame condensing (without instance-specific prompt) under various combinations of α{\alpha} and β{\beta}. We find the performance under the combination of {α=1,β=1}\{{\alpha}=1,{\beta}=1\} consistently exceeds others. Although a slightly higher performance is obtained on NME under {α=2,β=1}\{{\alpha}=2,{\beta}=1\}, it is difficult to achieve the same gain on the two different metrics after continuous adjustment.

Table 11: Sensitivity of the performance of Frame Condensing to α{\alpha} and β{\beta} on UCF101 with 10 steps. Default settings are marked in gray.
α(Lfc){\alpha}({L^{\text{c}}_{\text{f}})} 1 2 5 0.5 0.1 0.01 1 1 1 1 1
β(Lcec){\beta}(L^{\text{c}}_{\text{ce}}) 1 1 1 1 1 1 2 5 0.5 0.1 0.01
CNN 72.29 72.27 71.83 72.22 72.07 71.94 71.93 71.44 72.14 71.96 71.92
NME 76.42 76.43 76.36 76.25 76.21 76.13 76.37 76.23 76.22 75.91 75.84

Then we further analyze sensitivity of γ{\gamma} and η{\eta}, the balance weights for Lfp{L^{\text{p}}_{\text{f}}} and LcepL^{\text{p}}_{\text{ce}}, when setting α{\alpha} and β{\beta} as the optimal value, 11. Table 12 shows the performance of FrameMaker (frame condensing with instance-specific prompt) under various combinations of γ{\gamma} and η{\eta}.

Table 12: Sensitivity of the performance of FrameMaker to γ{\gamma} and η{\eta} (α=1,β=1)({\alpha}=1,{\beta}=1) on UCF101 with 10 steps. Default settings are marked in gray.
γ(Lfp){\gamma}({L^{\text{p}}_{\text{f}}}) 1 2 5 0.5 0.1 0.01 1 1 1 1 1
η(Lcep){\eta}(L^{\text{p}}_{\text{ce}}) 1 1 1 1 1 1 2 5 0.5 0.1 0.01
CNN 72.93 72.94 72.92 72.88 72.64 72.67 72.90 72.85 72.92 72.80 72.73
NME 76.64 76.62 76.58 76.61 76.59 76.61 76.65 76.58 76.61 76.55 76.51

We find the performance under the combination of {γ=1,η=1}\{{\gamma}=1,{\eta}=1\} always exceeds others. Some groups of weights make one of the test metrics higher but only {γ=1,η=1}\{{\gamma}=1,{\eta}=1\} obtain the best balance of them. Therefore, we set all the balance weights as 11 finally in our experiments.

Appendix C Implementation Details

This section provides some additional details about the experiments in the main body.

Training Details.

This is supplementary to Section 3.3 "Training". We further supplement the elaborate training process and knowledge distillation loss in this section. Figure 5 shows the overall framework of training process.

Refer to caption
Figure 5: An overview of the training process

In the training process of the incremental step kk, we use the dataset Dk=DkMk1{D^{\prime}}^{k}=D^{k}\bigcup{M^{k-1}} to update the model from F(;Θk1)F(;\Theta_{k-1}), where DkD^{k} is the dataset of the task kk which consists of videos belonging to current classes, and Mk1{M^{k-1}} is memory bank which contains condensed frames generated by FrameMaker after the step k1k-1. The model classifies the samples from the two datasets respectively and use exemplars in old classes for knowledge distillation.

For the classification task, we adopt cross entropy loss to supervise. LcedL^{\text{d}}_{{\text{ce}}} and LcemL^{\text{m}}_{{\text{ce}}} are the cross entropy losses for video example VikV_{i}^{k} in DkD^{k} and condensed frame exemplar Iik1I_{i}^{k-1} in Mk1{M^{k-1}}.

Lced=CrossEntropy(F(Vik;Θk),yik),L^{\text{d}}_{{\text{ce}}}={\text{CrossEntropy}(F(V_{i}^{k};\Theta_{k}),y_{i}^{k})}\ , (9)
Lcem=CrossEntropy(F(Iik1;Θk),yik1),L^{\text{m}}_{{\text{ce}}}={\text{CrossEntropy}(F(I_{i}^{k-1};\Theta_{k}),y_{i}^{k-1})}\ , (10)

where yiky_{i}^{k} and yik1y_{i}^{k-1} is the ground truth label of VikV_{i}^{k} and Iik1I_{i}^{k-1}, respectively.

In order to save the information of old classes better, the knowledge distillation function of PODNet [12] is employed on the old classes data Mk1{M^{k-1}} and it consists of two components, the spatial distillation function LspatialL_{\text{spatial}} and the final embedding distillation function LflatL_{\text{flat}}.

Lspatial(fl(Iik1;Θk),fl(Iik1;Θk1))=c=1Ch=1Hw=1Wfl,c,w,h(Iik1;Θk)w=1Wfl,c,w,h(Iik1;Θk1)2+c=1Cw=1Wh=1Hfl,c,w,h(Iik1;Θk)h=1Hfl,c,w,h(Iik1;Θk1)2,\begin{split}L_{\text{spatial}}&{(f_{l}(I_{i}^{k-1};\Theta_{k}),f_{l}(I_{i}^{k-1};\Theta_{k-1}))}\\ &=\sum_{c=1}^{C}\sum_{h=1}^{H}{\mid\mid\sum_{w=1}^{W}f_{l,c,w,h}(I_{i}^{k-1};\Theta_{k})-\sum_{w=1}^{W}f_{l,c,w,h}(I_{i}^{k-1};\Theta_{k-1})\mid\mid^{2}}\\ &+\sum_{c=1}^{C}\sum_{w=1}^{W}{\mid\mid\sum_{h=1}^{H}f_{l,c,w,h}(I_{i}^{k-1};\Theta_{k})-\sum_{h=1}^{H}f_{l,c,w,h}(I_{i}^{k-1};\Theta_{k-1})\mid\mid^{2}}\ ,\end{split} (11)

where fl(;Θk)f_{l}(;\Theta_{k}) is the output of intermediate convolution layer ll of model F(;Θk)F(;\Theta_{k}). C, H, W are the size of channel, height and width of the output feature of layer ll.

Lflat(f(Iik1;Θk),f(Iik1;Θk1))=f(Iik1;Θk)f(Iik1;Θk1)2,L_{\text{flat}}{(f(I_{i}^{k-1};\Theta_{k}),f(I_{i}^{k-1};\Theta_{k-1}))}=\mid\mid f(I_{i}^{k-1};\Theta_{k})-f(I_{i}^{k-1};\Theta_{k-1})\mid\mid^{2}\ , (12)

where f(;Θk)f(;\Theta_{k}) is the feature extractor of model F(;Θk)F(;\Theta_{k}).

The final distillation loss of the condensed frame exemplars is given by:

Ldistm=Lspatial(fl(Iik1;Θk),fl(Iik1;Θk1))+Lflat(f(Iik1;Θk),f(Iik1;Θk1)),L^{\text{m}}_{{\text{dist}}}=L_{\text{spatial}}{(f_{l}(I_{i}^{k-1};\Theta_{k}),f_{l}(I_{i}^{k-1};\Theta_{k-1}))}+L_{\text{flat}}{(f(I_{i}^{k-1};\Theta_{k}),f(I_{i}^{k-1};\Theta_{k-1}))}\ , (13)

To simplify, the hyperparameters used in PODNet [12] final distillation loss is ignored.

The total objective function for task kk is given by:

Lcil=Lced+Lcem+Ldistm,L_{\text{cil}}={L^{\text{d}}_{{\text{ce}}}+L^{\text{m}}_{{\text{ce}}}+L^{\text{m}}_{{\text{dist}}}}\ , (14)

which is the same as Eq. 8 in the manuscript.

Dataset Details.

Following TCD [41], we use three action recognition datasets, UCF101 [50], HMDB51 [28] and Something-Something V2 [18]. The HMDB51 dataset is a large collection of realistic videos from various sources which can be found in https://serre-lab.clps.brown.edu/resource/hmdb-a-large-human-motion-database/. UCF101 dataset is an extension of UCF50 [46] and consists of 13,320 video clips, which can be found in https://www.crcv.ucf.edu/data/UCF101.php. Something-SomethingV2 dataset is a large-scale motion-sensitive dataset that shows humans performing pre-defined basic actions with everyday objects, which can be found in https://developer.qualcomm.com/software/ai-datasets/something-something. For UCF101, we have three settings, 51+10×551+10\times 5 stages, 51+5×1051+5\times 10 stages and 51+1×2551+1\times 25 stages. For HMDB51, the number of initial task classes is 26, and the remaining 25 classes are separated into 5 or 25 groups. For Something-Something V2, the benchmarks are 84+10×984+10\times 9 stages and 84+5×1884+5\times 18 stages. The random seed we use is also the same as TCD [41].

Model and Hyperparameter Setup.

TSM [34] is employed as our backbone, and we follow the data preprocessing procedure and the basic training setting of TSM. We train a ResNet-34 TSM for UCF101 and a ResNet-50 TSM for HMDB51 and Something-Something V2. Each incremental training procedure takes 50 epochs. We use SGD optimizer with an initial learning rate of 0.04 for UCF101 and Something-Something V2, and 0.002 for HMDB51, which is reduced by half after 25 epochs. For incremental steps, we set the learning rate for all datasets as 0.001 and make the same reduction. The learning rate is set as 0.01 and 0.001 for condensing weights and the instance-specific prompt, respectively, and the total iteration for their optimization is set to be 8k.

Hardware Information.

We train all models on eight NVIDIA V100 GPUs and use PyTorch [42] for all our experiments.

Appendix D Visualizations

Refer to caption
Figure 6: Channel comparisons between different inputs on HMDB51 . Each line indicates the feature channel activation value obtained with the specified frame as input. FC and ISP are abbreviations of our proposed Frame Condensing and Instance-Specific Prompt, respectively.
Refer to caption
Figure 7: Visualizations of condensed frames and GradCAM maps on HMDB51. FC and ISP refer to proposed Frame Condensing and Instance-Specific Prompt, respectively.

Visual comparison of features. Figure 6 plots visual comparisons of the extracted features with different frames as model inputs. We can summarize that the average frame shows the largest feature difference compared to the features extracted from the original clips. Our proposed Frame Condensing and Instance-Specific Prompt can effectively reduce the feature differences with the original clips and thus effectively improve the feature quality of condensed frames.

Visualizations of condensed frames. Figure 7 plots the condensed frames, the sum of the condensed frames and their prompts, and the activation region generated by the GradCAM technology. From the results, we observe the learned prompts have very small magnitudes (aver- age value is 0.03), hence no significant difference is observed actually. However, from the GradCAM attention maps, we can observe the condensed frames without prompts may focus on background regions, while the learned prompts tend to increase the attention of motion regions or reduce the attention of the backgrounds. We think the reason behind this is that the collapsed temporal dynamics may easily lead the model pay more attention to the background to some extent, while the prompt can guide the model to refocus on the motion region and weaken the interference from backgrounds.

Appendix E Training Prompt without Condensed Frame

Refer to caption
Figure 8: Training prompts without condensed frames. (a): Per-task accuracy on HMDB51. The performance of Instance-Specific Prompt without condensed frames drops quickly. (b): The training loss of prompts on HMDB51. Instance-Specific Prompt without condensed frames is hard to optimize. (c) and (d): Learned prompts of two videos. There is no intuitive semantic information in prompt.

We attempt to train Instance-Specific Prompt with out condensed frames, and the learned prompting parameters are visualized in Figure 8(c) and (d). We observe that the learned prompts have no intuitive semantics. To further understand this phenomenon, we plot the per-task accuracy and the loss curve of the prompts in Figure 8(a) and (b), respectively. We can summarize two points: (i) the learned prompts without condensed frames cannot fight against catastrophic forgetting, which indicates that the semantic information of the learned prompts is weak; (ii) the loss curve of the prompts without condensed frames is much higher than that with condensed frames, which implies that it is difficult to train a single prompts from scratch. Thus, it is crucial that the condensed frames provide a good initialization for the prompts to avoid the local minima of prompts.