Learning a Condensed Frame for Memory-Efficient Video Class-Incremental Learning
Abstract
Recent incremental learning for action recognition usually stores representative videos to mitigate catastrophic forgetting. However, only a few bulky videos can be stored due to the limited memory. To address this problem, we propose FrameMaker, a memory-efficient video class-incremental learning approach that learns to produce a condensed frame for each selected video. Specifically, FrameMaker is mainly composed of two crucial components: Frame Condensing and Instance-Specific Prompt. The former is to reduce the memory cost by preserving only one condensed frame instead of the whole video, while the latter aims to compensate the lost spatio-temporal details in the Frame Condensing stage. By this means, FrameMaker enables a remarkable reduction in memory but keep enough information that can be applied to following incremental tasks. Experimental results on multiple challenging benchmarks, i.e., HMDB51, UCF101 and Something-Something V2, demonstrate that FrameMaker can achieve better performance to recent advanced methods while consuming only 20% memory. Additionally, under the same memory consumption conditions, FrameMaker significantly outperforms existing state-of-the-arts by a convincing margin.
1 Introduction
Training video action recognition models with all classes in a single fine-tuning stage have been widely studied [6, 34, 51, 62, 52, 30, 13, 37] in recent years. However, in many realistic scenarios, limited by privacy matters or technologies, the different classes can only be presented in sequence, where the previously trained classes are either unavailable or partially available in limited memory. Naively training models will overfit currently available data and cause performance deterioration on previously seen classes, which is named catastrophic forgetting [39]. Class-incremental learning is a machine learning paradigm to fight this challenge when fine-tuning a single model in a sequence of independent classes.
A branch of class-incremental learning methods [7, 9, 12, 21, 45, 33, 60, 5] has achieved remarkable performances in the image domain by re-training a portion of past examples. Meanwhile, some existing video incremental learning methods [41, 54] have demonstrated that storing more past examples can improve the ability to mitigate forgetting. Nevertheless, videos are information-redundant and require a high memory load to be saved, hence it is impractical to store a good deal of training videos for each class. Therefore, herding strategy [45] is adopted to select a small set of representative videos for the exemplar memory [41, 54]. vCLIMB [54] tries to down-sample frames for each video to reduce the memory consumption, and imposes a temporal-consistency regularization for better performance. Despite the remarkable performance, these methods still demand to conserve multiple (i.e., 8-16) frames for each representative video, leading to non-negligible memory overhead, which limits their further real-world applications.

To remedy the above weaknesses, we present FrameMaker, a memory-efficient video class-incremental learning approach. Concretely, the proposed FrameMaker contains two components, Frame Condensing and Instance-Specific Prompt. In Frame Condensing, FrameMaker assigns a learnable weight to each frame, and then these weights are adjusted to make the weighted sum frame i.e., the condensed frame, share the same embedding features with the original video clip. Obviously, the collapsed temporal dimension and mixed spatial cues by Frame Condensing may degrade the accuracy of action recognition. To compensate the missing spatio-temporal details, inspired by Task-Specific Prompt [36], we further propose Instance-Specific Prompt, which learns a set of parameters for each condensed frame to perform fine-grained pixel-wise modifications. In this way, the memory required to store a single video can be remarkably compressed. An intuitive comparison is given in Figure 1(a). On the premise of comparable performance with TCD [41], FrameMaker significantly reduces memory overhead by 80%. Moreover, the comparison curve displayed in Figure 1(b) further indicates that FrameMaker achieves a consistent performance improvement over recent TCD with the same memory consumption.
In summary, we make the following contributions: 1) We present a memory-efficient video class-incremental learning method FrameMaker, which significantly reduces the memory consumption of example videos by integrating multiple frames into one condensed frame; 2) For the first time, we show a novel perspective of Prompt, i.e., Instance-Specific Prompt, and demonstrate that it can embed effective spatio-temporal interactions into learned condensed frames for video class-incremental learning; 3) Experimental results reveal that FrameMaker consistently outperforms recent state-of-the-art methods with the same memory overhead on three challenging benchmarks (HMDB51, UCF101, and Something-Something V2).
2 Related Works
2.1 Class-Incremental Learning
The class-incremental learning has been well-studied in image domain [1, 32, 26, 63], existing approaches could be grouped into knowledge distillation-based methods [20], memory-based methods [40, 47, 45], and architecture-based methods. In video domain, class-incremental learning for action recognition is still an underexplored area. The existing solutions are designed for better temporal constraints, such as decomposing the spatio-temporal features [65], exploiting time-channel importance maps [41], and applying temporal consistency regularization [54]. Notably, both [41] and [54] have shown that more stored examples in memory can effectively encourage the performance. However, FrameMaker attempts to minimize the stored frames for each video by a pretty maneuver, which can provide decent performance with slight storage requirements.
2.2 Prompt-Based Learning
Prompting is an emerging transfer learning technique in natural language processing [36], which adapts the language model to downstream tasks by a fixed function. Since the design of prompting functions is challenging, the learnable prompts [29, 31] are proposed to learn task-specific knowledge parameters for the input. Recently, in the visual domain, some works have also emerged to adapt the pre-trained models to new tasks by adding prompt tokens [24] or pixels [3] in the data space. Besides, DualPrompt [58] and L2P [57] get rid of the rehearsal buffer in continual learning by learning a set of task-specific parameters. In this paper, we draw inspiration from prompting, and present an Instance-Specific Prompt to replenish the collapsed spatio-temporal cues for condensed frames.
2.3 Action Recognition
The approaches designed for action recognition task can be grouped into three categories: 2D CNN, 3D CNN and Transformer-based methods. 2D CNN based methods is more efficient [19, 49], while 3D CNN [16, 6, 15, 51, 53, 10, 62, 43, 56, 52] achieves better accuracy in action recognition at the cost of more computation cost. With the rapid development of vision transformer [35, 22, 38], researchers also apply it into the video recognition [2, 4, 37, 13], which is proven to be remarkable effective. In this work, we follow the settings in [41, 65], and employ the Temporal Shift Module (TSM) [34] based on 2D CNN for efficient experiments.
The video summarization and efficient action recognition tasks have provided some methods for key-frame selection and generation, which are related to this work. The video summarization [67, 55, 64] task selects the most representative frames for efficient human understanding, which has a different paradigm with action recognition task, hence we will mainly discuss efficient The efficient video classification task aims to learn a strategy to improve classification accuracy with fewer discriminative frames as input. The approaches designed for this task can be grouped into two categories: rule-based and model-based methods. The former [27, 66] select frames via some pre-defined rules such as motion distribution [66] or channel sampling [25], and the later design a submodel to select or generate key frames [44, 23, 61, 59, 14, 17]. In this paper, we propose model-free frame condensing method for video class-incremental learning, which only learns a few condensing weights and prompting parameters for each instance.

3 Method
This section presents the general framework of FrameMaker as shown in Figure 2, which mainly consists of two crucial components: Frame Condensing and Instance-Specific Prompt. Firstly, Frame Condensing is designed to integrate multiple frames into one frame for efficient storage. Then we introduce Instance-Specific Prompt to compensate the collapsed temporal information and the mixed spatial knowledge. Finally, the specific class-incremental training details for subsequent tasks using Condensed Frames are given.
3.1 Problem formulation
The purpose of video class-incremental learning is to train a model parameterized by as the tasks arrive in a sequence, where is the feature extractor and is the classification head. Each task has its specific dataset whose labels belong to a predefined label set and have not appeared in previous tasks. The memory-based methods [45, 54, 41] have been shown to be effective in preventing catastrophic forgetting. Specifically, after the incremental training step , a memory bank will be established for the dataset in this task to store the representative exemplars or features. Exemplars stored in previous datasets will be used in subsequent incremental tasks to mitigate forgetting. At the incremental step , the model is trained from with and evaluated on all seen classes.
3.2 FrameMaker
Frame Condensing.
We follow the standard protocol of video class-incremental methods [54, 41], which are based on memory-replay strategy and knowledge distillation. Differently, we devotes to reducing the number of stored frames for representative videos to achieve memory-efficient video class-incremental learning. Formally, we first select a subset of video instances by herding strategy [45] from after the incremental step . Given a video , we then sample frames uniformly and integrate them to condense a more representative one. To this end, we define learnable weights for the -th video instance in exemplar set . And the condensed frame can be calculated as follows:
(1) |
where , and are the size of channel, height and width of input frames, respectively. Next, the condensed frame is excepted to empower the same or much similar expressive ability as the original video clip. Hence, the embedding features of the condensed frame extracted from current model should be consistent with the features of the original video clip:
(2) |
where is the feature extractor of current model . This consistency regularization can foster the condensing weights to exploit memory-efficient expressions of video clips. To further improve the adaptability of the condensed frames to the current model, we employ cross entropy loss to supervise the classification confidence from the condensed frame:
(3) |
The full objective function for condensing weights is given by:
(4) |
With the methods described above, we can learn an effective representation, i.e., the condensed frame for each video based on current model.
Instance-Specific Prompt.
However, condensing a video into one frame would inevitably lose the the temporal dynamics and complete spatial information of the original video to a certain extent. Therefore, an intriguing perspective of prompt, termed as Instance-Specific Prompt, is proposed to replenish the missing spatio-temporal cues for the condensed frame.
Specifically, we first construct a learnable prompt for each exemplar video , which shares the same spatial resolution as its original clip. The prompt is then pixel-wise summed with the condensed frame , and learns to pull the embedding feature of the summed frame and the video clip together, which is similar to Eq. 2:
(5) |
Notably, the Eq. 2 for condensing weights is difficult to learn the satisfactory spatio-temporal features due to the collapse of the temporal dimension. Nevertheless, the Eq. 5 can enrich the representations of condensing frames by introducing more flexible learnable parameters in the input space. Besides, the Cross-Entropy loss is also utilized to enhance the semantic perception:
(6) |
Theoretically, the condensing weights and the prompting parameters can be jointly updated by . Interestingly, we observed in our practice that the flexible prompt leads to the condensing weights being under optimized. Hence, we give the final training objective function for frame condensing as:
(7) |
where additional and are added to achieve stronger constraint on for its effective training. , , and are the balance weights for each term, and they are empirically set to 1.0 unless otherwise specified.
After the optimization of condensing weights and prompting parameters, the prompt is added directly to the condensed frame and stored together. And the learned prompts will be frozen when the corresponding task is completed. Therefore, there is no extra memory cost for prompts.
3.3 Training
It is worth noting that FrameMaker aims to condense the stored frames for representative videos, and other class-incremental learning steps still follow the standard pipeline. Specifically, when training the incremental step , we use the dataset to update the model from , where is the dataset of the task which consists of videos belonging to , and is memory bank which contains condensed frames generated by FrameMaker. The training samples from and are alternately fed into the current model according to the proportion of their sample number. To further prevent the catastrophic forgetting, we also employ the knowledge distillation method proposed in PODNet [12] to transfer the knowledge from the previous model to current model . The overall objective function for task reads:
(8) |
where and are the Cross Entropy losses for new task data and condensed frame exemplars , respectively. And is the knowledge distillation loss [12] for condensed frame exemplars.
4 Experiments
4.1 Experimental Setup
Datasets and evaluation metrics.
The proposed FrameMaker is evaluated on three standard action recognition datasets, UCF101 [50], HMDB51 [28] and Something-Something V2 [18]. HMDB51 dataset consists of 6.8K videos belonging to 51 classes from YouTube or other websites. UCF101 dataset contains 13.3K videos from 101 classes. Something-Something V2 is a crowd-sourced dataset that includes 220K videos from 174 classes.
For UCF101, the model is trained on 51 classes first, and the remaining 50 classes are divided into 5, 10 and 25 tasks. For HMDB51, we train the base model using videos from 26 classes, and the remaining 25 classes are separated into 5 or 25 groups. For Something-Something V2, we first train 84 classes in the initial stage, and generate the groups of 10 and 5 classes.
To evaluate the performance of class-incremental learning, we infer the test videos from all seen categories after each task, and finally, report the average accuracy of all tasks. Following [41], two different metrics are reported, i.e., CNN and NME. CNN refers to training a fully-connected layer for extracted features, which is a standard classification protocol. NME is proposed by iCaRL [45], which assigns the labels for test data by comparing the feature embeddings with the mean-of-exemplars.
Implementation details.
TSM [34] is employed as our backbone, and we follow the data preprocessing procedure of TSM. For UCF101, we train a ResNet-34 TSM for 50 epochs with a batch size 256 from an initial learning rate 0.04. For HMDB51 and Something-Something V2, we train a ResNet-50 TSM for 50 epochs with a batch size of 128 from an initial learning rate of 1e-3 and 0.04, respectively. All used networks are first pre-trained on ImageNet [8] for initialization. These settings are consistent with TCD [41]. We train all models on eight NVIDIA V100 GPUs and use PyTorch [42] for all our experiments.
UCF101 | HMDB51 | |||||||||
Num. of Classes | stages | stages | stages | stages | stages | |||||
Classifier | CNN | NME | CNN | NME | CNN | NME | CNN | NME | CNN | NME |
Finetuning | 24.97 | - | 13.45 | - | 5.78 | - | 16.82 | - | 4.83 | - |
LwFMC [33] | 42.14 | - | 25.59 | - | 11.68 | - | 26.82 | - | 16.49 | - |
LwM [9] | 43.39 | - | 26.07 | - | 12.08 | - | 26.97 | - | 16.50 | - |
iCaRL [45] | - | 65.34 | - | 64.51 | - | 58.73 | - | 40.09 | - | 33.77 |
UCIR [21] | 74.31 | 74.09 | 70.42 | 70.50 | 63.22 | 64.00 | 44.90 | 46.53 | 37.04 | 37.15 |
PODNet [12] | 73.26 | 74.37 | 71.58 | 73.75 | 70.28 | 71.87 | 44.32 | 48.78 | 38.76 | 46.62 |
TCD [41] | 74.89 | 77.16 | 73.43 | 75.35 | 72.19 | 74.01 | 45.34 | 50.36 | 40.47 | 46.66 |
FrameMaker | 78.13 | 78.64 | 76.38 | 78.14 | 75.77 | 77.49 | 47.54 | 51.12 | 42.65 | 47.37 |
Num. of Classes | stages | stages | ||
---|---|---|---|---|
Classifier | CNN | NME | CNN | NME |
UCIR [21] | 26.84 | 17.98 | 20.69 | 12.57 |
PODNet [12] | 34.94 | 27.33 | 26.95 | 17.49 |
TCD [41] | 35.78 | 28.88 | 29.60 | 21.63 |
FrameMaker | 37.25 | 29.92 | 30.98 | 22.84 |
UCF101 | HMDB51 | |||||
---|---|---|---|---|---|---|
Frames | FC | ISP | CNN | NME | CNN | NME |
All | - | - | 72.09 | 75.70 | 43.38 | 47.00 |
Random | ✗ | ✗ | 68.64 | 73.96 | 39.59 | 43.48 |
Random | ✗ | ✓ | 70.71 | 75.04 | 39.81 | 43.74 |
Average | ✗ | ✗ | 70.82 | 75.45 | 41.84 | 45.45 |
Average | ✗ | ✓ | 71.51 | 76.23 | 42.59 | 46.46 |
Condensed | ✓ | ✗ | 72.29 | 76.42 | 42.18 | 46.27 |
Condensed | ✓ | ✓ | 72.93 | 76.64 | 43.39 | 46.88 |
4.2 Comparison with State-of-the-art Results
This section presents the qualitative comparison of our proposed FrameMaker with the existing class incremental learning approaches under multiple challenging settings on three datasets: UCF101 [50], HMDB51 [28] and Something-Something V2 [18]. For a fair comparison, we use the same exemplar memory size per class, model structure and pre-training initialization as the existing methods [41].
Table 1 summarizes the results on UCF101 and HMDB51, which shows that FrameMaker outperforms other methods consistently under different configurations in terms of both CNN and NME scores. The average accuracy of our FrameMaker surpasses TCD by around 3.0% on UCF101 and 2.0% on HMDB51 in terms of CNN, respectively. These results demonstrate that one condensed frame can also be equipped with an effective spatio-temporal representation.
FrameMaker is compared with recent advanced methods on Something-Something V2 in Table 3. FrameMaker sets new state-of-the-art performance under multiple challenging settings. Although FrameMaker only uses one condensed frame for future replay, it still achieves decent performance on the large-scale motion-sensitive dataset. We speculate that these gains mainly come from two aspects: (1) The proposed Instance-Specific Prompt effectively rescues the lost spatio-temporal details, which is also discussed in Section 4.3. (2) Memory-efficient FrameMaker allows us to store more exemplar videos with less memory consumption, which greatly alleviates catastrophic forgetting.
4.3 Ablation Study
In this section, we present ablation studies to analyze the properties and the effectiveness of FrameMaker. If not specified, the ablation studies are performed on UCF101 with 10 steps and HMDB51 with 5 steps. For comparison with the case of storing all frames, we select 5 exemplar videos for each class.
CNN | NME | ||
---|---|---|---|
✗ | ✗ | 70.82 | 75.45 |
✓ | ✗ | 71.35 | 75.76 |
✗ | ✓ | 71.96 | 76.07 |
✓ | ✓ | 72.29 | 76.42 |
CNN | NME | |||
---|---|---|---|---|
✗ | ✓ | ✓ | 71.83 | 75.94 |
✓ | ✓ | ✗ | 72.35 | 76.48 |
✓ | ✗ | ✓ | 72.58 | 76.57 |
✓ | ✓ | ✓ | 72.93 | 76.64 |
CNN | NME | |
---|---|---|
2 | 71.62 | 75.89 |
4 | 71.70 | 76.12 |
8 | 72.93 | 76.64 |
16 | 72.42 | 76.11 |
Backbone | Frames | CNN | NME | Mem. |
---|---|---|---|---|
TSM | All | 43.38 | 47.00 | 6.00Mb |
TSM | FC+ISP | 43.39 | 46.88 | 0.75Mb |
R3D50 | All | 39.85 | 45.64 | 6.00Mb |
R3D50 | FC+ISP | 39.88 | 45.08 | 0.75Mb |
ViT | All | 35.34 | 39.46 | 6.00Mb |
ViT | FC+ISP | 35.25 | 39.58 | 0.75Mb |
Position | Type | CNN | NME | Mem. |
---|---|---|---|---|
Feature | T.-Spec. | 41.67 | 45.71 | 77.07Mb |
Feature | C.-Spec. | 41.72 | 45.93 | 385.35Mb |
Feature | I.-Spec. | 42.19 | 46.53 | 1926.75Mb |
Frame | T.-Spec. | 41.16 | 45.34 | 0.75Mb |
Frame | C.-Spec. | 42.01 | 45.47 | 0.75Mb |
Frame | I.-Spec. | 43.39 | 46.88 | 0.75Mb |
Frame Condensing and Instance-Specific Prompt.
To show the effectiveness of Frame Condensing and Instance-Specific Prompt, we compare them with the following alternative methods to produce stored frames for future replay: (1) All Frames, which simply saves the whole video. (2) One Random Frame, which randomly selects one frame for each exemplar video. (3) One Average Frame, which averages the input frames for each exemplar video. (4) One Condensed Frame, which is generated by our Frame Condensing procedure. The results are reported in Table 3. From the table, in terms of CNN, we observe that Instance-Specific Prompt yields an improvement of 2.07% and 0.22% with randomly sampled frames, 0.69% and 0.75% with averaged frames on UCF101 and HMDB51, respectively, which implies that prompts can always replenish the ignored spatio-temporal cues in video frames. Further, Frame Condensing with Instance-Specific Prompt leads to at least on par or better performance than preserving all frames while using just one condensed frame under all settings on both datasets.
Loss terms.
To preserve the best spatio-temporal information in the condensed frames, the distillation loss and the cross entropy loss are both applied for Frame Condensing and Instance-Specific Prompt. Table 6 presents the results with Frame Condensing from several different combinations of loss terms. provides more explicit semantic supervision, making the accuracy to step further. Similar experiments for Instance-Specific Prompt are shown in Table 6. One noticeable thing is that once introducing prompt, we only need to calculate the final and to realize the joint update of condensing weights and prompt . However, we observe that the learned weights, in this case, tend to be average, which indicates that is under-optimized. We speculate that this is due to the strong plasticity of prompt, which leads to the insufficient optimization of . To this end, we further employ and for as the guidance. The improvement in Table 6 and the visualizations in Figure 3 show that receives effective learning.
The number of frames used for Frame Condensing.
As discussed in Section 3.2, the frames are uniformly sampled from a representative video for Frame Condensing. In Table 6, we show the effect of in video class-incremental learning. We observe that the performance increases as increases for both CNN and NME. However, the performance is saturated when using 16 frames, which is in line with the conclusion of existing methods [54, 41], i.e., storing more frames in a video does not necessarily deliver performance improvement. This phenomenon reveals that FrameMaker can empower one Condensed Frame to absorb plenty of spatio-temporal features from original clips.
Different backbones.
To evaluate the applicability of proposed FrameMaker, we replace the backbone with 3D CNN-based method R3D50 [16] and video transformer ViT[11], as shown in Table 8. In these experiments, we simply replicate the condensed frames multiple times along the the temporal dimension to meet the 3D models and Transformers. From the results, we can find that our FrameMaker can consistently achieve comparable performance on the three different backbones with only 12% memory cost. These results demonstrate the better generalization of FrameMaker.
Alternative prompting strategies.
We conduct experiments to explore the types and positions of the prompts in Table 8. (i) We compare our Instance-Specific (I.-Spec.) Prompt (ISP) with the Task- (T.-Spec.) and Class-Specific (C.-Spec.) prompts and observe that our ISP can achieve the best performance when operating on both features and frames. Since the actions in videos have great intra-class variance, simply sharing task and class-specific prompts is insufficient to reserve the important spatia-temporal cues for each instance; (ii) Positions of prompt: For the feature settings, we add the prompt on the features of the 4 stages of the ResNet. The feature-based prompt is worse than our frame-based prompt, which may be caused by the mismatch between the running model and fixed prompts during incremental procedure. Moreover, additional features require additional storage space because the corresponding prompt cannot be directly added on the changed feature.
Memory Per Class | =1.2Mb | =2.4Mb | =6.0Mb | |||
---|---|---|---|---|---|---|
Classifier | CNN | NME | CNN | NME | CNN | NME |
iCaRL [45] | - | 58.05 | - | 60.50 | - | 64.51 |
UCIR [21] | 61.92 | 65.52 | 66.43 | 67.58 | 70.42 | 70.50 |
PODNet [12] | 63.18 | 70.96 | 65.93 | 72.78 | 71.58 | 73.75 |
TCD [41] | 64.52 | 71.96 | 68.40 | 73.30 | 73.43 | 75.35 |
Memory Per Class | =0.15Mb | =0.3Mb | =0.75Mb | |||
FrameMaker | 49.37 | 70.78 | 62.06 | 74.18 | 72.93 | 76.64 |
Memory Per Class | =1.2Mb | =2.4Mb | =6.0Mb | |||
FrameMaker | 73.64 | 76.98 | 75.19 | 77.43 | 76.38 | 78.14 |
Memory budget comparison.
We now compare the memory budget of FramMaker with existing approaches. To make a direct comparison, we use the definition of the working memory size in terms of stored frames, following [54]. The results are summarized in Table 9. We can see that our FrameMaker only costs 1.2Mb memory, which surpasses TCD [41] using 6.0Mb by 0.21% and 1.63% in terms of CNN and NME. FrameMaker saves 80% of memory space with the comparable performance. Meanwhile, we further increase the stored videos but keep the same memory with existing methods. FrameMaker can still effectively promote performance and shows superior abilities to fight catastrophic forgetting. Our results are about 3% ahead of TCD in both indicators. Besides, we also attempt to select the same number of videos as the existing methods. However, the accuracy of the trained linear classifier, i.e., CNN, is weaker with a small number of video instances. Interestingly, NME evaluated by the exemplar class centre yielded fairly or even better results, implying that the model suffers only slight forgetting. We hypothesize that this is caused by the poor feature diversity that a few condensed frames are provided. Therefore, the classifier tends to overfit fixed sample features, resulting in poor generalization.

Visualization of condensing weights and instance-specific prompts.
To intuitively understand our FrameMaker, we provide visualization of condensing weights, Instance-Specific Prompts, condensed frames and learned prompts in Figure 3. To save space, we uniformly select the most typical 4 frames from frames for display. Based on this, we have the following fascinating observations: (1) For frames with strong scene bias, the learned weights are more uniform, while those with motion bias are not. For example, the action classification for (a) and (b) in Figure 3 does not depend on a specific frame. Hence, the condensing weight for each frame is roughly the same. In contrast, stair climbing and hugging in (c) and (d) only occur in some specific frames, and those keyframes are highlighted by condensing weights. This demonstrates that it is reasonable to condense multiple frames into one frame since our proposed Frame Condensing can assist the learning of which frame is meaningful for the action recognition task. (2) It is difficult for humans to absorb useful knowledge from the learned prompts shown in Figure 3, and even they seem to be the same. It should be emphasized that the prompts are different. We try to share the same prompt for all videos to verify this. The experiment is conducted on UCF101 with 10 steps, and the performances decrease by 0.83% and 0.78% in terms of CNN and NME, respectively. This fully validates that our Instance-Specific Prompt is to replenish the specific information for different instances.

Per-category and per-task analysis.
We depict the task-wise and category-wise accuracy with different forms of stored frames in Figure 4, i.e., "All Frames", "Frame Condensing without Instance-Specific Prompt" and "Frame Condensing with Instance-Specific Prompt". As expected, integrating multiple frames into one condensed frame does undermine performance in almost all cases unless Instance-Specific Prompt is introduced. As shown in Figure 4(a) and (b), the gains on incremental tasks yielded by Instance-Specific Prompt mainly fall in the later tasks, which implies that the spatio-temporal knowledge absorbed by prompt can effectively alleviate forgetting. Meanwhile, from Figure 4(c) and (d), we also observe that Frame Condensing can lead to the performance degradation of actions with large temporal and spatial variations or short duration, such as "CliffDiving", "Ride horse" and "Hug", etc. Therefore, Frame Condensing with only a few parameters is difficult to capture these complex spatio-temporal interactions. However, Instance-Specific Prompt helps alleviate this issue since it recovers the lost spatio-temporal details in Frame Condensing for each instance.
5 Discussions
Limitations.
Compared with the previous video incremental learning approaches, FrameMaker provides a simple framework that significantly reduces the amount of memory required for each exemplar video. Nevertheless, FrameMaker may fail when the memory budget is highly constrained, which needs to be further explored in future work.
Societal Impacts.
Although FrameMaker only stores condensed frames in memory buffer for experience replay, which still retains some information about the original video. Therefore, FrameMaker may be used in applications with privacy concerns [48].
Conclusion.
This paper proposes FrameMaker, a memory-efficient approach for video class-incremental learning. It explores how to learn an effective condensed frame with less memory for each video that can be applied to the future incremental task. FrameMaker mainly consists of two key components, i.e., Frame Condensing and Instance-Specific Prompt. The former learns a group of condensed weights for each video to integrate multiple frames into one condensed frame, while the latter learns to retrieve the collapsed spatio-temporal structure for the condensed frame. In this way, FrameMaker offers better results with only 20% memory consumption compared to recent advanced methods and sets a new state-of-the-art class-incremental learning performance on multiple challenging datasets. Memory-based class-incremental learning is one of the effective methods to fight forgetting. The efficient utilization of limited space is worth exploring in future work.
Acknowledgments.
This work is supported by the National Key R&D Program of China under Grant No. 2018AAA0101501, the National Natural Science Foundation of China undergrant 61871435 and 62272380, Fundamental Research Funds for the Central Universities no.2019kfyXKJC024, the Science and Technology Program of Xi’an, China under Grant 21RGZN0017, and by Alibaba Group through Alibaba Innovative Research Program.
References
- Aljundi et al. [2018] Rahaf Aljundi, Francesca Babiloni, Mohamed Elhoseiny, Marcus Rohrbach, and Tinne Tuytelaars. Memory aware synapses: Learning what (not) to forget. In Proceedings of the European Conference on Computer Vision, pages 139–154, 2018.
- Arnab et al. [2021] Anurag Arnab, Mostafa Dehghani, Georg Heigold, Chen Sun, Mario Lučić, and Cordelia Schmid. Vivit: A video vision transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6836–6846, 2021.
- Bahng et al. [2022] Hyojin Bahng, Ali Jahanian, Swami Sankaranarayanan, and Phillip Isola. Visual prompting: Modifying pixel space to adapt pre-trained models. arXiv preprint arXiv:2203.17274, 2022.
- Bertasius et al. [2021] Gedas Bertasius, Heng Wang, and Lorenzo Torresani. Is space-time attention all you need for video understanding. arXiv preprint arXiv:2102.05095, 2(3):4, 2021.
- Buzzega et al. [2020] Pietro Buzzega, Matteo Boschini, Angelo Porrello, Davide Abati, and Simone Calderara. Dark experience for general continual learning: a strong, simple baseline. Advances in neural information processing systems, 33:15920–15930, 2020.
- Carreira and Zisserman [2017] Joao Carreira and Andrew Zisserman. Quo vadis, action recognition? a new model and the kinetics dataset. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 6299–6308, 2017.
- Castro et al. [2018] Francisco M Castro, Manuel J Marín-Jiménez, Nicolás Guil, Cordelia Schmid, and Karteek Alahari. End-to-end incremental learning. In Proceedings of the European conference on computer vision, pages 233–248, 2018.
- Deng et al. [2009] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
- Dhar et al. [2019] Prithviraj Dhar, Rajat Vikram Singh, Kuan-Chuan Peng, Ziyan Wu, and Rama Chellappa. Learning without memorizing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5138–5146, 2019.
- Diba et al. [2018] Ali Diba, Mohsen Fayyaz, Vivek Sharma, M Mahdi Arzani, Rahman Yousefzadeh, Juergen Gall, and Luc Van Gool. Spatio-temporal channel correlation networks for action classification. In Proceedings of the European Conference on Computer Vision, pages 284–299, 2018.
- Dosovitskiy et al. [2020] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- Douillard et al. [2020] Arthur Douillard, Matthieu Cord, Charles Ollion, Thomas Robert, and Eduardo Valle. Podnet: Pooled outputs distillation for small-tasks incremental learning. In European Conference on Computer Vision, pages 86–102. Springer, 2020.
- Fan et al. [2021] Haoqi Fan, Bo Xiong, Karttikeya Mangalam, Yanghao Li, Zhicheng Yan, Jitendra Malik, and Christoph Feichtenhofer. Multiscale vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6824–6835, 2021.
- Fan et al. [2018] Hehe Fan, Zhongwen Xu, Linchao Zhu, Chenggang Yan, Jianjun Ge, and Yi Yang. Watching a small portion could be as good as watching all: Towards efficient video classification. In IJCAI International Joint Conference on Artificial Intelligence, 2018.
- Feichtenhofer [2020] Christoph Feichtenhofer. X3d: Expanding architectures for efficient video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 203–213, 2020.
- Feichtenhofer et al. [2019] Christoph Feichtenhofer, Haoqi Fan, Jitendra Malik, and Kaiming He. Slowfast networks for video recognition. In Proceedings of the IEEE/CVF international conference on computer vision, pages 6202–6211, 2019.
- Gowda et al. [2021] Shreyank N Gowda, Marcus Rohrbach, and Laura Sevilla-Lara. Smart frame selection for action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 1451–1459, 2021.
- Goyal et al. [2017] Raghav Goyal, Samira Ebrahimi Kahou, Vincent Michalski, Joanna Materzynska, Susanne Westphal, Heuna Kim, Valentin Haenel, Ingo Fruend, Peter Yianilos, Moritz Mueller-Freitag, et al. The" something something" video database for learning and evaluating visual common sense. In Proceedings of the IEEE international conference on computer vision, pages 5842–5850, 2017.
- He et al. [2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- Hinton et al. [2015] Geoffrey Hinton, Oriol Vinyals, Jeff Dean, et al. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2(7), 2015.
- Hou et al. [2019] Saihui Hou, Xinyu Pan, Chen Change Loy, Zilei Wang, and Dahua Lin. Learning a unified classifier incrementally via rebalancing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 831–839, 2019.
- Hu et al. [2019] Han Hu, Zheng Zhang, Zhenda Xie, and Stephen Lin. Local relation networks for image recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.
- Huang et al. [2018] De-An Huang, Vignesh Ramanathan, Dhruv Mahajan, Lorenzo Torresani, Manohar Paluri, Li Fei-Fei, and Juan Carlos Niebles. What makes a video a video: Analyzing temporal information in video understanding models and datasets. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 7366–7375, 2018.
- Jia et al. [2022] Menglin Jia, Luming Tang, Bor-Chun Chen, Claire Cardie, Serge Belongie, Bharath Hariharan, and Ser-Nam Lim. Visual prompt tuning. arXiv preprint arXiv:2203.12119, 2022.
- Kim et al. [2022] Kiyoon Kim, Shreyank N Gowda, Oisin Mac Aodha, and Laura Sevilla-Lara. Capturing temporal information in a single frame: Channel sampling strategies for action recognition. arXiv preprint arXiv:2201.10394, 2022.
- Kirkpatrick et al. [2017] James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, et al. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13):3521–3526, 2017.
- Korbar et al. [2019] Bruno Korbar, Du Tran, and Lorenzo Torresani. Scsampler: Sampling salient clips from video for efficient action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6232–6242, 2019.
- Kuehne et al. [2011] Hildegard Kuehne, Hueihan Jhuang, Estíbaliz Garrote, Tomaso Poggio, and Thomas Serre. Hmdb: a large video database for human motion recognition. In 2011 International conference on computer vision, pages 2556–2563. IEEE, 2011.
- Lester et al. [2021] Brian Lester, Rami Al-Rfou, and Noah Constant. The power of scale for parameter-efficient prompt tuning. arXiv preprint arXiv:2104.08691, 2021.
- Li et al. [2022] Kunchang Li, Yali Wang, Peng Gao, Guanglu Song, Yu Liu, Hongsheng Li, and Yu Qiao. Uniformer: Unified transformer for efficient spatiotemporal representation learning. arXiv preprint arXiv:2201.04676, 2022.
- Li and Liang [2021] Xiang Lisa Li and Percy Liang. Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190, 2021.
- Li and Hoiem [2017a] Zhizhong Li and Derek Hoiem. Learning without forgetting. IEEE transactions on pattern analysis and machine intelligence, 40(12):2935–2947, 2017a.
- Li and Hoiem [2017b] Zhizhong Li and Derek Hoiem. Learning without forgetting. IEEE transactions on pattern analysis and machine intelligence, 40(12):2935–2947, 2017b.
- Lin et al. [2019] Ji Lin, Chuang Gan, and Song Han. Tsm: Temporal shift module for efficient video understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7083–7093, 2019.
- Liu et al. [2022a] Chengxu Liu, Huan Yang, Jianlong Fu, and Xueming Qian. Learning trajectory-aware transformer for video super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022a.
- Liu et al. [2021a] Pengfei Liu, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. arXiv preprint arXiv:2107.13586, 2021a.
- Liu et al. [2021b] Ze Liu, Jia Ning, Yue Cao, Yixuan Wei, Zheng Zhang, Stephen Lin, and Han Hu. Video swin transformer. arXiv preprint arXiv:2106.13230, 2021b.
- Liu et al. [2022b] Ze Liu, Han Hu, Yutong Lin, Zhuliang Yao, Zhenda Xie, Yixuan Wei, Jia Ning, Yue Cao, Zheng Zhang, Li Dong, Furu Wei, and Baining Guo. Swin transformer v2: Scaling up capacity and resolution. In International Conference on Computer Vision and Pattern Recognition, 2022b.
- McCloskey and Cohen [1989] Michael McCloskey and Neal J Cohen. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of learning and motivation, volume 24, pages 109–165. Elsevier, 1989.
- Ostapenko et al. [2019] Oleksiy Ostapenko, Mihai Puscas, Tassilo Klein, Patrick Jahnichen, and Moin Nabi. Learning to remember: A synaptic plasticity driven framework for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11321–11329, 2019.
- Park et al. [2021] Jaeyoo Park, Minsoo Kang, and Bohyung Han. Class-incremental learning for action recognition in videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 13698–13707, 2021.
- Paszke et al. [2019] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
- Qiu et al. [2017] Zhaofan Qiu, Ting Yao, and Tao Mei. Learning spatio-temporal representation with pseudo-3d residual networks. In proceedings of the IEEE International Conference on Computer Vision, pages 5533–5541, 2017.
- Qiu et al. [2021] Zhaofan Qiu, Ting Yao, Yan Shu, Chong-Wah Ngo, and Tao Mei. Condensing a sequence to one informative frame for video recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 16311–16320, 2021.
- Rebuffi et al. [2017] Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, and Christoph H Lampert. icarl: Incremental classifier and representation learning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pages 2001–2010, 2017.
- Reddy and Shah [2013] Kishore K Reddy and Mubarak Shah. Recognizing 50 human action categories of web videos. Machine vision and applications, 24(5):971–981, 2013.
- Shin et al. [2017] Hanul Shin, Jung Kwon Lee, Jaehong Kim, and Jiwon Kim. Continual learning with deep generative replay. Advances in neural information processing systems, 30, 2017.
- Shokri and Shmatikov [2015] Reza Shokri and Vitaly Shmatikov. Privacy-preserving deep learning. In Proceedings of the 22nd ACM SIGSAC conference on computer and communications security, pages 1310–1321, 2015.
- Simonyan and Zisserman [2014] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
- Soomro et al. [2012] Khurram Soomro, Amir Roshan Zamir, and Mubarak Shah. Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402, 2012.
- Tran et al. [2015] Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE international conference on computer vision, pages 4489–4497, 2015.
- Tran et al. [2018] Du Tran, Heng Wang, Lorenzo Torresani, Jamie Ray, Yann LeCun, and Manohar Paluri. A closer look at spatiotemporal convolutions for action recognition. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pages 6450–6459, 2018.
- Tran et al. [2019] Du Tran, Heng Wang, Lorenzo Torresani, and Matt Feiszli. Video classification with channel-separated convolutional networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5552–5561, 2019.
- Villa et al. [2022] Andrés Villa, Kumail Alhamoud, Juan León Alcázar, Fabian Caba Heilbron, Victor Escorcia, and Bernard Ghanem. vclimb: A novel video class incremental learning benchmark. arXiv preprint arXiv:2201.09381, 2022.
- Wang et al. [2019] Junbo Wang, Wei Wang, Zhiyong Wang, Liang Wang, Dagan Feng, and Tieniu Tan. Stacked memory network for video summarization. In Proceedings of the 27th ACM International Conference on Multimedia, pages 836–844, 2019.
- Wang et al. [2018] Limin Wang, Wei Li, Wen Li, and Luc Van Gool. Appearance-and-relation networks for video classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1430–1439, 2018.
- Wang et al. [2021] Zifeng Wang, Zizhao Zhang, Chen-Yu Lee, Han Zhang, Ruoxi Sun, Xiaoqi Ren, Guolong Su, Vincent Perot, Jennifer Dy, and Tomas Pfister. Learning to prompt for continual learning. arXiv preprint arXiv:2112.08654, 2021.
- Wang et al. [2022] Zifeng Wang, Zizhao Zhang, Sayna Ebrahimi, Ruoxi Sun, Han Zhang, Chen-Yu Lee, Xiaoqi Ren, Guolong Su, Vincent Perot, Jennifer Dy, et al. Dualprompt: Complementary prompting for rehearsal-free continual learning. arXiv preprint arXiv:2204.04799, 2022.
- Wu et al. [2019a] Wenhao Wu, Dongliang He, Xiao Tan, Shifeng Chen, and Shilei Wen. Multi-agent reinforcement learning based frame sampling for effective untrimmed video recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6222–6231, 2019a.
- Wu et al. [2019b] Yue Wu, Yinpeng Chen, Lijuan Wang, Yuancheng Ye, Zicheng Liu, Yandong Guo, and Yun Fu. Large scale incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 374–382, 2019b.
- Wu et al. [2019c] Zuxuan Wu, Caiming Xiong, Chih-Yao Ma, Richard Socher, and Larry S Davis. Adaframe: Adaptive frame selection for fast video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1278–1287, 2019c.
- Xie et al. [2017] Saining Xie, Chen Sun, Jonathan Huang, Zhuowen Tu, and Kevin Murphy. Rethinking spatiotemporal feature learning for video understanding. arXiv preprint arXiv:1712.04851, 1(2):5, 2017.
- Zenke et al. [2017] Friedemann Zenke, Ben Poole, and Surya Ganguli. Continual learning through synaptic intelligence. In International Conference on Machine Learning, pages 3987–3995. PMLR, 2017.
- Zhao et al. [2021a] Bin Zhao, Haopeng Li, Xiaoqiang Lu, and Xuelong Li. Reconstructive sequence-graph network for video summarization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(5):2793–2801, 2021a.
- Zhao et al. [2021b] Hanbin Zhao, Xin Qin, Shihao Su, Yongjian Fu, Zibo Lin, and Xi Li. When video classification meets incremental classes. In Proceedings of the 29th ACM International Conference on Multimedia, pages 880–889, 2021b.
- Zhi et al. [2021] Yuan Zhi, Zhan Tong, Limin Wang, and Gangshan Wu. Mgsampler: An explainable sampling strategy for video action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1513–1522, 2021.
- Zhu et al. [2020] Wencheng Zhu, Jiwen Lu, Jiahao Li, and Jie Zhou. Dsnet: A flexible detect-to-summarize network for video summarization. IEEE Transactions on Image Processing, 30:948–962, 2020.
Checklist
The checklist follows the references. Please read the checklist guidelines carefully for information on how to answer these questions. For each question, change the default [TODO] to [Yes] , [No] , or [N/A] . You are strongly encouraged to include a justification to your answer, either by referencing the appropriate section of your paper or providing a brief inline description. For example:
-
•
Did you include the license to the code and datasets? [N/A]
Please do not modify the questions and only use the provided macros for your answers. Note that the Checklist section does not count towards the page limit. In your paper, please delete this instructions block and only keep the Checklist section heading above along with the questions/answers below.
-
1.
For all authors…
-
(a)
Do the main claims made in the abstract and introduction accurately reflect the paper’s contributions and scope? [Yes]
-
(b)
Did you describe the limitations of your work? [Yes]
-
(c)
Did you discuss any potential negative societal impacts of your work? [Yes]
-
(d)
Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes]
-
(a)
-
2.
If you are including theoretical results…
-
(a)
Did you state the full set of assumptions of all theoretical results? [N/A]
-
(b)
Did you include complete proofs of all theoretical results? [N/A]
-
(a)
-
3.
If you ran experiments…
-
(a)
Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [No]
-
(b)
Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes]
-
(c)
Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [No]
-
(d)
Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes]
-
(a)
-
4.
If you are using existing assets (e.g., code, data, models) or curating/releasing new assets…
-
(a)
If your work uses existing assets, did you cite the creators? [Yes]
-
(b)
Did you mention the license of the assets? [No]
-
(c)
Did you include any new assets either in the supplemental material or as a URL? [No]
-
(d)
Did you discuss whether and how consent was obtained from people whose data you’re using/curating? [No]
-
(e)
Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [No]
-
(a)
-
5.
If you used crowdsourcing or conducted research with human subjects…
-
(a)
Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A]
-
(b)
Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A]
-
(c)
Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]
-
(a)
Appendix A Additional Experiments for Different Memory Budgets
This is supplementary to Section 4.2 "Comparison with State-of-the-art Results". Limited by space, we only summarize the results of our approach using the same exemplar memory size per class as others and the experiment results of "Memory budget comparison" on UCF101 with 10 steps in the main body. Table 10 supplements the average accuracy of FrameMaker with different memory budgets under all benchmarks of UCF101 and HMDB51 datasets, which is compared with the best performance of TCD [41].
UCF101 | HMDB51 | |||||||||
Num. of Classes | stages | stages | stages | stages | stages | |||||
Memory Per Class | CNN | NME | CNN | NME | CNN | NME | CNN | NME | CNN | NME |
=6.0Mb [41] | 74.89 | 77.16 | 73.43 | 75.35 | 72.19 | 74.01 | 45.34 | 50.36 | 40.47 | 46.66 |
=0.15Mb | 58.75 | 73.97 | 49.37 | 70.78 | 48.31 | 65.86 | 36.08 | 43.82 | 32.53 | 39.00 |
=0.3Mb | 67.59 | 76.18 | 62.06 | 74.18 | 55.84 | 72.88 | 39.33 | 46.07 | 34.05 | 43.18 |
=0.75Mb | 73.05 | 76.70 | 72.93 | 76.64 | 68.79 | 75.36 | 43.39 | 46.88 | 38.95 | 44.18 |
=1.2Mb | 76.09 | 77.88 | 73.64 | 76.98 | 73.57 | 76.76 | 46.25 | 49.02 | 40.80 | 46.97 |
=2.4Mb | 77.24 | 78.06 | 75.19 | 77.43 | 74.92 | 76.92 | 46.97 | 50.42 | 41.96 | 47.03 |
=6.0Mb | 78.13 | 78.64 | 76.38 | 78.14 | 75.77 | 77.49 | 47.54 | 51.12 | 42.65 | 47.37 |
As can be seen from Table 10, under other settings, the results of our method are similar to that on UCF101 with 10 steps. When using about 20% of the storage space, our method can achieve better performance than TCD, and the accuracy is further improved when the number of saved exemplars is increased, which also proves the robustness of our approach.
Appendix B Additional Analysis of the Balance Weights
This is supplementary to Section 4.3 "Ablation Study". We further discuss the sensitivity to the balance weight for each term in on UCF101 with 10 steps.
We first discuss the sensitivity of and , the balance weights for and which are used to achieve stronger constraint on for its effective training. Table 11 shows the performance of frame condensing (without instance-specific prompt) under various combinations of and . We find the performance under the combination of consistently exceeds others. Although a slightly higher performance is obtained on NME under , it is difficult to achieve the same gain on the two different metrics after continuous adjustment.
1 | 2 | 5 | 0.5 | 0.1 | 0.01 | 1 | 1 | 1 | 1 | 1 | |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | 1 | 1 | 1 | 1 | 1 | 2 | 5 | 0.5 | 0.1 | 0.01 | |
CNN | 72.29 | 72.27 | 71.83 | 72.22 | 72.07 | 71.94 | 71.93 | 71.44 | 72.14 | 71.96 | 71.92 |
NME | 76.42 | 76.43 | 76.36 | 76.25 | 76.21 | 76.13 | 76.37 | 76.23 | 76.22 | 75.91 | 75.84 |
Then we further analyze sensitivity of and , the balance weights for and , when setting and as the optimal value, . Table 12 shows the performance of FrameMaker (frame condensing with instance-specific prompt) under various combinations of and .
1 | 2 | 5 | 0.5 | 0.1 | 0.01 | 1 | 1 | 1 | 1 | 1 | |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | 1 | 1 | 1 | 1 | 1 | 2 | 5 | 0.5 | 0.1 | 0.01 | |
CNN | 72.93 | 72.94 | 72.92 | 72.88 | 72.64 | 72.67 | 72.90 | 72.85 | 72.92 | 72.80 | 72.73 |
NME | 76.64 | 76.62 | 76.58 | 76.61 | 76.59 | 76.61 | 76.65 | 76.58 | 76.61 | 76.55 | 76.51 |
We find the performance under the combination of always exceeds others. Some groups of weights make one of the test metrics higher but only obtain the best balance of them. Therefore, we set all the balance weights as finally in our experiments.
Appendix C Implementation Details
This section provides some additional details about the experiments in the main body.
Training Details.
This is supplementary to Section 3.3 "Training". We further supplement the elaborate training process and knowledge distillation loss in this section. Figure 5 shows the overall framework of training process.

In the training process of the incremental step , we use the dataset to update the model from , where is the dataset of the task which consists of videos belonging to current classes, and is memory bank which contains condensed frames generated by FrameMaker after the step . The model classifies the samples from the two datasets respectively and use exemplars in old classes for knowledge distillation.
For the classification task, we adopt cross entropy loss to supervise. and are the cross entropy losses for video example in and condensed frame exemplar in .
(9) |
(10) |
where and is the ground truth label of and , respectively.
In order to save the information of old classes better, the knowledge distillation function of PODNet [12] is employed on the old classes data and it consists of two components, the spatial distillation function and the final embedding distillation function .
(11) |
where is the output of intermediate convolution layer of model . C, H, W are the size of channel, height and width of the output feature of layer .
(12) |
where is the feature extractor of model .
The final distillation loss of the condensed frame exemplars is given by:
(13) |
To simplify, the hyperparameters used in PODNet [12] final distillation loss is ignored.
The total objective function for task is given by:
(14) |
which is the same as Eq. 8 in the manuscript.
Dataset Details.
Following TCD [41], we use three action recognition datasets, UCF101 [50], HMDB51 [28] and Something-Something V2 [18]. The HMDB51 dataset is a large collection of realistic videos from various sources which can be found in https://serre-lab.clps.brown.edu/resource/hmdb-a-large-human-motion-database/. UCF101 dataset is an extension of UCF50 [46] and consists of 13,320 video clips, which can be found in https://www.crcv.ucf.edu/data/UCF101.php. Something-SomethingV2 dataset is a large-scale motion-sensitive dataset that shows humans performing pre-defined basic actions with everyday objects, which can be found in https://developer.qualcomm.com/software/ai-datasets/something-something. For UCF101, we have three settings, stages, stages and stages. For HMDB51, the number of initial task classes is 26, and the remaining 25 classes are separated into 5 or 25 groups. For Something-Something V2, the benchmarks are stages and stages. The random seed we use is also the same as TCD [41].
Model and Hyperparameter Setup.
TSM [34] is employed as our backbone, and we follow the data preprocessing procedure and the basic training setting of TSM. We train a ResNet-34 TSM for UCF101 and a ResNet-50 TSM for HMDB51 and Something-Something V2. Each incremental training procedure takes 50 epochs. We use SGD optimizer with an initial learning rate of 0.04 for UCF101 and Something-Something V2, and 0.002 for HMDB51, which is reduced by half after 25 epochs. For incremental steps, we set the learning rate for all datasets as 0.001 and make the same reduction. The learning rate is set as 0.01 and 0.001 for condensing weights and the instance-specific prompt, respectively, and the total iteration for their optimization is set to be 8k.
Hardware Information.
We train all models on eight NVIDIA V100 GPUs and use PyTorch [42] for all our experiments.
Appendix D Visualizations


Visual comparison of features. Figure 6 plots visual comparisons of the extracted features with different frames as model inputs. We can summarize that the average frame shows the largest feature difference compared to the features extracted from the original clips. Our proposed Frame Condensing and Instance-Specific Prompt can effectively reduce the feature differences with the original clips and thus effectively improve the feature quality of condensed frames.
Visualizations of condensed frames. Figure 7 plots the condensed frames, the sum of the condensed frames and their prompts, and the activation region generated by the GradCAM technology. From the results, we observe the learned prompts have very small magnitudes (aver- age value is 0.03), hence no significant difference is observed actually. However, from the GradCAM attention maps, we can observe the condensed frames without prompts may focus on background regions, while the learned prompts tend to increase the attention of motion regions or reduce the attention of the backgrounds. We think the reason behind this is that the collapsed temporal dynamics may easily lead the model pay more attention to the background to some extent, while the prompt can guide the model to refocus on the motion region and weaken the interference from backgrounds.
Appendix E Training Prompt without Condensed Frame

We attempt to train Instance-Specific Prompt with out condensed frames, and the learned prompting parameters are visualized in Figure 8(c) and (d). We observe that the learned prompts have no intuitive semantics. To further understand this phenomenon, we plot the per-task accuracy and the loss curve of the prompts in Figure 8(a) and (b), respectively. We can summarize two points: (i) the learned prompts without condensed frames cannot fight against catastrophic forgetting, which indicates that the semantic information of the learned prompts is weak; (ii) the loss curve of the prompts without condensed frames is much higher than that with condensed frames, which implies that it is difficult to train a single prompts from scratch. Thus, it is crucial that the condensed frames provide a good initialization for the prompts to avoid the local minima of prompts.