section [3em] \contentslabel2em \titlerule*[1pc].\contentspage \titlecontentssubsection [5em] \contentslabel2.6em \titlerule*[1pc].\contentspage
Dancing with Still Images:
Video Distillation via Static-Dynamic Disentanglement
Abstract
Recently, dataset distillation has paved the way towards efficient machine learning, especially for image datasets. However, the distillation for videos, characterized by an exclusive temporal dimension, remains an underexplored domain. In this work, we provide the first systematic study of video distillation and introduce a taxonomy to categorize temporal compression. Our investigation reveals that the temporal information is usually not well learned during distillation, and the temporal dimension of synthetic data contributes little. The observations motivate our unified framework of disentangling the dynamic and static information in the videos. It first distills the videos into still images as static memory and then compensates the dynamic and motion information with a learnable dynamic memory block. Our method achieves state-of-the-art on video datasets at different scales, with a notably smaller memory storage budget. Our code is available at https://github.com/yuz1wan/video_distillation.
1 Introduction
Dataset distillation, as an emerging direction recently, compresses the original dataset into a smaller one while maintaining training effectiveness. It alleviates the challenges of costly training due to the increasingly large datasets and models. It is widely adopted in various downstream fields including federated learning and continual learning.
Recent works on dataset distillation mainly focus on distilling images [25, 31, 30, 4, 32, 16, 17, 15, 29, 10, 6, 23, 13, 7]. Though some methods seem to seamlessly adapt to other data formats or modalities [12, 27, 28], few works studied video distillation. Compared to image data, videos possess an additional temporal dimension, which significantly adds to the time and space complexity of the distillation algorithms and is already hardly affordable when the instance-per-class (IPC) is large. Besides, the scale of video datasets [19, 11, 3] is usually more intimidating. However, the high temporal redundancy of videos is very suitable for and can be well exploited by dataset distillation methods, providing a good opportunity for dataset distillation. Therefore, in this work, we firstly and systematically study the dataset distillation on video data, especially involving the compression of the video temporal redundancy.

Currently, dataset distillation approaches directly align the training dynamics (gradient [31], trajectory [4], feature [30], etc.) of real and synthetic. On video datasets, these methods simply match all the real and synthetic frames and the frame correspondence can be depicted as a complete bipartite graph. Therefore, to condense the temporal dimension, we can either reduce the synthetic frames or prune the bipartite correspondence graph between real and synthetic frames. For the further analysis of temporal condensation in video distillation, we put these two schemes under one unified framework, namely “segmented matching and interpolation” (Fig. 1(b)): the real videos are synthetic videos are partitioned to multiple segments and distillation are applied within each real-synthetic segments pair; the synthetic videos are then interpolated to the required video length. This framework can cover most scenarios at different levels of temporal condensation.
Then, to thoroughly study temporal condensation, we build a taxonomy for various temporal correspondences and classify the potential methods based on four essential dimensions: the numbers of synthetic frames and real frames for matching, the number of segments, and the interpolation algorithm. Along these four dimensions, we conduct comparisons of the distillation performance and computation cost. Empirical analysis shows that, though increasing frame numbers does enhance distillation performance, the improvement is marginal and comes at the expense of considerably longer training times and higher costs; and frame segmentation reduces the training costs, but brings a substantial decrease in model performance. These important taxonomies and observations could guide the research and the efficient algorithm design of video distillation.
In light of the above observation, we propose a unified framework for video distillation to exploit the unique temporal redundancy of videos. Our observation implies that dense frame correspondence is non-critical in the video distillation task. Hence, to maximize the efficiency, we reduce the real frames, synthetic frames, and segment length to 1. This is equivalent to image distillation with which we can distill the static information in the first stage. Second, we use a learnable dynamic memory to compensate for the loss of dynamic information. The static and dynamic memories are then combined with an integrator network. Our paradigm can be effortlessly applied to existing algorithms to enhance performance with a memory storage budget (referred to as storage here) used. We embed our method to various data distillation algorithms including DM [30], MTT [4], FRePo [32], and achieve state-of-the-art with less storage. Our approach could achieve comparable performance with 50 memory storage budget and bring substantial improvement with a comparable one.
Overall, our contributions are: (1) We propose the very first work that systematically studies the video dataset distillation. (2) We introduce a novel taxonomy for temporal condensation in video distillation methods, which guides our and future works. (3) We propose a novel paradigm, enabling existing image distillation techniques to achieve improved results when applied to video distillation while using an even smaller memory storage budget.
2 Related work
Dataset Distillation/Condensation. Dataset distillation [25], endeavors to condense large datasets into smaller ones while maintaining comparable training performance. The algorithms fall into the following categories:
(1) Performance Matching: following the very first work of DD [25], a broad category of techniques employs bi-level optimization. A few methods integrate kernel ridge regression (KRR) to reduce the computational complexity of bi-level optimization, where KIP [16, 17] employs the Neural Tangents Kernel (NTK), while RFAD [15] adopts the Empirical Neural Network Gaussian Process (NNGP). FRePo [32] separates a neural network into a feature extractor and a linear classifier to optimize.
(2) Parameter Matching: DC [31] aligns the single-step gradients of synthetic and real data. In line with DC, DSA [29] enhances this approach through symmetrical image augmentation, and IDC [10] enhances by storing synthetic data in lower resolutions. MTT [4] first applies multi-step parameter matching, and TESLA [6] reduces memory usage and uses learnable soft labels.
(3) Distribution Matching: DM [30] directly aligns the features of real and synthetic data, while CAFE [23] ensures statistical feature properties from synthetic and real samples are consistent across all network layers except the final one.
(4) Factorization Methods decompress the full dataset into two components: base and hallucinator. HaBa [13] employs task-input-like bases and ConvNet hallucinators, while LinBa [7] integrates a linear hallucinator with given predictands. Inspired by these methods, we factorize the static and dynamic information in video distillation to minimize temporal redundancy and reduce storage costs.
Video Recognition. Video recognition involves the classification of videos into semantic classes, e.g., human actions and scenes. Currently, there are several main design philosophies for video recognition with deep learning: (1) 2D Convolution-Based: The most intuitive approach is to break down the video into individual frames and process the frames individually. Then temporal aggregation (pooling, LSTM, GRU, etc.) is used for getting video-level features, which is then used for classification. (2) 3D Convolution-Based. To adopt early aggregation of temporal features, the presence of an additional temporal dimension in videos naturally suggests the possibility of employing 3D convolutional networks. Tran et al. [21] propose C3D, and then Carreira and Zisserman [3] extend the pre-trained models of 2D convolutional networks to 3D, with consideration of alleviating the challenges posed by the large number of parameters in 3D convolutions. Meanwhile, efforts are dedicated to low-rank approximations for 3D convolutions. R(2+1)D [22] employs a spatial 2D convolutional structure combined with a temporal 1D convolution to achieve pseudo-3D convolutions. (3) Transformer-Based. With the success of attention mechanisms in natural language processing, the long-range effectiveness of self-attention determines its suitability for video recognition. Therefore, there have also emerged some models [24, 2, 20] for videos based on Transformers.
3 Pre-analysis
The temporal redundancy has been widely discussed for video understanding, while we focus on analyzing the temporal compression in dataset distillation. In this section, we first propose some basic principles for temporal compression (Sec. 3.1). We use a generic paradigm to describe the temporal compression strategies and further propose a taxonomy of compression according to four factors (Sec. 3.2), along with which we conduct comprehensively comparisons and obtain our observations and conclusions, supporting further study on video distillation (Sec. 3.3).
3.1 Preliminaries
Video Temporal Redundancy. Most current works are dedicated to developing and improving methods for compressing image datasets in terms of quantity. Video data have an additional temporal dimension as its major difference from images, suggesting the presence of temporal redundancy. Video temporal redundancy study has a long history [14, 9]. Researchers have long observed the significant temporal redundancy, which has diverse causes. For example, videos inherently exhibit substantial similarity between adjacent frames, leading to low temporal information utilization during data usage. Thus, we focus on, analyze, and exploit the temporal compression for video distillation.
Precondition for Temporal Correspondence. To study the temporal compression, we are categorizing the temporal correspondence for the distillation matching (e.g. gradient/trajectory/distribution matching) between real and synthetic videos. We can put all scenarios into one formulation, which covers all possible methods from the temporal aspect:
Definition 1.
Compressed video distillation involves real video and synthetic video with asymmetrical lengths. Multiple frame sequences are drawn from , and paired: , , where , , . Distillation is to apply matching algorithms to each frame sequence pair individually.

Fig. 2(a) gives an example of compressed video distillation. Though the existing distillation methods usually produce irregular and “freeform” patterns that are cryptic for humans, on video data, we intuitively desire an algorithm that obeys temporal consistency. Otherwise, the video algorithm could degrade due to the loss of correct temporal information. The temporal consistency is ensured by two preconditions: the real and synthetic frames sampled for distillation should be in the correct order and follow a uniform flow of time. More specifically, the orderedness is:
Precondition 1.
(Orderedness) During the compressed video distillation of real video and synthetic video , given two frame sequence pairs , , we define a partial order : iff any frame in occurs earlier than any frame in and any frame in occurs earlier than any frame in . The distillation process obeys orderedness iff all the frame sequence pairs for and yield a total ordering associated with .
That is, all frame sequence pairs are comparable and associated with and both real and synthetic sequences are in the correct time order. The matching strategy in Fig. 2(b) does not meet orderedness as the two matching pairs at right are in the wrong temporal order.
Precondition 2.
(Uniformity) Compressed video distillation of real video and synthetic video , obeys uniformity iff , .
All the real frame sequences have the same length, as the same to synthetic frame sequences. This ensures that the synthetic video we learn follows a uniform flow of time, e.g., the matching strategy in Fig. 2(c) leads to non-uniform FPS, and the synthetic frames at left will learn a much smaller FPS than the frames at right. We also justify the uniformity with experiments in the supplementary.
These two preconditions narrow the searching space of temporal condensation strategies, enabling us to systematically categorize and analyze the compressed video distillation. In Fig. 3(a), the naive algorithm may match all synthetic frames to real frames. To compress the temporal dimension, we can either reduce the number of frames (Fig. 3(b)) or reduce the correspondence by temporal segmentation (Fig. 3(c)). Note that we do not force orderedness and uniformity within each frame sequence pair for matching, since the distillation matching algorithms themselves could drive the synthetic data to follow the consistency.

3.2 Segmented Matching and Interpolation
To take a step further, we propose Segmented Matching and Interpolation framework for the quantitative comparison and analysis of temporal condensation (Fig. 3(e)). Given a target synthetic frame number , to approach flexible compression rate, we distill real frames to frames and interpolate the frames to our target . Specifically, the real video and synthetic video are segmented evenly for the pairwise distillation matching, which is the only valid strategy for temporal consistency. The interpolation enables the compression of synthetic video. This paradigm covers most video distillation methods from the perspective of temporal compression, and the extent of compression can be parameterized by the four factors:
(1) Number of Independent Synthetic Frames () is the size of the trainable synthetic frames. These frames are learned with dataset distillation algorithms and will be interpolated to the target length of synthetic video . Smaller indicates a larger temporal compression rate.
(2) Number of Real Frames () is the size of real frames for distillation matching. Larger implies a larger receptive field for synthetic video during distillation.
(3) Number of Segments (). We cut the real and synthetic videos into the same number of segments and apply distillation between the pairs of real and synthetic segments. Larger could reduce the training time since the distillation algorithm receives smaller segments with fewer frames.
(4) Interpolation Algorithm () interpolates the independent synthetic frames to the required synthetic video length. Our algorithms are detailed in Sec. 3.3.
So the level of temporal compression can be uniquely determined by a quadruplet . We show some examples in Fig. 4 with different combinations. In the following section, we will compare and analyze these four axes and put forward some empirical conclusions to drive future studies on video distillation.







3.3 Comparison of Temporal Compression
To investigate the effects of different temporal compression levels, we conduct comprehensive experiments with different , , , and . In the following experiments, the DM [30] algorithm is adopted and we use ConvNet (3-layer convolutional network from [31]) with one layer GRU head [5] as our backbone network.
Comparison of and values. We first compare the models with different (number of independent synthetic frames) and (number of real frames) and visualize in Fig. 5. We implement the segmented matching segment number . The performance is Fig. 5(a) shows that: (1) distilling the video to still image could yield decent accuracy (over 17%), (2) larger and brings minor performance gain (less than 3%), (3) the model degrades when (left-top of the figure), potentially due to insufficient temporal information in the real data. Fig. 5(b) and (c) show that larger and lead to significant memory consumption, e.g. model with takes GPU memory than a model. And models with large also take more training time. And we can also read Fig. 5 diagonally to fix the per-frame receptive field , and basically larger ratio leads to better performance.
Comparison of values (segments). We study the effects of in Fig. 6. The segmentation could notably reduce memory consumption (up to 40%) while maintaining the training speed. However, the efficiency is achieved at the cost of performance as the segmentation decreases the “receptive field” of each synthetic frame.
Comparison of Interpolation Algorithms . is critical to compressed video distillation, especially when is small. We use various interpolation methods: (1) Duplication is simply copying the learned synthetic frames to the required length, or namely nearest interpolation. e.g. a 2-frame video can be interpolated to 4-frame video . (2) Linear interpolation generates intermediate frames by blending adjacent reference frames. The frame at time will be the weighted sum of nearest reference frames , according to their temporal distance and . e.g. a 2-frame video can be interpolated to 4-frame video . (3) Parametric interpolator is a pre-trained interpolation network on the real video dataset. For each video data with frames, we evenly sample frames and duplicate them to length . We train a TimeSformer [2] model on these real data and it learns to recover the original video from the duplicated one. The pretrained model can be utilized for interpolation, e.g. a 2-frame video can be interpolated to 4-frame video .
and | 1 | 2 | 4 | 8 |
Duplication | ||||
Linear | - | |||
Parametric |
We compare the three interpolators in Tab. 1. The simple duplication method outperforms linear interpolation. The parametric method performs the best, especially on larger frame numbers, as it encodes dataset-specific inductive bias to compensate for the loss of temporal information.
Discussion. With the comparison of the four dimensions of temporal compression, we have some fundamental observations that could offer direction for our further study of video distillation: (1) Temporal compression confers significant advantages to dataset distillation and static images could encode more than little knowledge for video datasets; (2) Segmented distillation could reduce the distillation cost, but significantly sacrifice the model performance (3) Parametric interpolation could compensate for the loss of temporal dynamic information in the video.
4 Methodology
Dataset distillation is a lossy data compression process. With synthetic data as an intermediary, only part of the information in a real dataset could finally be learned by the model according to the data processing inequality. Thus, based on the analysis in Sec. 3.3 and considering the trade-off between efficiency and efficacy, we propose a video dataset distillation paradigm by disentangling the static and dynamic information in videos. We put more effort into the learning of static information with low cost (Sec. 4.1), and then compensate for the dynamic information (Sec. 4.2). We give an overview of our method in Fig. 7.

4.1 Static Learning
The results in Sec. 3.3 indicate that static information in videos is more critical to the distillation task, given the limited capacity of small synthetic data. Hence, we use a setting to learn a static memory with only one frame. The specific distillation process involves selecting one frame randomly from each video segment to form an image dataset in each epoch. The DC [31] method is then applied for gradient matching on a convolutional network. Since a new image dataset is created in each epoch, the “image” we distill has ideally observed all video frames and distilled the static memory from them.
4.2 Dynamic Fine-tuning
The choice of dynamic memory can be diverse. In this paper, our dynamic memory is represented as multiple frames of single-channel images. We use a network which takes static memory and dynamic memory as input and outputs video clips. At this stage, we fix the static memory and use different matching methods (performance matching, distribution matching, and parameter matching) to simultaneously update the network and dynamic memory.
We use a concrete formula to explain our paradigm. We refer as a matching loss of distillation where is synthetic dataset and indicates origin dataset , , . Given a frame selection method selecting frames from to obtain a dataset with , we summarize our paradigm in Alg. 1.
5 Experiments
Dataset | MiniUCF | HMDB51 | ||
IPC | 1 | 5 | 1 | 5 |
Full Dataset | 9.81 GB | 4.93 GB | ||
Random | 115 MB | 586 MB | 115 MB | 586 MB |
Herding [26] | ||||
K-Center [18] | ||||
DM [30] | 115 MB | 586 MB | 115 MB | 586 MB |
MTT [4] | ||||
FRePo [32] | ||||
Static-DC | 8 MB | 36 MB | 8 MB | 36 MB |
DM [30]+Ours | 94 MB | 455 MB | 94 MB | 455 MB |
MTT [4]+Ours | 94 MB | 455 MB | 94 MB | 455 MB |
FRePo [32]+Ours | 48 MB | 228 MB | 48 MB | 228 MB |
Dataset | Kinetics-400 | SSv2 | ||
IPC | 1 | 5 | 1 | 5 |
Full Dataset | ||||
Random | ||||
DM [30] | ||||
MTT [4] | ||||
Static-DC | ||||
DM[30]+Ours | ||||
MTT[4]+Ours |
5.1 Datasets and Metrics
We adopt small video datasets UCF101 [19] and HMDB51 [11], and large-scale Kinetics [3] and Something-Something V2 [8] in our experiments. UCF101 [19] consists of 13,320 video clips in 101 action categories while HMDB51 [11] consists of 6849 video clips in 51 action categories. Kinetics [3] is a collection of video clips that cover 400/600/700 human action classes while SSv2 [8] covers 174 motion-heavy classes. To evaluate the distillation algorithms on more diversified data scales, and considering the efficiency of experiments and the clarity of model comparisons, following the scale of pioneering work on image distillation [25], we build a miniaturized version of UCF101, named MiniUCF, including 50 most common classes from the UCF101 dataset. This miniaturization enables rapid iterations of our method and facilitates observing relatively significant changes in performance. We report the top-1 classification accuracy for MiniUCF and HMDB51, and the top-5 classification accuracy for Kinetics400 and SSv2.
5.2 Baselines
The baseline methods involve: (1) coreset selection methods (random selection, Herding [26] and K-Center [18]) following the implementation for image distillation in DC [31]. (2) direct adaptation of the common image distillation methods (DM [30], MTT [4], FRePo [32]) to the video distillation task. (3) image distillation method (DC [31]) with frame duplication for a “boring videos” proposed by us, namely “Static-DC”.
5.3 Implementation Details
Data. For MiniUCF and HMDB51, the videos are sampled to 16 frames with sampling interval 4 dynamically, . the frames indices vary in different epochs. Following the setup in C3D [21], each of these frames is cropped and resized to 112x112. For Kinetics-400 and Something-Something V2, the videos are sampled to 8 frames before the distillation, and the frames are cropped to 64x64. We only use horizontal flipping with a 50% probability for data augment.
Static Learning. We use DC [31] to distill static memory with random real frame initialization. We utilize a 4-layer 2D convolutional neural network for distillation (ConvNetD4) and perform an early stop in the distillation training when the loss converges. Interestingly, our experiments indicate that static memory is not necessarily trained to full convergence, as dynamic memory compensates for it.
Dynamic Finetuning. In dynamic fine-tuning, we adopt distillation methods in various types, including DM [30], MTT [4], and FRePo [32] to evaluate the broad applicability of our paradigm. Dynamic memory is initialized with random noise. We use a small 3D CNN (referred to as MiniC3D) for distillation. The network used for combining static and dynamic memory is also a MiniC3D. For more details, please refer to the supplementary.
Fair Comparison between the baseline and our method. We rigorously ensure that the total storage space for static memory, dynamic memory, and the network is smaller than the corresponding IPC (Instance Per Class). Specifically, on DM and MTT, we use no more than 82% of the storage space corresponding to the baseline, which amounts to 2 static memory with 2 dynamic memory for every instance. On FRePo, we use no more than 42% of the storage space corresponding to the baseline, which means 1 static memory with 1 dynamic memory for every instance.
Evaluation of Distilled Dataset. Naturally, we evaluate how well our synthetic data performs on architectures used to distill it. When evaluating data distilled by FRePo, the results should be considered as a reference only due to the label learning conducted by the method itself and the use of a different optimizer. We also evaluate how well our synthetic data performs on different architectures from the one used to distill it on the MiniUCF, 1 instance per class task.
Hyper-Parameters. Considering the numerous parameters involved in the experiments, we detail the parameter settings for all experiments in the supplementary.
5.4 Results
We show the results of our small-scale experiments in Tab. 2 and large-scale in Tab. 3. The full dataset indicates the accuracy of the network trained on the full dataset.
Comparison to Coreset Method. Following image distillation, we compare our method with coreset selection methods on MiniUCF and HMDB51 in Tab. 2. In the majority of cases, our approach outperforms the coreset selection. Regarding coreset methods, we also observe: (1) On average, the herding method proves to be the most effective coreset approach. (2) With an increase in IPC, the performance of herding exhibits a notable improvement. These conclusions align with earlier experiments [31] in image distillation, lending credibility to our findings.
Comparison to Other Methods. We compare our final method with other methods we proposed in Fig. 1. Among the three methods we proposed, Static-DC (Fig. 1(b)) exhibits the poorest performance. Compared to both the coreset method and Static-DC, the naively adapted method (Fig. 1(a)) shows a significant improvement. This underscores the applicability of existing image distillation techniques to videos and the effects of dynamic information in video understanding. In comparison, our final method (corresponding to Fig. 1(c)) achieves a remarkable superiority over all other methods through static and dynamic disentangle learning. In most cases, our method could enhance the current distillation methods while requiring less storage. However, the performance of our method on Kinetics-400 is not as strong as on the other two smaller datasets. This is because Kinetics-400 itself has a much larger number of categories and samples compared to the other two smaller datasets, which increases the difficulty and cost of distillation. Since our network is shared, having 400 categories sharing one network in Kinetics-400 might cause interference during learning. Allocating different networks to different categories or using a more complex network could potentially offer better improvement in our method. However, this approach would impose a greater training burden and significantly reduce the practical value of distillation, so we do not conduct larger-scale experiments.
Cross Architecture Generalization. We also show the result of cross-architecture generalization in Tab. 4. The experimental results indicate that the data obtained by static-dynamic disentanglement performs much better on other networks compared to the naively adapted method.
5.5 Ablation Study
Ratios of Static and Dynamic. We report results for MiniUCF 1 Instance per class task with different static and dynamic memory ratios in Tab. 5. We can find that increasing the quantity of static and dynamic memory both improves the scores. However, as their numbers increase, the time and GPU memory required for Static Learning and the training convergence time for Dynamic Fine-tuning under the same computational power will also increase (Tab. 5). Considering training efficiency and effectiveness while ensuring the storage does not exceed the corresponding IPC, a balanced choice is to use 2 static with 2 dynamics for every instance.
SPC | DPC | Acc (%) | Storage | S1 Time | S1 GPU Memory |
1 | 0 | 8 MB | 68 s/iter | 4,651 MiB | |
1 | 48 MB | ||||
2 | 79 MB | ||||
3 | 111 MB | ||||
2 | 0 | 14 MB | 156 s/iter | 6,579 MiB | |
1 | 54 MB | ||||
2 | 94 MB | ||||
3 | - | 117 MB |
Impact of Video Dynamics on Distillation. Actions in the video exhibit varying degrees of dynamics. To explore the impact of video dynamics, we categorized all classes of MiniUCF into two groups (relatively static, highly dynamic) based on their level of dynamics, calculated by the average Hamming distance between inter-frame features for each class, and compared their test accuracy on networks trained with distilled data (Fig. 8). We observe that (1) data distilled by the Static-DC method is more sensitive for static classes, which aligns with our expectations as this method generates data that lacks dynamic information, akin to a “boring video”. (2) The naively adapted MTT, in comparison, can distill more useful information but still shows higher scores for static classes than dynamic classes. (3) MTT+Ours, however, demonstrates better distillation of dynamic information compared to the previous methods and exhibits significant improvements on dynamic classes.
5.6 Visualization
To observe the temporal changes in the distilled videos, we sampled frames from the videos obtained using different methods and visualized their inter-frame differences. We show two examples in Fig. 9 and more in the supplementary. Although visually abstract, we can still conclude that the distilled videos indeed exhibit temporal variations.


6 Conclusion
In this work, we provide the first systematic study of video distillation. We propose a taxonomy to categorize methods based on key factors including the number of frames and segment length. With extensive experiments, we revealed that more frames provide marginal gains at greatly increased costs. Then, we proposed our method that disentangles static and dynamic information and achieves SOTA with efficient storage. We believe our paradigm will pave a new way for video distillation.
Acknowledgments
This work is supported in part by the National Natural Science Foundation of China under Grants 62306175, National Key Research and Development Project of China (No.2022ZD0160102, No.2021ZD0110704), Shanghai Artificial Intelligence Laboratory, XPLORER PRIZE grants.
References
- Baker et al. [2011] Simon Baker, Daniel Scharstein, JP Lewis, Stefan Roth, Michael J Black, and Richard Szeliski. A database and evaluation methodology for optical flow. IJCV, 92:1–31, 2011.
- Bertasius et al. [2021] Gedas Bertasius, Heng Wang, and Lorenzo Torresani. Is space-time attention all you need for video understanding? In ICML, page 4, 2021.
- Carreira and Zisserman [2017] João Carreira and Andrew Zisserman. Quo vadis, action recognition? a new model and the kinetics dataset. In CVPR, pages 4724–4733, 2017.
- Cazenavette et al. [2022] George Cazenavette, Tongzhou Wang, Antonio Torralba, Alexei A Efros, and Jun-Yan Zhu. Dataset distillation by matching training trajectories. In CVPR, 2022.
- Cho et al. [2014] Kyunghyun Cho, Bart Van Merriënboer, Dzmitry Bahdanau, and Yoshua Bengio. On the properties of neural machine translation: Encoder-decoder approaches. arXiv preprint arXiv:1409.1259, 2014.
- Cui et al. [2023] Justin Cui, Ruochen Wang, Si Si, and Cho-Jui Hsieh. Scaling up dataset distillation to imagenet-1k with constant memory. In ICML, pages 6565–6590, 2023.
- Deng and Russakovsky [2022] Zhiwei Deng and Olga Russakovsky. Remember the past: Distilling datasets into addressable memories for neural networks. In NeurIPS, 2022.
- Goyal et al. [2017] Raghav Goyal, Samira Ebrahimi Kahou, Vincent Michalski, Joanna Materzyńska, Susanne Westphal, Heuna Kim, Valentin Haenel, Ingo Fruend, Peter Yianilos, Moritz Mueller-Freitag, Florian Hoppe, Christian Thurau, Ingo Bax, and Roland Memisevic. The ”something something” video database for learning and evaluating visual common sense, 2017.
- Huang et al. [2018] De-An Huang, Vignesh Ramanathan, Dhruv Mahajan, Lorenzo Torresani, Manohar Paluri, Li Fei-Fei, and Juan Carlos Niebles. What makes a video a video: Analyzing temporal information in video understanding models and datasets. In CVPR, pages 7366–7375, 2018.
- Kim et al. [2022] Jang-Hyun Kim, Jinuk Kim, Seong Joon Oh, Sangdoo Yun, Hwanjun Song, Joonhyun Jeong, Jung-Woo Ha, and Hyun Oh Song. Dataset condensation via efficient synthetic-data parameterization. In ICML, 2022.
- Kuehne et al. [2011] H. Kuehne, H. Jhuang, E. Garrote, T. Poggio, and T. Serre. Hmdb: A large video database for human motion recognition. In ICCV, pages 2556–2563, 2011.
- Li and Li [2021] Yongqi Li and Wenjie Li. Data distillation for text classification. arXiv preprint arXiv:2104.08448, 2021.
- Liu et al. [2022] Songhua Liu, Kai Wang, Xingyi Yang, Jingwen Ye, and Xinchao Wang. Dataset distillation via factorization. In NeurIPS, 2022.
- Liu et al. [2021] Xin Liu, Silvia L. Pintea, Fatemeh Karimi Nejadasl, Olaf Booij, and Jan C. van Gemert. No frame left behind: Full video action recognition. In CVPR, pages 14892–14901, 2021.
- Loo et al. [2022] Noel Loo, Ramin Hasani, Alexander Amini, and Daniela Rus. Efficient dataset distillation using random feature approximation. In NeurIPS, 2022.
- Nguyen et al. [2020] Timothy Nguyen, Zhourong Chen, and Jaehoon Lee. Dataset meta-learning from kernel ridge-regression. arXiv preprint arXiv:2011.00050, 2020.
- Nguyen et al. [2021] Timothy Nguyen, Roman Novak, Lechao Xiao, and Jaehoon Lee. Dataset distillation with infinitely wide convolutional networks. In NeurIPS, 2021.
- Sener and Savarese [2018] Ozan Sener and Silvio Savarese. Active learning for convolutional neural networks: A core-set approach. In ICLR, 2018.
- Soomro et al. [2012] Khurram Soomro, Amir Roshan Zamir, and Mubarak Shah. UCF101: A dataset of 101 human actions classes from videos in the wild. CoRR, abs/1212.0402, 2012.
- Tong et al. [2022] Zhan Tong, Yibing Song, Jue Wang, and Limin Wang. Videomae: Masked autoencoders are data-efficient learners for self-supervised video pre-training. In NeurIPS, pages 10078–10093. Curran Associates, Inc., 2022.
- Tran et al. [2015] Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. Learning spatiotemporal features with 3d convolutional networks. In ICCV, pages 4489–4497, 2015.
- Tran et al. [2018] Du Tran, Heng Wang, Lorenzo Torresani, Jamie Ray, Yann LeCun, and Manohar Paluri. A closer look at spatiotemporal convolutions for action recognition. In CVPR, pages 6450–6459, 2018.
- Wang et al. [2022] Kai Wang, Bo Zhao, Xiangyu Peng, Zheng Zhu, Shuo Yang, Shuo Wang, Guan Huang, Hakan Bilen, Xinchao Wang, and Yang You. Cafe: Learning to condense dataset by aligning features. In CVPR, 2022.
- Wang et al. [2021] Mengmeng Wang, Jiazheng Xing, and Yong Liu. Actionclip: A new paradigm for video action recognition. CoRR, abs/2109.08472, 2021.
- Wang et al. [2018] Tongzhou Wang, Jun-Yan Zhu, Antonio Torralba, and Alexei A Efros. Dataset distillation. arXiv preprint arXiv:1811.10959, 2018.
- Welling [2009] Max Welling. Herding dynamical weights to learn. In ICML, 2009.
- Wu et al. [2023] Xindi Wu, Zhiwei Deng, and Olga Russakovsky. Multimodal dataset distillation for image-text retrieval. arXiv preprint arXiv:2308.07545, 2023.
- Xu et al. [2023] Zhe Xu, Yuzhong Chen, Menghai Pan, Huiyuan Chen, Mahashweta Das, Hao Yang, and Hanghang Tong. Kernel ridge regression-based graph dataset distillation. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 2850–2861, 2023.
- Zhao and Bilen [2021] Bo Zhao and Hakan Bilen. Dataset condensation with differentiable siamese augmentation. In ICML, 2021.
- Zhao and Bilen [2023] Bo Zhao and Hakan Bilen. Dataset condensation with distribution matching. In WACV, 2023.
- Zhao et al. [2020] Bo Zhao, Konda Reddy Mopuri, and Hakan Bilen. Dataset condensation with gradient matching. arXiv preprint arXiv:2006.05929, 2020.
- Zhou et al. [2022] Yongchao Zhou, Ehsan Nezhadarya, and Jimmy Ba. Dataset distillation using neural feature regression. arXiv preprint arXiv:2206.00719, 2022.
Supplementary Material
1
7 Selection of MiniUCF
We train and test on split 1 of UCF101 on MiniC3D and select the top 50 classes based on accuracy. MiniUCF can reach an accuracy of 57.2% on MiniC3D. We provide the 50 categories we have selected in Suppl. Tab. 6.
ApplyEyeMakeup | BalanceBeam | BandMarching |
BaseballPitch | Basketball | BasketballDunk |
Biking | Billiards | BlowingCandles |
Bowling | BreastStroke | CleanAndJerk |
CliffDiving | CricketShot | Diving |
FloorGymnastics | FrisbeeCatch | GolfSwing |
HammerThrow | HighJump | HorseRace |
HorseRiding | HulaHoop | IceDancing |
JumpingJack | Knitting | MilitaryParade |
Mixing | ParallelBars | PlayingPiano |
PlayingViolin | PoleVault | PommelHorse |
Punch | Rafting | Rowing |
SkateBoarding | Skiing | Skijet |
SkyDiving | SoccerPenalty | StillRings |
SumoWrestling | Surfing | Swing |
TennisSwing | TrampolineJumping | UnevenBars |
VolleyballSpiking | WritingOnBoard |
8 Ablation Study of Uniformity
We conduct experiments to justify the necessity of uniformity in Supp. Tab. 7. Eight real frames are split evenly or non-evenly into 4 segments and aligned to 4 synthetic frames, among which uniform segmentation achieves the best performance.
Segment Sizes | Acc. | Segment Sizes | Acc. |
1,1,5,1 | 15.80.2 | 1,1,1,5 | 15.70.5 |
1,1,2,4 | 14.10.2 | 1,4,1,2 | 15.10.2 |
3,1,2,2 | 17.60.7 | 1,2,2,3 | 16.80.1 |
2,2,2,2 (uniform) | 17.60.2 |
SPC | DPC | Acc | S_Acc | D_Acc |
1 | 0 | 13.7 | 14.9 | 13.0 |
1 | 17.5 | 18.0 | 16.9 | |
2 | 19.6 | 21.1 | 17.1 | |
3 | 20.6 | 23.1 | 19.5 | |
2 | 1 | 22.3 | 23.0 | 21.4 |
2 | 23.3 | 24.1 | 23.1 |
9 Impact of Video Dynamics on Distillation
9.1 Selection of Static and Dynamic Group
We utilize a pre-trained four-layer 2D convolutional network to extract features from individual frames. Subsequently, we compute the Hamming distance between the features of consecutive frames. Then, we derive the average inter-frame Hamming distance for each video segment, allowing us to ascertain the average inter-frame Hamming distance for each class. We classify 50% of the classes with smaller average inter-frame Hamming distances into the static group, while the remaining 50% with larger distances are designated as the dynamic group. In Suppl. Tab. 6, we highlight the static and dynamic groups using distinct colors.
9.2 More Detailed Results
We show more detailed results in Suppl. Tab. 8. We can observe that both increasing static memory and dynamic memory concurrently enhance the accuracy of static and dynamic classes. Additionally, we note that when the disparity in quantity between static memory and dynamic memory becomes larger, there is an imbalance in the accuracy of the static group and dynamic group. The reason for this lies in the random pairing of dynamic memory and static memory during sampling. If there is an excess of dynamic memory, multiple instances of dynamic memory may end up learning the same content during training, essentially augmenting the static information. Therefore, we recommend maintaining a 1:1 ratio between static memory and dynamic memory when utilizing our paradigm.

10 Visualization
10.1 Optical Flows
10.2 Inter-frame Differences
11 Implementation Details
11.1 Network Structure
In this section, we provide a detailed introduction to models used in the experiments, including MiniC3D, CNN+GRU, and CNN+LSTM. Additionally, we compare their performance with C3D[21] on the UCF101[19] and HMDB51[19] classification tasks.

MiniC3D | CNN+GRU | CNN+LSTM | C3D⋆[21, 3] | |
UCF101 [19] | 33.7 | 30.4 | 27.92 | 51.6 |
HMDB51 [11] | 28.6 | 24.0 | 23.2 | 24.3 |
MiniC3D.
Suppl. Fig. 11(a) shows the structure of MiniC3D. For Conv3D blocks except for Conv3D 4, we use 37 7 kernels with 1 2 2 strides and 1 3 3 paddings. Conv3D 4 is used for classification, which has 11 1 kernel and 1 1 1 stride. Channels are denoted below block names in Suppl. Fig. 11. For pooling operations, we employ a 12 2 kernel for max-pooling in pool3D 1, and 2 2 2 kernels for max-pooling in both pool3D 2 and pool3D 3. In contrast, we utilize average pooling in pool3D 4.
CNN+GRU.
Suppl. Fig. 11(b) shows the structure of CNN+GRU. For Conv blocks, we use 33 kernels with 1 1 strides and 11 paddings. Channels are denoted below block names in Suppl. Fig. 11. For pooling operations, we employ 2 2 kernels for average pooling in both pool 1, pool 2, and pool 3. The GRU block is a single-layer GRU. A linear layer is used for classification.
CNN+LSTM.
Suppl. Fig. 11(c) shows the structure of CNN+LSTM. For Conv blocks, we use 33 kernels with 1 1 strides and 1 1 paddings. Channels are denoted below block names in Suppl. Fig. 11. For the pooling operations, we employ 2 2 kernels for average pooling in both pool 1, pool 2, and pool 3. The LSTM is a single-layer LSTM. A linear layer is used for classification.
Comparison With Full Model.
We show the classification results of UCF101 and HMDB51 on these models and full C3D in Suppl. Tab. 9. In Suppl. Tab. 9, we observe that (1) on MiniC3D, the classification accuracy of UCF101 is lower than that on full C3D; (2) on MiniC3D, the classification accuracy of HMDB51 can even exceed that on full C3D.
Dataset | IPC | lr_img | batch_syn | syn_steps |
MiniUCF | 1 | 1e5 | 50 | 10 |
5 | 1e5 | 128 | 5 | |
HMDB51 | 1 | 1e4 | 51 | 10 |
5 | 1e6 | 128 | 5 | |
Kinetics400 | 1 | 5e5 | 256 | 10 |
5 | 1e7 | 256 | 5 | |
SSv2 | 1 | 1e5 | 256 | 10 |
5 | 1e6 | 256 | 5 |
Dataset | IPC | lr_dynamic | lr_hal | batch_syn | syn_steps |
MiniUCF | 1 | 1e4 | 1e-3 | 50 | 10 |
5 | 1e4 | 1e-3 | 128 | 5 | |
HMDB51 | 1 | 1e5 | 1e-2 | 51 | 10 |
5 | 1e6 | 1e-2 | 128 | 5 | |
Kinetics400 | 1 | 1e3 | 1e-2 | 256 | 10 |
5 | 1e7 | 1e-2 | 256 | 5 | |
SSv2 | 1 | 1e4 | 1e-2 | 256 | 10 |
5 | 1e5 | 1e-2 | 256 | 5 |
11.2 Details of Temporal Analysis
The temporal analysis experiments in Sec. 3.3 are conducted with DM [30] algorithm and CNN+GRU model as detailed before. 16 frames are sampled from each video with temporal stride 12, and we set the target synthetic video length also 16. For a fair comparison of time and space complexity, all the experiments are run on one NVIDIA V100 GPU (16GB), 8 cores of Intel Xeon 5218 CPU, and 20 GB memory. The learning rate for synthetic images is set to 1.0 and that for network updating is set to 0.01. The models are trained for 10,000 iterations with a real batch size of 64, which is observed as enough for the training convergence in our experiments.
Dataset | IPC | lr_img | batch_real |
MiniUCF | 1 | 30 | 64 |
5 | 100 | 64 | |
HMDB51 | 1 | 30 | 64 |
5 | 300 | 64 | |
Kinetics400 | 1 | 100 | 64 |
5 | 500 | 128 | |
SSv2 | 1 | 10 | 64 |
5 | 100 | 128 |
Dataset | IPC | lr_dynamic | lr_hal | batch_real |
MiniUCF | 1 | 1e-4 | 1e-5 | 64 |
5 | 1e3 | 1e-6 | 64 | |
HMDB51 | 1 | 10 | 1e-6 | 64 |
5 | 10 | 1e-5 | 64 | |
Kinetics400 | 1 | 1 | 1e-5 | 64 |
5 | 10 | 1e-5 | 128 | |
SSv2 | 1 | 1 | 1e-5 | 64 |
5 | 100 | 1e-5 | 128 |
Dataset | IPC | lr_img |
MiniUCF | 1 | 1e-3 |
5 | 1e-3 | |
HMDB51 | 1 | 1e-3 |
5 | 1e-3 |
Dataset | IPC | lr_dynamic | lr_hal |
MiniUCF | 1 | 1e-4 | 1e-3 |
5 | 1e-1 | 1e-3 | |
HMDB51 | 1 | 1e-1 | 1e-3 |
5 | 1e-2 | 1e-3 |
11.3 Details of Full Experiments
In the experiments, we initially fine-tune the parameters for methods without our paradigm (naively adapted methods) to achieve the best possible results. To ensure a fair comparison, we strive to maintain consistency in all other parameter settings, making adjustments only to the learning rates associated with the unique dynamic information and in the methods with our paradigm.
Structure of .
We employ two different . The simple one only has one Conv3D block with 33 3 kernels, 1 1 1 stride, and 1 1 1 padding, while the other one has one Conv3D block and one ConvTranspose3d block with middle channel 8. With the exception of DM+Ours for MiniUCF IPC=1, we consistently utilize the former.
Hyper-parameters for Distillation.
We have thoroughly documented the parameters employed in our experiments in Suppl. Tab. 10 11 12. Parameters not explicitly mentioned default to the values specified in the original implementation code. The specific meanings of all mentioned parameters are detailed below:
lr_img: learning rate used to update distilled video.
lr_teacher: learning rate used to train expert trajectories of real videos and training trajectories of distilled videos. We set it to 0.01 by default.
batch_syn: number of distilled videos to match real videos at every iteration.
batch_real: number of real videos to be matched at every iteration.
syn_steps: steps of training trajectories of distilled videos to match expert trajectories at every iteration.
lr_dynamic: learning rate used to update dynamic memory.
lr_hal: learning rate used to update .
expert_epochs: steps of expert trajectories to be matched at every iteration. We set it to 1 by default.
max_start_epoch: the max starting step of expert trajectories to be matched. We set it to 10 by default.
We train 30 expert trajectories for MiniUCF and HMDB51 [11], and 20 for Kinetics400 [3] and Something-Something V2 [8]. Regarding whether to update lr_teacher during the training process, we retain relatively better results for each task, such as in the case of MTT for MiniUCF IPC=1, where the performance without updating lr_teacher surpasses that with updating.
Hyper-parameters for Evaluation.
For evaluation on FRePo and FRePo+Ours, we set the learning rate to 1e-4 and trained for 1,000 epochs on MiniC3D. For other evaluations, we configure the learning rate to be 1e-2 and conduct training for 500 epochs on MiniC3D. In the case of the cross-architecture generalization test on CNN+GRU and CNN+LSTM, we set the learning rate to 1e-2 and trained for 100 epochs.

