Duong H. [email protected]
\addauthorTrung-Nhan [email protected]
\addauthorNam [email protected]
\addinstitution
Ho Chi Minh City University of Technology
Ho Chi Minh, Vietnam
Pay Attention to Snapshots of Pruning
Paying more Attention to Snapshots of Iterative Pruning: Improving Model Compression via Ensemble Distillation
Abstract
Network pruning is one of the most dominant methods for reducing the heavy inference cost of deep neural networks. Existing methods often iteratively prune networks to attain high compression ratio without incurring significant loss in performance. However, we argue that conventional methods for retraining pruned networks (i.e., using small, fixed learning rate) are inadequate as they completely ignore the benefits from snapshots of iterative pruning. In this work, we show that strong ensembles can be constructed from snapshots of iterative pruning, which achieve competitive performance and vary in network structure. Furthermore, we present a simple, general and effective pipeline that generates strong ensembles of networks during pruning with large learning rate restarting, and utilizes knowledge distillation with those ensembles to improve the predictive power of compact models. In standard image classification benchmarks such as CIFAR and Tiny-Imagenet, we advance state-of-the-art pruning ratio of structured pruning by integrating simple -norm filters pruning into our pipeline. Specifically, we reduce 75-80% of total parameters and 65-70% MACs of numerous variants of ResNet architectures while having comparable or better performance than that of original networks. Code is available at https://github.com/lehduong/kesi.
1 Introduction
Motivation Researchers have extensively exploited deep and wide networks for the sake of achieving superior performance on various tasks. Most of state-of-the-art networks are extremely computationally expensive and require excessive memory. However, real-world applications usually require running deep neural networks on edge devices for various reasons: user privacy, security, real-time analysis, offline capability, reducing cost for server deployment, and so on. Adopting large and cumbersome networks to such resource-constrained environments is challenging due to restrictions of memory, computational power, energy consumption, and so on.
Background
Network pruning [LeCun et al. (1990), Reed (1993), Han et al. (2015), Li et al. (2016)] reduces a cumbersome and over-parameterized network to compact one by removing unnecessary weights and connections of networks. It is widely believed that small networks pruned from large, over-parameterized networks achieve superior performance than those trained from scratch [Frankle and Carbin (2018), Renda et al. (2020), Li et al. (2016), Luo et al. (2017)]. A plausible explanation to this phenomenon is the lottery ticket hypothesis [Frankle and Carbin (2018)] i.e. large, over-parameterized networks contain many optimal sub-networks i.e. winning tickets. In particular, network pruning could be done in two manners: one-shot pruning - prune a network with the desired compression ratio and retrain it only one time, or iterative pruning - only prune small ratio of the original network, retrain and repeat that process until the target size is reached. It has been shown that iterative pruning could lead to a greater compression ratio compare to one-shot pruning approaches [Han et al. (2015), Luo et al. (2017), Li et al. (2016), Renda et al. (2020)]. Furthermore, Frankle et al\bmvaOneDot[Frankle and Carbin (2018)] point out that iteratively-pruned-winning-tickets learn faster and reach higher test accuracy at smaller network size.
On the other hand, ensembles of neural networks are known to be much more robust and accurate than individual networks [Huang et al. (2017), Ashukha et al. (2020), Snoek et al. (2019)]. In spite of their superior performance, the tremendous cost of training and inference of ensembles makes them less attractive in practice. For the purpose of accelerating training time of ensembles, prior works proposed methods encouraging models to converge to different local minimums during training [Huang et al. (2017), Garipov et al. (2018), Yang et al. (2019b)]. To reduce inference time of ensembles, one could use a single network to mimic behavior of ensembles as pioneered by born-again tree [(6)] and knowledge distillation [Hinton et al. (2015), Balan et al. (2015), Buciluǎ et al. (2006), Malinin et al. (2019)]. In above approaches, although small networks can not achieve comparable performance with ensembles of networks, dark knowledge transferred from teachers to student network could bridge the gap between their predictive powers.
Our proposal
While existing methods of iterative pruning are more effective than one-shot pruning, the snapshots at each pruning iteration are mostly overlooked. We consider leveraging the snapshots of iterative pruning to take the performance of compact models to the next level.
In this work, we propose a simple pipeline for model compression by slightly modifying the standard approach. Specifically, we make use of large learning rate restarting at each pruning iteration to retrain pruned networks. Hence, each retraining step could be considered as a cycle of Snapshot ensemble [Huang et al. (2017)]. Utilizing both large learning rate restarting and pruning foster the diversity between snapshots, thus, constructing strong ensembles. Once achieved the desired compression ratio, we then distill the knowledge from the ensembles of snapshots of iterative pruning to the final model. Our method acquires the advantages of network pruning, ensembles learning, and knowledge distillation. To the best of our knowledge, this is the first work attempting to exploit snapshots of iterative pruning to further improve the performance of pruned networks.
Our main contributions
The contributions of our work are summarized as below:
-
1.
We empirically show that fine-tuning with large learning rate restarting can achieve competitive or better results than the common strategy (i.e. small, fixed learning rate) on a range of standard datasets and architectures. Surprisingly, such simple modification can create very strong baselines for both structured and unstructured pruning.
-
2.
We demonstrate that snapshots of iterative pruning could construct strong ensembles.
-
3.
We propose a simple pipeline to combine knowledge distillation from ensembles and iterative pruning. We empirically show that our approach can achieve state-of-the-art pruning ratio by reducing of parameters and MACs on numerous variants of ResNet while having comparable or better results than original networks.
2 Related Work
Knowledge Distillation
The approach of training small, efficient student network to mimic behavior of large, over-parameterized network has been proposed for a long time [Buciluǎ et al. (2006)] and was recently repopularized in [Hinton et al. (2015), Ba and Caruana (2014)]. Later, knowledge distillation was extended to various aspects, transferring knowledge from intermediate layers [Romero et al. (2014), Zagoruyko and Komodakis (2016a)], allowing teachers and students to guide each others [Zhang et al. (2018)], using teacher and student with the same architecture [Furlanello et al. (2018), Yang et al. (2019b), Yang et al. (2019a), Bagherinezhad et al. (2018)], distilling knowledge in multiple steps [Mirzadeh et al. (2019)]. To address the cost of training two networks in knowledge distillation, [Zhu et al. (2018), Zhang et al. (2018), Yang et al. (2019b)] propose online approaches to train the student and teacher networks in one generation. Furthermore, Anil et al\bmvaOneDot[Anil et al. (2018)] adopt knowledge distillation to accelerate the training of large scale neural networks. Universally Slimmable networks [Yu and Huang (2019)] provide an ensemble of sub-networks that has implicit knowledge distillation through shared weights.
Network Pruning
The idea behind network pruning is to reduce the redundant weights and connections of original network to achieve compact networks without losing much performance [Han et al. (2015), Li et al. (2016)]. In general, pruning can be divided into two categories: structured pruning and unstructured pruning. Unstructured pruning [Hanson and Pratt (1989), LeCun et al. (1990), Han et al. (2015), Srinivas and Babu (2015), Guo et al. (2016)] always results in sparse weight matrices, which can not directly accelerate the inference efficiency without specialized hardware/libraries. In contrast, structured pruning approaches [Li et al. (2016), He et al. (2017), Yu et al. (2018), Lin et al. (2019), Molchanov et al. (2016)] remove the redundant weights at the level of filters/channels/layers, thus, speeding up the inference of networks directly. There are numerous approaches to determine redundant filters/weights: [Luo et al. (2017)] use statistic information of the next filters to select unimportant filters, [Li et al. (2016)] prune the filters that have smallest norms in each layer, [Molchanov et al. (2016)] select the filters to minimize the construction loss estimated with Taylor expansion. As these criteria are rough estimations of weight’s importance, pruning a large number of filters/weights at once might break down and lead to inferior performance compare to iterative pruning [Han et al. (2015), Li et al. (2016)]. Recently, Liu et al\bmvaOneDot[Liu et al. (2018)] empirically show that training the pruned model from scratch can also achieve comparable or even better performance than fine-tuning. While the efficacy of network pruning remains an open question, in this work, we propose exploiting the benefit of having multiple networks through iterative pruning for constructing ensembles of networks.
3 Knowledge Distillation
Consider the classification problem in which we need to determine the correct category for input image among classes. The probability of class for sample given by neural network parameterized by is computed as:
(1) |
Where is the temperature of softmax function, higher values of lead to softer output distribution. Conventional approaches optimize the parameters by sampling mini-batches from the dataset and update the parameters to minimize cross-entropy objective:
(2) |
The target distribution of a sample is usually represented by one-hot vector i.e. only the true class is and all other classes are . Since input images might differ in term of noise, complexity, and multi-modality, enforcing networks to excessively fit the delta distribution of ground truth for all samples might deteriorate their generalization. Besides that, the similarity between classes provides rich information for learning and potentially prevent overfitting [Yang et al. (2019a)]. Knowledge distillation [Buciluǎ et al. (2006), Hinton et al. (2015)] uses a trained (teacher) network, which usually has high capacity, to guide the training of other (student) network. Let be the probability of class for image given by the teacher network, which is parameterized by . The objective function of knowledge distillation is defined as:
(3) |
In case the teacher is an ensemble of networks, the target distribution of knowledge distillation is the average of outputs of all networks: .
An alternative approach is optimizing the mean of Kullback-Leibler divergence between the student and each teacher network:
(4) |
We experimented with two above objectives but did not observe significant difference in performance of student networks, thus, we only report results of the second approach.
4 Snapshots of Iterative Pruning

In contrast to previous works, which mainly focus on the aforementioned usage of iterative pruning (i.e. alleviating the noise of weight’s importance estimation), we exploit the benefits of generating multiple models varying in structure and capacity to construct strong ensembles.
Inspired by the prior works of [Smith (2015), Loshchilov and Hutter (2016)] in which the authors show that promising local optimums could be found in a small number of epochs after restarting the learning rate. Furthermore, Huang et al\bmvaOneDot[Huang et al. (2017)] demonstrate that utilizing large learning rate restarting during training can construct strong ensembles without much additional cost.
Broadly speaking, the performance of ensembles depends on: the performance of individual network and the diversity of them. On the other hand, network pruning generates snapshots varying in structure and achieving competitive performance. Hence, if pruned networks could achieve minimal loss in predictive power relative to the original network, the ensemble of them could potentially outperforms the ensemble of networks having identical architecture (and trained with large learning rate restarting).
Prior works such as [Han et al. (2015), Liu et al. (2018), Molchanov et al. (2016)] retrain the pruned networks for more epochs with a fixed learning rate, which is usually the final learning rate of the training. However, this approach might result in multiple snapshots being stuck in similar local optimums, thus, leading to very weak ensembles as shown in our experiments. Similar to [Huang et al. (2017)], we adopt the large learning rate restarting at each pruning iteration to encourage each snapshot to converge to different optimum. For learning rate restarting, we utilize the One-cycle policy [Smith and Topin (2019)], which is proved to increase convergence speed of several models. Due to the similarity of our proposed method and Snapshot Ensembling [Huang et al. (2017)], we refer to each pruning and retraining step as a cycle. One-cycle policy adjusts learning rate at each mini-batch update and has two phases:
Increasing learning rate The learning rate and momentum of optimizer will be initialized to and respectively. During the first iterations of fine-tuning, learning rate and momentum gradually increase from initial values to , . The learning rate and momentum at -th step with cosine annealing strategy are given by:
(5) |
(6) |
Decreasing learning rate After iterations, learning rate and momentum will be gradually decreased from and to and in iterations where is total number of iterations for fine-tuning.
(7) |
(8) |
It is worth noticing that differs from previous works [Huang et al. (2017), Yang et al. (2019b)], which use cosine annealing schedule, by using One-cycle policy, we also "warm-up" learning rate at the start of each cycle. In our experiments, warming up learning rate is extremely important to achieve high accuracy with deep and large networks.
Surprisingly, retraining with One-cycle policy does not only generate significantly stronger ensembles, but also consistently outperforms the standard strategy for finetuning in terms of predictive accuracy of individual snapshots. We hypothesize that the (local) optimums of pruned networks are actually far from those of original networks, thus, large learning rate is needed to guarantee the convergence of pruned networks. We leave rigorous evaluation to investigate this phenomenon for future works.
5 Effective Pipeline for Model Compression
Since we already obtain strong ensembles during pruning, it is straightforward to distill the knowledge from them to the final pruned network. Our proposed pipeline can be summarized as follow:
-
1.
Train the baseline model to completion.
Prune redundant weights of the network based on some criteria.
Retrain the pruned network with large learning rate.
Repeat step 2 and 3 until desired compression ratio is reached.
Distill knowledge from ensembles of snapshots of pruning.
From now, we refer to our pipeline for model compression as Knowledge Distillation from Ensembles of Snapshots of Iterative Pruning (KESI). An overview of our approach is depicted in Figure 1. Our approach is extremely simple, easy to implement and can be adopted with any pruning mechanisms. We discuss the reasons why ensembles of snapshots of pruning are naturally suited for knowledge distillation.
Quality of Teacher In knowledge distillation, student can either learn to jointly optimize the supervised loss (Equation 2) and knowledge distillation loss (Equation 4) or only optimize the distillation objective. In the former case, if the teacher is poorly trained, mathematically speaking, the two objectives will conflict with each other. In the latter case, a poor teacher provides weak supervision (noisy label), making it’s harder to learn from the student’s perspective. Furthermore, ensembles provide more robust predictions on noisy labeled datasets [Lee and Chung (2019)] and out-of-distribution examples [Lakshminarayanan et al. (2017)].
Student and Teacher Gap Although ensembles of snapshots have superior performance than the original network, it is not sufficient to guarantee the improvement in the performance of the student network with Knowledge Distillation. In fact, many works such as [(33), Cho and Hariharan (2019), Yang et al. (2019a)] show that a powerful teacher might impair the performance of its student if there is a large gap between their predictive powers. However, ensembles of snapshots of pruning consist of models varying in capacity. Hence, teacher’s predictions of hard-to-learn samples (because of their complexity, multi-modality) will have softer distributions as the small networks could not "remember" those samples and would be more uncertain about them.
In this work, we only investigate knowledge distillation from ensembles of fixed-weights teachers, however, we can also jointly train all models and allow them to guide each other, which is referred to as deep mutual learning [Zhang et al. (2018)].
6 Experiments
We conduct experiments on CIFAR-10, CIFAR-100 [Krizhevsky et al. (2009)] and Tiny-Imagenet 111https://tiny-imagenet.herokuapp.com datasets.
The two CIFAR datasets [Krizhevsky et al. (2009)] consist of colored natural images sized at pixels. CIFAR-10 (C10) and CIFAR-100 (C100) images are drawn from 10 and 100 classes, respectively. For each dataset, there are 50,000 training images and 10,000 images reserved for testing.
The Tiny ImageNet dataset consists of a subset of ImageNet images [Deng et al. (2009)]. There are 200 classes, each of which has 500 training images and 50 validation images. Each image is resized to and augmented with random crops, horizontal mirroring, and RGB intensity scaling.
We run each experiment 3 times then report mean and standard deviation of each network. In our experiments, we prune all networks in 5 cycles unless otherwise stated.
6.1 Experiment setup
Training baselines
We adopt the training and pruning code from [Liu et al. (2018)] 222https://github.com/Eric-mingjie/rethinking-network-pruning. We train all networks with Stochastic Gradient Descent (SGD), learning rate is dropped from to at training and to at . The batch size is set to and weight decay is similar to [He et al. (2016a), He et al. (2016b)].
CIFAR In order to create strong baseline models, we extend the training schedule of all models to 300 epochs. For WideResnet, we use same configurations as described in [Zagoruyko and Komodakis (2016b)].
Tiny-Imagenet we adopt Pytorch’s pretrained models on ImageNet and only replace the last fully-connected layer and train networks for more epochs. We warm up learning rate from to in epochs. Other configurations are adopted from CIFAR training recipe.
Pruning
Structured pruning we use -norm based filters pruning [Li et al. (2016)] for simplicity. In each layer, a fixed number of filters having smallest -norm will be pruned. Since the bulk of networks tend to be last layers, we increase the percentage of filters that will be pruned as the layer goes deeper to achieve higher compression ratio.
Unstructured pruning, we exploit (global) magnitude-based weight pruning [Han et al. (2015)] i.e. pooling parameters across all layers and pruning weights with lowest magnitude. Specifically, we only prune parameters of convolutional layers similar to (Liu et al., 2018).
Retraining
The budget for fine-tuning of each cycle is and epochs on CIFAR and Tiny-Imagenet datasets respectively regardless of model architectures. In standard policy, the learning rate is set to 0.001 and fixed during retraining.
For One-cycle policy, we set the initial learning rate , gradually increase it to the maximum learning rate in of total (retrain) epochs, then decrease it to the minimum learning rate for remaining epochs. Other configurations are identical to those of training.
Knowledge Distillation
We use Adam optimizer Kingma and Ba (2014) for ensemble distillation since it gives better results than vanilla SGD in our experiments. For knowledge distillation, we also adopt One-cycle policy where we set to respectively. We do not explicitly use regularization for knowledge distillation. Other configurations e.g\bmvaOneDotbatch size, number of retraining epochs,… are similar to normal finetuning.
In our experiments, we use temperature . The teachers i.e. ensembles of snapshots consist of 6 models including the original (unpruned) network and 5 snapshots of pruning.
6.2 Results
6.2.1 Effectiveness of large learning rate
![]() |
![]() |
![]() |
(a) Resnet-56 on CIFAR-10 | (b) Resnet-110 on CIFAR-10 | (c) WideResnet-16-8 on CIFAR-10 |
![]() |
![]() |
![]() |
(a) Resnet-56 on CIFAR-100 | (b) Resnet-110 on CIFAR-100 | (c) WideResnet-16-8 on CIFAR-100 |
![]() |
![]() |
![]() |
(a) Resnet-56 on CIFAR-10 | (b) Resnet-110 on CIFAR-10 | (c) WideResnet-16-8 on CIFAR-10 |
We conduct experiments to empirically evaluate the performance of pruned networks trained with large learning rate compare to networks fine-tuned with small learning rate. Figure 2 and 3 demonstrate results of pruned networks with different compression ratios for both structured and unstructured pruning. Exhaustive results are reported in supplementary documents.
6.2.2 Performance of ensembles of snapshots
We compare the performance of ensembles of snapshots with different approaches: snapshots of pruned networks trained with small learning rate, snapshots of pruned networks trained with large learning rate restarting and snapshots of unpruned networks retrained with large learning rate (i.e. all snapshots have same architecture as the original network). Figure 4 presents the result of this experiment.
We can see that although the network capacity is decreased at each cycle, the ensembles of snapshots of iterative pruning achieve competitive or even better than snapshots of networks with same architecture. Detailed results of performance of ensembles are reported in supplementary documents.
6.2.3 Performance of compact networks trained with our pipeline

Model | Methods | % Params | % FLOPs | baseline | pruned |
---|---|---|---|---|---|
Resnet-56 | CP (He et al., 2017) | - | 50.6 | 92.80 | 91.80 |
FPEC (Li et al., 2016) | 14.1 | 27.6 | 93.04 | 93.06 | |
NISP (Yu et al., 2018) | 42.4 | 35.5 | 93.26 | 93.01 | |
GAL-0.8 (Lin et al., 2019) | 65.9 | 60.2 | 93.26 | 91.58 | |
GBN (You et al., 2019) | 66.7 | 70.3 | 93.10 | 93.07 | |
HRank (Lin et al., 2020) | 42.4 | 50.0 | 93.26 | 93.17 | |
PFEC+KESI (our) | 67.1 | 61.5 | 93.42 | ||
Resnet-110 | PFEC (Li et al., 2016) | 32.6 | 38.7 | 93.53 | 93.30 |
GAL-0.5 (Lin et al., 2019) | 44.8 | 48.5 | 93.50 | 92.74 | |
HRank (Lin et al., 2020) | 68.7 | 68.6 | 93.52 | 92.65 | |
PFEC+KESI (our) | 77.5 | 65.4 | 94.01 |
Model | Structured Pruning | Unstructured Pruning | |||||||
---|---|---|---|---|---|---|---|---|---|
Method | # Params(M) | % MACs | C10 | C100 | Method | # Params(M) | C10 | C100 | |
Resnet-56 | baseline | 0.85 | 0.00 | 93.42 | 71.07 | baseline | 0.85 | 93.42 | 71.07 |
PFEC Li et al. (2016) | 0.28 | 61.5 | MWP Han et al. (2015) | 0.29 | |||||
PFEC+One-cycle | 0.28 | 61.5 | MWP+One-cycle | 0.29 | |||||
PFEC+KESI(our) | 0.28 | 61.5 | MWP+KESI(our) | 0.29 | |||||
Resnet-110 | baseline | 1.73 | 0.00 | 94.01 | 72.35 | baseline | 1.73 | 94.01 | 72.35 |
PFEC Li et al. (2016) | 0.39 | 65.38 | MWP Han et al. (2015) | 0.36 | |||||
PFEC+One-cycle | 0.39 | 65.38 | MWP+One-cycle | 0.36 | |||||
PFEC+KESI(our) | 0.39 | 65.38 | MWP+KESI(our) | 0.36 | |||||
Preresnet-164 | baseline | 1.70 | 0.00 | 95.06 | 76.35 | baseline | 1.70 | 95.06 | 76.35 |
PFEC Li et al. (2016) | 0.31 | 69.23 | MWP Han et al. (2015) | ||||||
PFEC+One-cycle | 0.31 | 69.23 | MWP+One-cycle | ||||||
PFEC+KESI(our) | 0.31 | 69.23 | MWP+KESI(our) | ||||||
WideResnet-16-8 | baseline | 10.96 | 0.00 | 95.62 | 79.57 | baseline | 10.96 | 95.62 | 79.57 |
PFEC Li et al. (2016) | 2.48 | 64.52 | MWP Han et al. (2015) | 2.53 | |||||
PFEC+One-cycle | 2.48 | 64.52 | MWP+One-cycle | 2.53 | |||||
PFEC+KESI(our) | 2.48 | 64.52 | MWP+KESI(our) | 2.53 | |||||
VGG-16 | baseline | 14.99 | 0.00 | 94.23 | 73.24 | baseline | 14.99 | 94.23 | 73.24 |
PFEC Li et al. (2016) | 2.71 | 45.16 | MWP Han et al. (2015) | 1.02 | |||||
PFEC+One-cycle | 2.71 | 45.16 | MWP+One-cycle | 1.02 | |||||
PFEC+KESI(our) | 2.71 | 45.16 | MWP+KESI(our) | 1.02 |
Table 3: Performance of compact models on Tiny-Imagenet Model Method #Params (M) MACs(G) Acc Resnet-18 baseline 11.01 1.82 67.22 FPEC Li et al. (2016) 2.71 0.83 FPEC+One-cycle 2.71 0.83 FPEC+KESI (our) 2.71 0.83 Resnet-34 baseline 21.39 3.68 68.81 FPEC Li et al. (2016) 5.40 1.57 FPEC+One-cycle 5.40 1.57 FPEC+KESI (our) 5.40 1.57 Table 4: Knowledge distillation with ensembles teacher and single model teacher Model Method #Params (M) C10 C100 Resnet-56 baseline 0.85 93.42 71.07 single teacher 0.28 ensemble teacher 0.28 Resnet-110 baseline 1.73 94.01 72.35 single teacher 0.39 ensemble teacher 0.39 WRN-16-8 baseline 19.96 95.62 79.57 single teacher 2.48 ensemble teacher 2.48
In this section, we demonstrate that the smaller models trained with our pipeline (KESI) achieve comparable or even better results than the original model. Each final model is iteratively pruned and retrained in 5 cycles with different strategies. Table 2 and 4 present the performance of compact models on CIFAR-10, CIFAR-100 and Tiny-Imagenet. Specifically, we compare the iteratively-pruned-models retrained with small learning rate, large learning rate and our pipeline (i.e. large learning rate + knowledge distillation). Our pipeline consistently outperforms the standard strategy by a large margin for both structured and unstructured pruning.
Although our approach is general and can be applied to any (iterative) pruning mechanism, we also give a comparison of model trained with our pipeline and conventional approaches in table 1. We conduct experiment to compare performance of student networks trained with single teacher (i.e. original/unpruned networks) and ensembles teacher in table 4 for ablation study. We can see that compact models learn from ensembles outperform those learn from a single teacher by a large margin.
7 Conclusion
We propose a simple pipeline by slightly modifying the standard approach to acquire the advantages of network ensembles, knowledge distillation and network pruning. Our experiments show that small and compact networks trained with our pipeline significantly outperform the standard approach and create very strong baselines for model compression. Specifically, our method reduces nearly of parameters and FLOPs of several models by structured pruning without incurring loss in performance.
Acknowledgement
The authors thank anonymous reviewers and area chairs for their useful feedback. We also want to express our appreciation to Ms. Le Thi Tham Quynh for her valuable aids in the final preparation of the paper.
References
- Anil et al. (2018) Rohan Anil, Gabriel Pereyra, Alexandre Passos, Robert Ormandi, George E Dahl, and Geoffrey E Hinton. Large scale distributed neural network training through online distillation. arXiv preprint arXiv:1804.03235, 2018.
- Ashukha et al. (2020) Arsenii Ashukha, Alexander Lyzhov, Dmitry Molchanov, and Dmitry Vetrov. Pitfalls of in-domain uncertainty estimation and ensembling in deep learning. arXiv preprint arXiv:2002.06470, 2020.
- Ba and Caruana (2014) Jimmy Ba and Rich Caruana. Do deep nets really need to be deep? In Advances in neural information processing systems, pages 2654–2662, 2014.
- Bagherinezhad et al. (2018) Hessam Bagherinezhad, Maxwell Horton, Mohammad Rastegari, and Ali Farhadi. Label refinery: Improving imagenet classification through label progression. arXiv preprint arXiv:1805.02641, 2018.
- Balan et al. (2015) Anoop Korattikara Balan, Vivek Rathod, Kevin P Murphy, and Max Welling. Bayesian dark knowledge. In Advances in Neural Information Processing Systems, pages 3438–3446, 2015.
- (6) Leo Breiman and Nong Shang. Born again trees.
- Buciluǎ et al. (2006) Cristian Buciluǎ, Rich Caruana, and Alexandru Niculescu-Mizil. Model compression. In Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 535–541, 2006.
- Cho and Hariharan (2019) Jang Hyun Cho and Bharath Hariharan. On the efficacy of knowledge distillation. In Proceedings of the IEEE International Conference on Computer Vision, pages 4794–4802, 2019.
- Deng et al. (2009) Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
- Frankle and Carbin (2018) Jonathan Frankle and Michael Carbin. The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635, 2018.
- Furlanello et al. (2018) Tommaso Furlanello, Zachary C Lipton, Michael Tschannen, Laurent Itti, and Anima Anandkumar. Born again neural networks. arXiv preprint arXiv:1805.04770, 2018.
- Garipov et al. (2018) Timur Garipov, Pavel Izmailov, Dmitrii Podoprikhin, Dmitry P Vetrov, and Andrew G Wilson. Loss surfaces, mode connectivity, and fast ensembling of dnns. In Advances in Neural Information Processing Systems, pages 8789–8798, 2018.
- Guo et al. (2016) Yiwen Guo, Anbang Yao, and Yurong Chen. Dynamic network surgery for efficient dnns. In Advances in neural information processing systems, pages 1379–1387, 2016.
- Han et al. (2015) Song Han, Jeff Pool, John Tran, and William Dally. Learning both weights and connections for efficient neural network. In Advances in neural information processing systems, pages 1135–1143, 2015.
- Hanson and Pratt (1989) Stephen José Hanson and Lorien Y Pratt. Comparing biases for minimal network construction with back-propagation. In Advances in neural information processing systems, pages 177–185, 1989.
- He et al. (2016a) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016a.
- He et al. (2016b) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Identity mappings in deep residual networks. In European conference on computer vision, pages 630–645. Springer, 2016b.
- He et al. (2017) Yihui He, Xiangyu Zhang, and Jian Sun. Channel pruning for accelerating very deep neural networks. In Proceedings of the IEEE International Conference on Computer Vision, pages 1389–1397, 2017.
- Hinton et al. (2015) Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.
- Huang et al. (2017) Gao Huang, Yixuan Li, Geoff Pleiss, Zhuang Liu, John E Hopcroft, and Kilian Q Weinberger. Snapshot ensembles: Train 1, get m for free. arXiv preprint arXiv:1704.00109, 2017.
- Kingma and Ba (2014) Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Krizhevsky et al. (2009) Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009.
- Lakshminarayanan et al. (2017) Balaji Lakshminarayanan, Alexander Pritzel, and Charles Blundell. Simple and scalable predictive uncertainty estimation using deep ensembles. In Advances in neural information processing systems, pages 6402–6413, 2017.
- LeCun et al. (1990) Yann LeCun, John S Denker, and Sara A Solla. Optimal brain damage. In Advances in neural information processing systems, pages 598–605, 1990.
- Lee and Chung (2019) Jisoo Lee and Sae-Young Chung. Robust training with ensemble consensus. arXiv preprint arXiv:1910.09792, 2019.
- Li et al. (2016) Hao Li, Asim Kadav, Igor Durdanovic, Hanan Samet, and Hans Peter Graf. Pruning filters for efficient convnets. arXiv preprint arXiv:1608.08710, 2016.
- Lin et al. (2020) Mingbao Lin, Rongrong Ji, Yan Wang, Yichen Zhang, Baochang Zhang, Yonghong Tian, and Ling Shao. Hrank: Filter pruning using high-rank feature map. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1529–1538, 2020.
- Lin et al. (2019) Shaohui Lin, Rongrong Ji, Chenqian Yan, Baochang Zhang, Liujuan Cao, Qixiang Ye, Feiyue Huang, and David Doermann. Towards optimal structured cnn pruning via generative adversarial learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2790–2799, 2019.
- Liu et al. (2018) Zhuang Liu, Mingjie Sun, Tinghui Zhou, Gao Huang, and Trevor Darrell. Rethinking the value of network pruning. arXiv preprint arXiv:1810.05270, 2018.
- Loshchilov and Hutter (2016) Ilya Loshchilov and Frank Hutter. Sgdr: stochastic gradient descent with restarts. corr abs/1608.03983 (2016). arXiv preprint arXiv:1608.03983, 2016.
- Luo et al. (2017) Jian-Hao Luo, Jianxin Wu, and Weiyao Lin. Thinet: A filter level pruning method for deep neural network compression. In Proceedings of the IEEE international conference on computer vision, pages 5058–5066, 2017.
- Malinin et al. (2019) Andrey Malinin, Bruno Mlodozeniec, and Mark Gales. Ensemble distribution distillation. arXiv preprint arXiv:1905.00076, 2019.
- (33) Seyed-Iman Mirzadeh, Mehrdad Farajtabar, Ang Li, Nir Levine, Akihiro Matsukawa, and Hassan Ghasemzadeh. Improved knowledge distillation via teacher assistant.
- Mirzadeh et al. (2019) Seyed-Iman Mirzadeh, Mehrdad Farajtabar, Ang Li, and Hassan Ghasemzadeh. Improved knowledge distillation via teacher assistant: Bridging the gap between student and teacher. arXiv preprint arXiv:1902.03393, 2019.
- Molchanov et al. (2016) Pavlo Molchanov, Stephen Tyree, Tero Karras, Timo Aila, and Jan Kautz. Pruning convolutional neural networks for resource efficient inference. arXiv preprint arXiv:1611.06440, 2016.
- Reed (1993) Russell Reed. Pruning algorithms-a survey. IEEE transactions on Neural Networks, 4(5):740–747, 1993.
- Renda et al. (2020) Alex Renda, Jonathan Frankle, and Michael Carbin. Comparing rewinding and fine-tuning in neural network pruning. arXiv preprint arXiv:2003.02389, 2020.
- Romero et al. (2014) Adriana Romero, Nicolas Ballas, Samira Ebrahimi Kahou, Antoine Chassang, Carlo Gatta, and Yoshua Bengio. Fitnets: Hints for thin deep nets. arXiv preprint arXiv:1412.6550, 2014.
- Smith (2015) Leslie N Smith. No more pesky learning rate guessing games. CoRR, abs/1506.01186, 5, 2015.
- Smith and Topin (2019) Leslie N Smith and Nicholay Topin. Super-convergence: Very fast training of neural networks using large learning rates. In Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications, volume 11006, page 1100612. International Society for Optics and Photonics, 2019.
- Snoek et al. (2019) Jasper Snoek, Yaniv Ovadia, Emily Fertig, Balaji Lakshminarayanan, Sebastian Nowozin, D Sculley, Joshua Dillon, Jie Ren, and Zachary Nado. Can you trust your model’s uncertainty? evaluating predictive uncertainty under dataset shift. In Advances in Neural Information Processing Systems, pages 13969–13980, 2019.
- Srinivas and Babu (2015) Suraj Srinivas and R Venkatesh Babu. Data-free parameter pruning for deep neural networks. arXiv preprint arXiv:1507.06149, 2015.
- Yang et al. (2019a) Chenglin Yang, Lingxi Xie, Siyuan Qiao, and Alan L Yuille. Training deep neural networks in generations: A more tolerant teacher educates better students. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 5628–5635, 2019a.
- Yang et al. (2019b) Chenglin Yang, Lingxi Xie, Chi Su, and Alan L Yuille. Snapshot distillation: Teacher-student optimization in one generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2859–2868, 2019b.
- You et al. (2019) Zhonghui You, Kun Yan, Jinmian Ye, Meng Ma, and Ping Wang. Gate decorator: Global filter pruning method for accelerating deep convolutional neural networks. In Advances in Neural Information Processing Systems, pages 2133–2144, 2019.
- Yu and Huang (2019) Jiahui Yu and Thomas S Huang. Universally slimmable networks and improved training techniques. In Proceedings of the IEEE International Conference on Computer Vision, pages 1803–1811, 2019.
- Yu et al. (2018) Ruichi Yu, Ang Li, Chun-Fu Chen, Jui-Hsin Lai, Vlad I Morariu, Xintong Han, Mingfei Gao, Ching-Yung Lin, and Larry S Davis. Nisp: Pruning networks using neuron importance score propagation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 9194–9203, 2018.
- Zagoruyko and Komodakis (2016a) Sergey Zagoruyko and Nikos Komodakis. Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. arXiv preprint arXiv:1612.03928, 2016a.
- Zagoruyko and Komodakis (2016b) Sergey Zagoruyko and Nikos Komodakis. Wide residual networks. arXiv preprint arXiv:1605.07146, 2016b.
- Zhang et al. (2018) Ying Zhang, Tao Xiang, Timothy M Hospedales, and Huchuan Lu. Deep mutual learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4320–4328, 2018.
- Zhu et al. (2018) Xiatian Zhu, Shaogang Gong, et al. Knowledge distillation by on-the-fly native ensemble. In Advances in neural information processing systems, pages 7517–7527, 2018.
Appendix A Results
Model | #Param | % MACs(G) | C10-SLR | C10-LLR | C100-SLR | C100-LLR |
---|---|---|---|---|---|---|
Resnet-110 (baseline) | 1.73M | 0.00 | 94.01 | 94.01 | 72.35 | 72.35 |
Resnet-110 #1 | 1.26M | 23.08 | ||||
Resnet-110 #2 | 0.93M | 38.46 | ||||
Resnet-110 #3 | 0.70M | 50.00 | ||||
Resnet-110 #4 | 0.52M | 57.69 | ||||
Resnet-110 #5 | 0.39M | 65.38 | ||||
Resnet-110 Ensemble | - | - | 93.82 | 73.32 | ||
Resnet-56 (baseline) | 0.85M | 0.00 | 93.42 | 93.42 | 71.07 | 71.07 |
Resnet-56 #1 | 0.66M | 23.07 | ||||
Resnet-56 #2 | 0.52M | 30.77 | ||||
Resnet-56 #3 | 0.42M | 46.15 | ||||
Resnet-56 #4 | 0.35M | 53.85 | ||||
Resnet-56 #5 | 0.29M | 61.54 | ||||
Resnet-56 Ensemble | - | - | 93.25 | 71.37 | ||
VGG-16 (baseline) | 14.99M | 0.00 | 94.23 | 94.23 | 73.24 | 73.24 |
VGG-16 #1 | 9.46M | 0.00 | ||||
VGG-16 #2 | 6.27M | 0.00 | ||||
VGG-16 #3 | 4.43M | 0.00 | ||||
VGG-16 #4 | 3.36M | 0.00 | ||||
VGG-16 #5 | 2.71M | 0.00 | ||||
VGG-16 Ensemble | - | - | 94.29 | |||
PreResnet-164 (baseline) | 1.7M | 0.00 | 95.06 | 95.06 | 76.35 | 76.35 |
PreResnet-164 #1 | 1.09M | 26.92 | ||||
PreResnet-164 #2 | 0.74M | 46.15 | ||||
PreResnet-164 #3 | 0.54M | 57.69 | ||||
PreResnet-164 #4 | 0.4M | 65.38 | ||||
PreResnet-164 #5 | 0.31M | 69.23 | ||||
PreResnet-164 Ensemble | - | - | - | - | ||
WideResnet-16-8 (baseline) | 11.01 | 0.00 | 95.62 | 95.62 | 79.57 | 79.57 |
WideResnet-16-8 #1 | 8.01 | 20.00 | ||||
WideResnet-16-8 #2 | 5.89 | 35.48 | ||||
WideResnet-16-8 #3 | 4.38 | 47.74 | ||||
WideResnet-16-8 #4 | 3.28 | 57.42 | ||||
WideResnet-16-8 #5 | 2.48 | 64.52 | ||||
WideResnet-16-8 Ensemble | - | - | 95.63 |
Model | #Active Params (M) | C10-SLR | C10-LLR | C100-SLR | C100-LLR |
---|---|---|---|---|---|
Resnet-56 (baseline) | 0 | 93.42 | 93.42 | 71.07 | 71.07 |
Resnet-56 #1 | 0.66 | ||||
Resnet-56 #2 | 0.52 | ||||
Resnet-56 #3 | 0.42 | ||||
Resnet-56 #4 | 0.35 | ||||
Resnet-56 #5 | 0.29 | ||||
Resnet-56 Ensemble | - | 93.39 | |||
Resnet-110 (baseline) | 0 | 94.01 | 94.01 | 72.35 | 72.35 |
Resnet-110 #1 | 1.27 | ||||
Resnet-110 #2 | 0.94 | ||||
Resnet-110 #3 | 0.69 | ||||
Resnet-110 #4 | 0.50 | ||||
Resnet-110 #5 | 0.36 | ||||
Resnet-110 Ensemble | - | ||||
WideResnet-110 (baseline) | 11.01 | 95.62 | 95.62 | 79.57 | 79.57 |
WideResnet-110 #1 | 8.05 | ||||
WideResnet-110 #2 | 5.97 | ||||
WideResnet-110 #3 | 4.44 | ||||
WideResnet-110 #4 | 3.34 | ||||
WideResnet-110 #5 | 2.53 | ||||
WideResnet-110 Ensemble | - |
Model | #Param | SLR | LLR |
---|---|---|---|
Resnet-18 (baseline) | 11.01 | 67.22 | 67.22 |
Resnet-18 #1 | 8.30 | ||
Resnet-18 #2 | 6.17 | ||
Resnet-18 #3 | 4.64 | ||
Resnet-18 #4 | 3.52 | ||
Resnet-18 #5 | 2.71 | ||
Resnet-18 Ensemble | - | ||
Resnet-34 (baseline) | 21.39 | 68.81 | 68.81 |
Resnet-34 #1 | 15.97 | ||
Resnet-34 #2 | 12.02 | ||
Resnet-34 #3 | 9.13 | ||
Resnet-34 #4 | 6.99 | ||
Resnet-34 #5 | 5.40 | ||
Resnet-34 Ensemble | - |