Learning to Teach with Student Feedback
Abstract
Knowledge distillation (KD) has gained much attention due to its effectiveness in compressing large-scale pre-trained models. In typical KD methods, the small student model is trained to match the soft targets generated by the big teacher model. However, the interaction between student and teacher is one-way. The teacher is usually fixed once trained, resulting in static soft targets to be distilled. This one-way interaction leads to the teacher’s inability to perceive the characteristics of the student and its training progress. To address this issue, we propose Interactive Knowledge Distillation (IKD), which also allows the teacher to learn to teach from the feedback of the student. In particular, IKD trains the teacher model to generate specific soft target at each training step for a certain student. Joint optimization for both teacher and student is achieved by two iterative steps: a course step to optimize student with the soft target of teacher, and an exam step to optimize teacher with the feedback of student. IKD is a general framework that is orthogonal to most existing knowledge distillation methods. Experimental results show that IKD outperforms traditional KD methods on various NLP tasks.
Introduction
Large-scale pre-trained language models (PLMs) such as GPT (Radford et al. 2019), BERT (Devlin et al. 2019), RoBERTa (Liu et al. 2019) have achieved significant improvement on various NLP tasks. Despite their power, they are computationally expensive due to their enormous size, which limits their deployment in real-time scenarios.
As an effective technique to tackle this problem, Knowledge Distillation (KD) (Bucila, Caruana, and Niculescu-Mizil 2006; Hinton, Vinyals, and Dean 2015) has gained much attention in the community. In common KD methods for compressing large-scale PLMs, the original large-scale PLM often serves as the teacher model, which is first trained on the downstream task then uses its generated soft targets to teach the student model, which is usually a smaller PLM. Previous works such as DistilBERT (Sanh et al. 2019), BERT-PKD (Sun et al. 2019), MobileBERT (Sun et al. 2020), and TinyBERT (Jiao et al. 2020) have explored informative features that can be distilled, including the output logits, word embeddings, hidden states, attention maps, etc.

However, in these existing methods the interaction between teacher and student remains one-way. The teacher model is usually fixed once trained on the downstream data, resulting in static soft targets regardless of the characteristics of the student and its training progress. Though there are works that insert a scheduled temperature into the Softmax function to control the smoothness of the soft targets such as Annealing-KD (Jafari et al. 2021), the design of the temperature and its schedule does not utilize the feedback from the student and highly depends on expert experience.
In this paper, we propose the Interactive Knowledge Distillation (IKD) to implement a bidirectional interaction between teacher and student. IKD allows the teacher to learn to teach based on the feedback from the student. In particular, the teacher model is trained to generate specific soft targets at each training step with the help of meta-learning, especially MAML (Finn, Abbeel, and Levine 2017). Figure 1 illustrates main difference between our proposed IKD and traditional KD methods.
The central idea of IKD is to make the student model generalize well (like achieving a lower loss) on a batch of unseen samples after learned from the teacher model, while the measurable performance on unseen samples is back-propagated to the teacher model in order to update its teaching strategy. More specifically, IKD consists of two update steps: course step and exam step. In course step, the student is trained to match the soft targets generated by the teacher on a batch of data called course data. In exam step, the student after one gradient step is evaluated on another batch of data called exam data, the cross entropy of the student on the exam data provides meta-gradients to be back-propagated to the teacher model such that the teacher model could get updated to generate better soft targets. By iteratively conducting the two steps, we can continually improve the student model via the soft targets generated by the teacher, and improve the teacher model via the feedback generated by the student.
We conduct experiments on various NLP tasks. Experimental results on GLUE benchmark demonstrate that IKD consistently outperforms vanilla knowledge distillation. Further analysis is then conducted to shed some light on the dynamic soft targets and the student feedback.
To sum up, our contributions are as follows:
-
•
We propose a co-interactive method for teacher-student framework, namely Interactive Knowledge Distillation (IKD), which allows the teacher model to learn to generate specific soft targets at each training step.
-
•
We take an approximation to convert the iterative optimization into a joint optimization, which is more efficient for training.
-
•
Our proposed IKD is orthogonal to most existing knowledge distillation works that distill different feature sets, in which we empirically evaluated the effectiveness of IKD based on vanilla knowledge distillation.
Method

The intuition behind our framework is straight-forward: First, the teacher teaches the course to the student who then updates its knowledge according to the course. Second, the student takes exams and produces scores for the teacher to adjust its teaching strategy. Such two steps are common in real-world education.
Based on the intuition above, IKD can be formulated in the context of machine learning as follows. Denote the teacher model and the student model as and respectively. Assume teacher model is parameterized by , student model is parameterized by . In the course step, we optimize the student model by vanilla knowledge distillation. The student model is trained to match the soft targets generated by the teacher model on a batch of data drawn from the course data set . In the exam step, we evaluate the student model on a batch of data drawn from the exam data set . The test score of the exam, which can be instantiated as cross entropy, provides meta gradients to optimize the teacher model . The overall illustration of our method is depicted by Figure 2.
Course Step: Student Optimization by Vanilla Knowledge Distillation
In our setting, the teacher model is implemented as a deep encoder such as BERT (Devlin et al. 2019), and the student model is implemented as a lightweight encoder such as a Transformer (Vaswani et al. 2017) encoder with fewer layers.
Given a training sample drawn from the course data set , the teacher model computes a contextualized embedding 111We use the [CLS] token embedding of the last layer as the sentence representation., which is then fed into a Softmax layer to obtain the probability for each category, i.e. the soft targets, that is , where is a learnable parameter matrix and is the logits. For simplicity, is abbreviated to . In practice, the temperature is often introduced as a hyper-parameter to control the smoothness of the soft targets,
(1) |
where denotes the number of classes, is the correct class.
Similarly, we can obtain the output of the student model , where . In vanilla knowledge distillation, the KL divergence of the teacher’s prediction and the student’s prediction should be minimized,
(2) |
where means Kullback-Leibler divergence.
In practice, the cross entropy between the student’s prediction and the ground truth is also incorporated to be minimized, i.e.,
(3) |
Thus the overall loss function for the student model can be written as:
(4) |
where is a hyper-parameter to balance the knowledge distillation loss and the cross entropy loss.
The student model is then updated by taking one gradient descent step with the loss function above, i.e.
(5) |
where is the learning rate of the student model.
Exam Step: Teacher Optimization by Student’s Feedback
It is expected that the student with the post-update parameters will generalize well on unseen samples. Therefore, the updated student is then evaluated on another batch of samples drawn from the exam data set . The performance of the student is measured by the cross entropy loss on the batch of exam data, which is called meta loss since it is calculated on the post-update parameter thus provides meta gradient w.r.t. the teacher’s parameters.
(6) | ||||
(7) |
Our goal is to minimize by optimizing the teacher model, therefore the gradient of is w.r.t. the teacher’s parameters , so is a second-order gradient (gradient of gradient). To explicitly show this, we can re-write in the above equation as:
(8) |
in which we know that the gradient can flow into through the knowledge distillation in the course step. Figure 3 shows the computational graph of the forward pass and the backward pass.

Although it is computationally available by conducting an additional backward pass under current deep learning libraries that support automatic differentiation, we use the first-order approximation for its efficiency (Finn, Abbeel, and Levine 2017; Nichol, Achiam, and Schulman 2018).
In the context of meta-learning, the teacher can be regarded as the meta-learner, while and can be analogous to support set and query set respectively.
Also, the standard cross entropy loss between the teacher’s prediction and the ground truth can be incorporated,
(9) |
Thus, the overall loss function for the teacher model is
(10) |
where is another hyper-parameter to balance the meta loss and the cross entropy loss.
Thus, the teacher model can also be optimized by taking gradient descent:
(11) |
where is the learning rate of the teacher model.
Putting the student optimization in course step and the teacher optimization in exam step together, the overall loss function of IKD is defined as
(12) |
By iterating the course step and the exam step, the student model learns from the soft targets generated by the teacher while the teacher model learns from the feedback generated by the student.
Discussion and Implementation
To take a closer look, in this section we further analyze the expanded form of the meta-gradient, which provides insights into the student’s feedback and sheds light on how to take an approximation to achieve efficient training.
According to the chain rule, we can unfold the meta gradient as follows222For the simplicity of derivation, we set and to 1, and omit the constant scalar .,
(13) | ||||
Note that is the Jacobi matrix of shape and is of shape .333For matrix multiplication, should be in the transpose form, which is omitted here for the simplicity. Therefore, is actually a weighting vector of shape to adjust the gradient of teacher’s prediction w.r.t. its parameters . It is easy to find that the weighting vector only depends on the student’s gradient, therefore it is exactly the student’s feedback.
The behavior relationship between the weighting vector and the teacher’s prediction is analyzed. It is worth noticing that the weighting vector should be detached from the computational graph to forbid the gradient flow into the student. By this, the meta loss will only contribute to the update of the teacher’s parameters.
In practice, the difference between and is usually very small, which motivates us to take an approximation such that . By this approximation, Eq. (13) can be simplified as
(14) |
Besides, note that , and are mutually independent, therefore we can calculate them independently and achieve joint optimizing in an efficient way. In our experiments we set for efficient optimization. Though, it should be mentioned that in our framework the course data set can be unlabeled, i.e. . In this case, the cross entropy loss between the student’s prediction and the ground truth is removed such that the loss for the student is exactly the knowledge distillation loss. Thus, only the supervision signal of is required. We leave IKD in this semi-supervised setting as future work.
Experiments
Though IKD is orthogonal to most current knowledge distillation methods to compressing large-scale PLMs such as DistilBERT (Sanh et al. 2019) and TinyBERT (Jiao et al. 2020), which distill different feature sets (such as logits, embeddings, hidden states, attention maps, etc.), it is exhaustive to test IKD with these different distillation features and their combinations. Thus, in our experiments we mainly evaluate the effectiveness of IKD based on the vanilla knowledge distillation and BERT-PKD (Sun et al. 2019).
Datasets and Models
We conduct experiments on the General Language Understanding Evaluation (GLUE) benchmark (Wang et al. 2019) with the backbone of BERT (Devlin et al. 2019), and on ChemProt dataset (Schneider et al. 2020) and SciCite dataset (Cohan et al. 2019) with backbone of SciBERT (Beltagy, Lo, and Cohan 2019).
GLUE Benchmark
We use seven text classification tasks in the GLUE benchemark: The Corpus of Linguistic Acceptability (CoLA) (Warstadt, Singh, and Bowman 2019), The Stanford Sentiment Treebank (SST-2) (Socher et al. 2013), Microsoft Research Paraphrase Corpus (MRPC) (Dolan and Brockett 2005), Quora Question Pairs (QQP) (Wang et al. 2019), Multi-Genre Natural Language Inference (MNLI) (Williams, Nangia, and Bowman 2018), Question Natural Language Inference (QNLI) (Rajpurkar et al. 2016) and Recognizing Textual Entailment (RTE) (Wang et al. 2019).
Domain Tasks
We use ChemProt (Schneider et al. 2020) and SciCite (Cohan et al. 2019) to evaluate the performance of our method on domain specific tasks. ChemProt consists of 1,820 PubMed abstracts with chemical-protein interactions annotated by domain experts and was used in the BioCreative VI text mining chemical-protein interactions shared task. SciCite is a dataset containing 11k annotated citation intents in biochemical and computer science domains. The standard metrics are micro F1 for ChemProt and macro F1 for SciCite. We evaluate our method on both tasks with the backbone of SciBERT (Beltagy, Lo, and Cohan 2019).
Experimental Setup
Our implementation is based on PyTorch (Paszke et al. 2019). The training and evaluation are performed on a single RTX 2080Ti or RTX 3090 GPU. For downstream tasks in GLUE we first fine-tune the teacher model for 3 epochs with learning rate of 2e-5. For domain tasks the teacher is fine-tuned for 4 epochs. The batch size is 32 for all tasks. The student model is initialized with the parameters from the first layers of BERT base(12 layer), where is the number of the student’s layer. In our experiments we mainly evaluate our method when and . The total number of parameters of BERT24, BERT12, BERT6 and Bert4 are 340M, 110M, 67M and 52M respectively. The batch size is set to 4 for all downstream tasks. We conduct 3 experiments for each dataset and the median is reported.
Method | MNLI | QQP | QNLI | SST-2 | CoLA | MRPC | RTE | Macro |
---|---|---|---|---|---|---|---|---|
(393k) | (364k) | (105k) | (67k) | (8.5k) | (3.7k) | (2.5k) | Score | |
Dev Set | ||||||||
BERT24 (Teacher) | 87.0 | 88.7 | 92.6 | 93.2 | 63.4 | 88.8 | 70.0 | 83.4 |
BERT4-TAKD† | 72.4 | 87 | 82.6 | 89.1 | 34.2 | 85.2 | 59.6 | 72.9 |
BERT4-Annealing KD† | 74.4 | 86.5 | 83.1 | 89.4 | 36.0 | 86.2 | 61.0 | 73.8 |
BERT4-FT | 78.2 | 85.9 | 85.4 | 88.0 | 28.6 | 85.1 | 63.2 | 73.5 |
BERT4-KD | 78.5 | 86.4 | 85.2 | 88.8 | 30.4 | 84.6 | 63.5 | 73.9 |
BERT4-IKD | 79.1 | 86.5 | 85.5 | 89.1 | 30.9 | 86.0 | 66.4 | 74.7 |
Test Set | ||||||||
BERT4-KD | 78.2 | 68.8 | 84.6 | 90.1 | 27.7 | 82.5 | 63.5 | 70.8 |
BERT4-IKD | 78.4 | 68.9 | 85.3 | 90.1 | 30.6 | 83.2 | 62.9 | 71.3 |
Experimental Results
We report our results on GLUE benchmark. The results are presented in Table 1. We denote BERTk as a BERT with only layers. BERT4-FT is to directly fine-tune the first 4 layers of the pre-trained BERT. BERT4-KD represents student trained with vanilla knowledge distillation loss, i.e., Eq. (4). We also exhibit the reported result of TAKD (Mirzadeh et al. 2019) and Annealing KD (Jafari et al. 2021) for a direct comparision.
Results on GLUE Benchmark
As shown in Table 1, IKD surpasses the vanilla knowledge distillation method on all tasks. For RTE, IKD outperforms directly finetuning by 3.2% and knowledge distillation by 2.9%. For QNLI and MRPC, although the performance of knowledge distillation compared with directly fine-tuning is even detrimental, IKD can consistently improve the students. We also evaluate our method on the GLUE test server for comparision on the test set and we show the results in Table 1. On the GLUE leaderboard our method consistently improves the performance.
Results on Domain Tasks
Experimental results on the ChemProt and SciCite test set are shown in Table 2, which demonstrate the effectiveness of IKD on domain specific task. It’s worth noting that on the SciCite dataset our approach is even able to make the student outperform the teacher.
Method | ChemProt | SciCite |
---|---|---|
(4.2k) | (7.3k) | |
SciBERT12(Teacher) | 84.5 | 86.1 |
SciBERT6-FT | 78.1 | 85.0 |
SciBERT6-KD | 79.3 | 85.7 |
SciBERT6-IKD | 79.9 | 86.6 |
Compatibility with Patient Knowledge Distillation
To evaluate the compatibility of IKD with more competitive distillation methods, we further apply IKD on BERT-PKD (Sun et al. 2019) (denoted as BERT-PKD-IKD) with teacher model as BERT-base (12 layers) and student as BERT6. BERT-PKD exploits information from hidden features for distillation. We follow PKD-Skip strategy which is shown to be better than PKD-Last as suggested by Sun et al. (2019). Experimental results are listed in Table 3, which demonstrate the versatility of IKD with stronger baselines.
Method | QNLI | SST-2 | MRPC | RTE |
---|---|---|---|---|
BERT12(Teacher) | 90.4 | 92.4 | 88.9 | 68.2 |
BERT6-PKD | 88.3 | 89.1 | 87.6 | 66.8 |
BERT6-PKD-IKD | 88.7 | 90.3 | 87.7 | 67.1 |
Analysis
In this section, we take a closer look at IKD by conducting empirical analysis to provide some insights to understand how it works. In particular, we focus on the dynamic of generating soft targets and the feedback from the student.
About the Soft Targets

Dataset | Epoch 1 | Epoch 2 | Epoch 3 |
---|---|---|---|
MRPC | 0.2637 | 0.1076 | 0.0343 |
RTE | 0.4461 | 0.2313 | 0.0845 |
ChemProt | 0.5049 | 0.2686 | 0.1451 |
Since IKD allows the teacher to dynamically adjust its soft targets by using the feedback from the student, it would be interesting to see how the output of the teacher changes as the training goes on. In Figure 4 we visualize the output soft targets of the teacher output in different training epochs. We can find that the soft targets are relatively smooth at the beginning and become sharper as the training progresses. This is consistent with the hand-designed smoothing schedules (Dogan et al. 2020; Jafari et al. 2021), which introduce a hyper-parameter to control how quickly the soft targets converge to the one-hot distribution. In contrast, our method automatically learns to adjust the smoothness of the soft targets based on the feedback from the student. In Table 4 we further use entropy to measure the smoothness of output targets in different epochs on full training sets of MRPC, RTE and ChemProt. The average entropy goes down over the training process as shown, which confirms our observation as well.
About the Feedback
As mentioned in Implementation Details, the gradient of the meta loss can be viewed as a weighted sum of the gradient of teacher’s prediction w.r.t. its parameters, i.e. (See Eq. (14)). The weighting vector can be regarded as the feedback of the student. Assume , Eq. (14) can be written as
(15) |
where , , is the number of classes.
In this section, we attempt to analyze the relationship between the student’s feedback and the teacher’s prediction . In particular, we plot for each in Figure 5. The data points are collected from the 7,500 training steps on CoLA dataset. CoLA is a binary classification task so we can plot data points for each training step such that there are 15,000 data points in total.


Note that The greater the student’s feedback for class , the greater the penalty of teacher’s prediction on class . As shown in Figure 5(a), the student’s feedback is always non-positive for teacher’s prediction on target label, which can also be revealed by the expanded form of the feedback:
(16) |
where is the target label. Besides, we can see that the student’s feedback is usually non-negative for teacher’s prediction on non-target label. Thus, the teacher’s prediction is always encouraged to approach the one-hot distribution where only the target label is 1, which has been verified in analysis. Also, we find that the range of feedback on target prediction is approximately three times larger than the range of feedback on non-target prediction.
In addition, the student’s feedback can be different even when the teacher’s predictions are the same. To take a deeper look, we demonstrate the value of feedback of the student at different training stage in Figure 5(b). We can see that the student feedback converges to zero as the training goes, which implies that IKD works mainly at the immediate training stage.
Related Work
Our method draws on the idea of meta-learning and applies it to knowledge distillation for language understanding. Thus, there are two lines of work that are related: (a) Knowledge distillation methods for language understanding, (b) Optimization-based meta-learning. We will introduce the two lines of work in Transformer Based KD and Optimization-Based Meta-Learning, and the intersect of the both lines in Teacher-Student Framework with Meta-Learning.
Knowledge Distillation for NLP
Knowledge distillation (KD) was originally proposed in domains other than NLP (Bucila, Caruana, and Niculescu-Mizil 2006; Hinton, Vinyals, and Dean 2015). With the emergence of large-scale pre-trained language models (PLMs) such as GPT (Radford et al. 2019), BERT (Devlin et al. 2019), and RoBERTa (Liu et al. 2019), KD has gain much attention in the NLP community due to its simplicity and efficiency in compressing the large PLMs. Much effort has been devoted to compress BERT via KD, including BERT-PKD (Sun et al. 2019), DistilBERT (Sanh et al. 2019), MobileBERT (Sun et al. 2020), and TinyBERT (Jiao et al. 2020). These methods have achieved great success by exploring informative features to be distilled, such as word embeddings, attention maps, hidden states, and output logits. Different from these work, our proposed IKD focuses on the interaction between teacher and student therefore is orthogonal to these work. Gou et al. (2021) provides a survey of online knowledge distillation where both the teacher and the student are updated simulatenously. Our IKD can be regarded as a special case of online KD.
Optimization-Based Meta-Learning
Optimization-based meta-learning intends to learn a group of initial parameters that can fast converge on a new task with only a few gradient steps, in which the learner and meta-learner share the same architecture. As a representative method, Model-Agnostic Meta-Learning (MAML) (Finn, Abbeel, and Levine 2017) uses a nested loop to find an optimal set of parameters on a variety of learning tasks, such that they can adapt well after a small number of training steps. To tackle the high-order derivative, Finn, Abbeel, and Levine (2017) also propose First-Order MAML (FOMAML) to perform the first-order approximation to reduce the computation. Similar to MAML, Reptile (Nichol, Achiam, and Schulman 2018) utilizes the weights in the inner loop instead of gradient to update the meta parameters and achieves competitive performance on several few-shot classification benchmarks.
Teacher-Student Framework with Meta-Learning
Meta-learning has been proved efficient when combined with the teacher-student framework. Liu et al. (2020) employ meta-learning to optimize a label generator that generates dynamic soft targets for intermediate layers. Their method is proposed for self-boosting. Therefore, the teacher and the student are required to share the same architecture and parameters. Pan et al. (2020) focus on training a meta-teacher possessing multi-domain knowledge by metric-based meta-learning and then teach single-domain students. Meta Pseudo Labels (Pham et al. 2020) leverage a teacher network to generate pseudo labels on unsupervised images to enlarge the training set of the student. They aim to train a pseudo labels generator to while IKD dedicates to compress a large model into a smaller one. Thus the pseudo label generator is trained from scratch and shares same architecture with the student while in our IKD the teacher must be a larger pretrained model which leads to the impossibility of implementation of gradient substitution in MAML. Moreover, Meta Pseudo Labels generates hard labels and uses Monte Carlo approximation to obtain gradient while soft targets from the teacher in KD are necessary for passing dark knowledge to student and gradients in IKD are obtained by adopting the chain rule manually.
Conclusion
This paper proposes the Interactive Knowledge Distillation (IKD) to empower the teacher to learn to teach with student feedback. Inspired by MAML, IKD involves two optimization steps: (1) a course step for student optimization with the guidance of teacher and (2) an exam step for teacher optimization with the feedback from student. IKD shows superiority over vanilla KD on a suit of NLP tasks. By iterating the course step and exam step we can jointly optimize the teacher and student. In addition, we have discussed a semi-supervised generalization of our method, in which the course data is allowed to be unlabeled.
References
- Beltagy, Lo, and Cohan (2019) Beltagy, I.; Lo, K.; and Cohan, A. 2019. SciBERT: A Pretrained Language Model for Scientific Text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 3615–3620. Hong Kong, China: Association for Computational Linguistics.
- Bucila, Caruana, and Niculescu-Mizil (2006) Bucila, C.; Caruana, R.; and Niculescu-Mizil, A. 2006. Model compression. In Proceedings of the Twelfth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Philadelphia, PA, USA, August 20-23, 2006, 535–541. ACM.
- Cohan et al. (2019) Cohan, A.; Ammar, W.; van Zuylen, M.; and Cady, F. 2019. Structural Scaffolds for Citation Intent Classification in Scientific Publications. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 3586–3596. Minneapolis, Minnesota: Association for Computational Linguistics.
- Devlin et al. (2019) Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171–4186. Minneapolis, Minnesota: Association for Computational Linguistics.
- Dogan et al. (2020) Dogan, Ü.; Deshmukh, A. A.; Machura, M.; and Igel, C. 2020. Label-Similarity Curriculum Learning. In Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part XXIX, volume 12374 of Lecture Notes in Computer Science, 174–190. Springer.
- Dolan and Brockett (2005) Dolan, W. B.; and Brockett, C. 2005. Automatically Constructing a Corpus of Sentential Paraphrases. In Proceedings of the Third International Workshop on Paraphrasing, IWP@IJCNLP 2005, Jeju Island, Korea, October 2005, 2005. Asian Federation of Natural Language Processing.
- Finn, Abbeel, and Levine (2017) Finn, C.; Abbeel, P.; and Levine, S. 2017. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. In Precup, D.; and Teh, Y. W., eds., Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, volume 70 of Proceedings of Machine Learning Research, 1126–1135. PMLR.
- Gou et al. (2021) Gou, J.; Yu, B.; Maybank, S. J.; and Tao, D. 2021. Knowledge Distillation: A Survey. International Journal of Computer Vision, 129(6): 1789–1819.
- Hinton, Vinyals, and Dean (2015) Hinton, G. E.; Vinyals, O.; and Dean, J. 2015. Distilling the Knowledge in a Neural Network. CoRR, abs/1503.02531.
- Jafari et al. (2021) Jafari, A.; Rezagholizadeh, M.; Sharma, P.; and Ghodsi, A. 2021. Annealing Knowledge Distillation. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, EACL 2021, Online, April 19 - 23, 2021, 2493–2504. Association for Computational Linguistics.
- Jiao et al. (2020) Jiao, X.; Yin, Y.; Shang, L.; Jiang, X.; Chen, X.; Li, L.; Wang, F.; and Liu, Q. 2020. TinyBERT: Distilling BERT for Natural Language Understanding. In Findings of the Association for Computational Linguistics: EMNLP 2020, 4163–4174. Online: Association for Computational Linguistics.
- Liu et al. (2020) Liu, B.; Rao, Y.; Lu, J.; Zhou, J.; and Hsieh, C. 2020. MetaDistiller: Network Self-Boosting via Meta-Learned Top-Down Distillation. In Vedaldi, A.; Bischof, H.; Brox, T.; and Frahm, J., eds., Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part XIV, volume 12359 of Lecture Notes in Computer Science, 694–709. Springer.
- Liu et al. (2019) Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; and Stoyanov, V. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR, abs/1907.11692.
- Mirzadeh et al. (2019) Mirzadeh, S.-I.; Farajtabar, M.; Li, A.; Levine, N.; Matsukawa, A.; and Ghasemzadeh, H. 2019. Improved Knowledge Distillation via Teacher Assistant. arXiv:1902.03393.
- Nichol, Achiam, and Schulman (2018) Nichol, A.; Achiam, J.; and Schulman, J. 2018. On First-Order Meta-Learning Algorithms. CoRR, abs/1803.02999.
- Pan et al. (2020) Pan, H.; Wang, C.; Qiu, M.; Zhang, Y.; Li, Y.; and Huang, J. 2020. Meta-KD: A Meta Knowledge Distillation Framework for Language Model Compression across Domains. CoRR, abs/2012.01266.
- Paszke et al. (2019) Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; Desmaison, A.; Köpf, A.; Yang, E.; DeVito, Z.; Raison, M.; Tejani, A.; Chilamkurthy, S.; Steiner, B.; Fang, L.; Bai, J.; and Chintala, S. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 8024–8035.
- Pham et al. (2020) Pham, H.; Xie, Q.; Dai, Z.; and Le, Q. V. 2020. Meta Pseudo Labels. CoRR, abs/2003.10580.
- Radford et al. (2019) Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; and Sutskever, I. 2019. Language Models are Unsupervised Multitask Learners.
- Rajpurkar et al. (2016) Rajpurkar, P.; Zhang, J.; Lopyrev, K.; and Liang, P. 2016. SQuAD: 100, 000+ Questions for Machine Comprehension of Text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, EMNLP 2016, Austin, Texas, USA, November 1-4, 2016, 2383–2392. The Association for Computational Linguistics.
- Sanh et al. (2019) Sanh, V.; Debut, L.; Chaumond, J.; and Wolf, T. 2019. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108.
- Schneider et al. (2020) Schneider, E. T. R.; de Souza, J. V. A.; Knafou, J.; Oliveira, L. E. S. e.; Copara, J.; Gumiel, Y. B.; Oliveira, L. F. A. d.; Paraiso, E. C.; Teodoro, D.; and Barra, C. M. C. M. 2020. BioBERTpt - A Portuguese Neural Language Model for Clinical Named Entity Recognition. In Proceedings of the 3rd Clinical Natural Language Processing Workshop, 65–72. Online: Association for Computational Linguistics.
- Socher et al. (2013) Socher, R.; Perelygin, A.; Wu, J.; Chuang, J.; Manning, C. D.; Ng, A. Y.; and Potts, C. 2013. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, EMNLP 2013, 18-21 October 2013, Grand Hyatt Seattle, Seattle, Washington, USA, A meeting of SIGDAT, a Special Interest Group of the ACL, 1631–1642. ACL.
- Sun et al. (2019) Sun, S.; Cheng, Y.; Gan, Z.; and Liu, J. 2019. Patient Knowledge Distillation for BERT Model Compression. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 4323–4332. Hong Kong, China: Association for Computational Linguistics.
- Sun et al. (2020) Sun, Z.; Yu, H.; Song, X.; Liu, R.; Yang, Y.; and Zhou, D. 2020. MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020, 2158–2170. Association for Computational Linguistics.
- Vaswani et al. (2017) Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, L.; and Polosukhin, I. 2017. Attention is All you Need. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, 5998–6008.
- Wang et al. (2019) Wang, A.; Singh, A.; Michael, J.; Hill, F.; Levy, O.; and Bowman, S. R. 2019. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net.
- Warstadt, Singh, and Bowman (2019) Warstadt, A.; Singh, A.; and Bowman, S. R. 2019. Neural Network Acceptability Judgments. Trans. Assoc. Comput. Linguistics, 7: 625–641.
- Williams, Nangia, and Bowman (2018) Williams, A.; Nangia, N.; and Bowman, S. R. 2018. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, New Orleans, Louisiana, USA, June 1-6, 2018, Volume 1 (Long Papers), 1112–1122. Association for Computational Linguistics.