A Quantitative Perspective on Values of Domain Knowledge for Machine Learning
Abstract
With the exploding popularity of machine learning, domain knowledge in various forms has been playing a crucial role in improving the learning performance, especially when training data is limited. Nonetheless, there is little understanding of to what extent domain knowledge can affect a machine learning task from a quantitative perspective. To increase the transparency and rigorously explain the role of domain knowledge in machine learning, we study the problem of quantifying the values of domain knowledge in terms of its contribution to the learning performance in the context of informed machine learning. We propose a quantification method based on Shapley value that fairly attributes the overall learning performance improvement to different domain knowledge. We also present Monte-Carlo sampling to approximate the fair value of domain knowledge with a polynomial time complexity. We run experiments of injecting symbolic domain knowledge into semi-supervised learning tasks on both MNIST and CIFAR10 datasets, providing quantitative values of different symbolic knowledge and rigorously explaining how it affects the machine learning performance in terms of test accuracy.
1 Introduction
Machine learning has achieved great success in numerous fields, such as computer vision, natural language processing, security, and robotics. The current success of machine learning, or more generally artificial intelligence (AI), owes much to the huge amount of available data. That being said, a priori domain knowledge, which comes from various sources and has multiple forms, also plays an increasingly more crucial role in the development of machine learning and may lead to new breakthroughs von Rueden et al. (2019); Wang et al. (2020b); Borghesi et al. (2020).
Domain knowledge has already been playing a part in current machine learning pipelines in many real-world applications. Specifically, machine learning with domain knowledge injection (a.k.a informed machine learning von Rueden et al. (2019)) can be accomplished in different ways. For example, in computer vision, image pre-processing such as flipping Benaim and Wolf (2018); Shyam et al. (2017), translation Benaim and Wolf (2018), cropping, Qi et al. (2018), scalingZhang et al. (2018) can help to augment data, which exploits the prior domain knowledge of image invariance. For another example, differential equations and logic rules from physical sciences and/or common knowledge provide additional constraints or new regularization terms for model training Battaglia et al. (2016); Borghesi et al. (2020); Silvestri et al. (2020); Muralidhar et al. (2018); Xu et al. (2018). Moreover, knowledge graphs can provide relational knowledge for reasoning or feature representation in deep learning Liang et al. (2018); Choi et al. (2017). In recent years, prior (domain) knowledge from one or more multiple teacher networks can also be injected into a learning task (i.e., student network) via distillation Hinton et al. (2015); You et al. (2017). As evidenced by many examples, such prior domain knowledge can effectively address the lack of labeled data commonly encountered in machine learning tasks, improving the learning performance in terms of various goals such as accuracy, generalization, and/or robustness.
In machine learning tasks, we need not only a model with good performance, but also model explainability Molnar (2020); Lipton (2018). The model explainability can help us better understand the model, use the model with more trust, or more quickly debug the deployed model. There have been many prior studies to address model explainability (especially in the context of deep neural networks or DNNs) from various perspectives, such as model features Molnar (2020), model inference Camburu (2020), and contribution of individual data samples Ghorbani and Zou (2019). Nonetheless, despite the increasingly wider usage of informed machine learning and and the clear advantages over its uninformed counterpart, a quantitative understanding of the actual contribution of domain knowledge injected into machine learning tasks has been lacking. This creates ambiguity about the role of domain knowledge in machine learning and decreases the explainability of the resulting models.
A foundation of domain knowledge explainability is to quantify the value or contribution of domain knowledge for machine learning tasks. In the first place, quantification of domain knowledge contribution gives us a better transparency of informed machine learning, which is part of the “right to explanation” legal requirement imposed in Europe for automated decision systems Goodman and Flaxman (2017). In addition, by domain knowledge quantification, we can know the actual contribution of different domain knowledge, which provides us with a principled guideline about what kind of domain knowledge is mostly needed to be collected in similar machine learning tasks. Last but not least, domain knowledge injected in machine learning may be implicit, inaccurate, and/or even corrupted. With domain knowledge quantification, we can make an informed decision as to which domain knowledge to utilize, since acquiring domain knowledge (e.g., coming from experts) may be quite expensive.
Nonetheless, it is non-trivial to quantify the value of domain knowledge in terms of its contribution of the performance of a learning task. First, domain knowledge has various forms and can be injected in machine learning in different ways von Rueden et al. (2019). Moreover, different domain knowledge is combined together and injected into a machine learning task, and the overall performance is typically not a simple linear combination of the performance corresponding to each piece of domain knowledge. Last but not least, we need to guarantee fairness for the quantification of domain knowledge in terms of its contribution to model performance.
To address these challenges, we leverage a classic tool from game theory — Shapley value — which provides a fair distribution of total payoff among several participants Shapley (1953). Shapley value has been recently used in machine learning model explainability to quantify the contribution of features Štrumbelj and Kononenko (2014), training data samples Ghorbani and Zou (2019), algorithms Yona et al. (2019), and neurons Ghorbani and Zou (2020). But, to our knowledge, quantifying the contribution of domain knowledge for machine learning is a novel perspective on model explainability that has not been considered. It will only become more important as informed machine learning becomes increasingly ubiquitous.
Our main contributions are summarized as follows.
-
•
We formulate the problem of quantifying the value of domain knowledge in informed machine learning, with a focus on the common approach to using regularization for domain knowledge injection into machine learning tasks.
-
•
We present a quantification method based on Shapley value to measure the value of domain knowledge in a fair manner. Additionally, we address the computational cost of Shapley value and give an approximated algorithm based on Monte-Carlo sampling, which achieves a polynomial time complexity.
-
•
As examples of quantifying the value of domain knowledge, we perform experiments on image classification tasks using semi-supervised learning over datasets of MNIST MNI and CIFAR10 CIF . Specifically, we consider symbolic domain knowledge and inject it into a learning task by adding a knowledge-related loss to the learning objective. Our experiments show the performance improvement by injecting different combinations of domain knowledge, from which we calculate the value of each domain knowledge for quantitative explainability. Our results also quantify and explain the performance degradation due to imperfect domain knowledge (on MNIST).
2 Related Work
Explanation of DNN models. Since most of the current DNN models are black-box and hard to explain, model interpretability/explainability has attracted much attention from both academia and industry Molnar (2020); Lipton (2018). The techniques to explain DNN or AI models can be broadly categorized into intrinsic interpretation techniques and post-hoc interpretation Molnar (2020), depending on whether the adopted explainability method is integrated with the learning process and affects the model. Intrinsic explainability techniques usually rely on some explainable/interpretable models like regression models, decision tree Zhang et al. (2019), attention models Heo et al. (2018), etc. More recently, intrinsic explainability techniques are also used in deep learning to achieve self-explanatory neural networks Alvarez Melis and Jaakkola (2018); Shen et al. (2019). Intrinsic explainability can have more accurate explanations for the output of DNN models, but it may also decrease the learning performance to some extent.
Unlike intrinsic explainability, post-hoc explainability uses a new explainable model to explain another more complex model after the model is trained. Many model-based methods have been proposed for post-hoc interpretability/explainability Greenwell et al. (2018); Ribeiro et al. (2018, 2016). By contrast, explainability based on Shapley value is a model-free method and has been becoming increasingly popular for AI explanation due to its flexibility. Shapley value is proved to be the only method that satisfies a set of axioms of fairness in payoff distribution problems Shapley (1953). Among existing techniques based on Shapley value, Štrumbelj and Kononenko (2014) explains the features of model input, Ghorbani and Zou (2019) studies explanation of the value of data samples, and Ghorbani and Zou (2020) quantifies the contribution of neurons in neural networks and proposes a Bayesian optimization algorithm to reduce the evaluation complexity. The joint contribution of learning algorithms and dataset are studied in Yona et al. (2019) using Shapley value. Also, Ancona et al. (2019) propose a polynomial algorithm to compute Shapley values and explain input features for DNNs. In this paper, we leverage Shapley value for model explainability from a new perspective — quantifying the value of domain knowledge.
Informed Machine Learning Many techniques have been developed to inject domain knowledge in machine learning tasks von Rueden et al. (2019); Wang et al. (2020b). One naive approach is to generate additional synthetic training data based on domain knowledge. For example, samples generated from simulations or demonstration can help mitigate the data shortage problem in reinforcement learning Deist et al. (2019); Gao et al. (2018b); Hester et al. (2017); Karpatne et al. (2017); Rai et al. (2019). Additionally, data can be generated by knowledge graphs Jiang et al. (2018), generative models Gao et al. (2018a), etc. Domain knowledge can also benefit machine learning tasks by choosing hypothesis sets like regression model and neural architectures. A good hypothesis set can improve the learning performance as well as reduce the training complexity. Examples include domain knowledge based neural architecture search Chen et al. (2020); Towell and Shavlik (1994) and hyper-parameter tuning Mendoza et al. (2016).
Injecting domain knowledge into machine learning tasks via a loss term in the training objective function is another common method. For example, AI models applied in real world are often trained on the optimization objective with constraints originated from domain knowledge Borghesi et al. (2020); Silvestri et al. (2020); Muralidhar et al. (2018). Knowledge distillation from one or more multiple teacher networks also falls into this category of knowledge injection Hinton et al. (2015); You et al. (2017). Moreover, recent studies have shown that logical rules can also be integrated into machine learning by including a domain knowledge related loss term Xu et al. (2018); Hu et al. (2016).
To our knowledge, however, the values of domain knowledge injected into informed machine learning have not been rigorously or fairly quantified.
3 Informed Machine Learning
A key goal of machine learning is to find the intrinsic hypothesis for a learning task. Typically, a hypothesis represents a mapping from an input variable to an output variable . In some cases, a hypothesis is a distribution over the mapping. In conventional machine learning, a training dataset is provided and a hypothesis set is determined, usually by choosing a learning model . With a certain loss function and learning algorithm, the model is learnt to approximate the optimal hypothesis which has the optimal performance among the chosen hypothesis set . For example, given a performance metric (which can be risk, cost, or negative reward/utility), the optimal hypothesis in the chosen hypothesis set is . By training based on the dataset , we get a final hypothesis .
For a machine learning task, various forms of domain knowledge from multiple sources can be integrated into the machine learning pipeline for improving the model performance. For example, domain knowledge can be used for data augmentation von Rueden et al. (2019); Wang et al. (2020b) by generating synthetic data samples via simulations Deist et al. (2019); Hester et al. (2017); Karpatne et al. (2017), data transforming Qi et al. (2018); Zhang et al. (2018), and generative models Goodfellow et al. (2014); Gao et al. (2018a). With more (independent) samples in the augmented dataset , the generalization error can be reduced Guedj and Shawe-Taylor (2019). Also, domain knowledge is important for choosing a hypothesis set . This is usually done by choosing a learning model , such as knowledge-based neural architecture search Chen et al. (2020) and hyper-parameter tuning Mendoza et al. (2016). Moreover, domain knowledge can be used to decide whether some final hypothesis should be discarded von Rueden et al. (2019).
In this paper, we focus on another important approach to domain knowledge injection: adding knowledge-based constraints or loss terms as part of the objective function Borghesi et al. (2020); Silvestri et al. (2020); Muralidhar et al. (2018); Xu et al. (2018). Specifically, with a data-based loss used in conventional machine learning and a loss based on domain knowledge applied to its corresponding (possibly unlabelled) dataset , for , the training objective based on domain knowledge can be expressed as
(1) |
where is used to balance the pure data-based loss and knowledge-based loss. In addition to knowledge-based constraints or losses, this formulation also applies to data augmentation based on domain knowledge: is the synthetic training dataset generated based on domain knowledge and we have for .
Assume that is the performance metric of interest in a machine learning task given a set containing pieces of domain knowledge . For conciseness, represents the -th domain knowledge as well as the corresponding (possibly unlabelled) dataset . With the final hypothesis learnt based on both training dataset and domain knowledge in , the learning performance is denoted as .
4 The Value of Domain Knowledge
For informed machine learning with domain knowledge, the model performance highly relies on the included domain knowledge information. Naturally, different domain knowledge can have different contributions to the learning performance, and the performance metric can change a lot by adding, removing or revising a subset of the domain knowledge. Thus, to explain the learnt model and achieve better transparency, it is crucial to quantify the value of each domain knowledge for the considered learning task in terms of the metric of interest. Moreover, the quantification of domain knowledge values should be fair. In this section, we give a set of axioms for fair quantification of domain knowledge values and propose a method based on Shapley value to achieve fair quantification.
4.1 Axioms for Fairness
If a set of knowledge , together with a labelled training dataset , jointly determines the model performance, the performance improvement contributed by the domain knowledge set is denoted as
(2) |
where is the performance without domain knowledge set . Alternatively, if the domain knowledge is already included into a machine learning task, the performance improvement due to the specific domain knowledge set can be expressed as
(3) |
where is the hypothesis averaged over all the possible domain knowledge.
To quantify the contribution of individual domain knowledge,
the performance improvement should be attributed
into individual contributions by different domain knowledge in the set . Importantly, a fair attribution should meet the following basic axioms Shapley (1953); Ghorbani and
Zou (2019).
Axiom 1 (Efficiency)
The performance improvement contributed by the entire
domain knowledge set is the sum of contributions by
individual domain knowledge, i.e.
(4) |
Axiom 2 (Symmetry) If two different domain knowledge and are exchangeable and not distinguishable in terms of the performance improvement, they should have equal contributions. Thus, for any knowledge subset , we have
(5) |
Axiom 3 (Null-player) If individual domain knowledge in has no contribution to the performance improvement , the attributed contribution of is zero. That is, for any domain knowledge subset , we have
(6) |
4.2 Domain Knowledge Shapley
Since domain knowledge in the set combined together achieves performance improvement , it is non-trivial to find a contribution attribution method that satisfies all the above three axioms. To address this, we provide a attribution method based on Shapley value — Domain Knowledge Shapley — to quantify the contribution of different domain knowledge. Importantly, Shapley value, a classic tool from cooperative game theory, is proved to be the only method that satisfies the axioms (along with other properties) for fair attribution, as will be introduced in detail next Shapley (1953).
In our problem, the performance improvement resulting from the domain knowledge set can be seen as the total payoff to be distributed and each domain knowledge in the set can be seen as a player in a cooperative game. Thus, domain knowledge contribution is also the payoff distributed to each player. Shapley value with respect to the -th player representing domain knowledge is calculated as
(7) |
where is the number of combinations with size from the total knowledge set , and is the marginal contribution of domain knowledge on the basis of a domain knowledge subset . Intuitively, Shapley value quantifies the contribution of domain knowledge as the average marginal contribution of over all the possible . The reason to average over the domain knowledge subset is fairness: the order of adding each domain knowledge for quantifying its marginal contribution should be equal for all domain knowledge.
4.3 Approximate Method
Although Shapley value is a provably fair in quantifying the contribution of domain knowledge, it needs evaluations in total, accounting for different combinations of domain knowledge Shapley (1953). Thus, the computational cost of Shapley value is exponentially increases with the size of the knowledge set , and can quickly become intolerable for some practical machine learning tasks that have both high training and testing costs. As a result, explaining DNN models (e.g., the contribution of training data samples Ghorbani and Zou (2019)) based on Shapley value typically leverage approximate methods.
Monte-Carlo sampling is a common method to approximate the computation of Shapley values in the context of DNN explainability Castro et al. (2009); Maleki et al. (2013). Specifically, we denote as the set of permutation of and rewrite domain knowledge Shapley in Eqn. (7) based on Castro et al. (2009) as follows:
(8) |
where is the set of predecessors of in a permutation . Based on Eqn. (8), the domain knowledge Shapley for can be estimated by sampling from the permutation of domain knowledge set . The approximated knowledge Shapley algorithm is given in Algorithm 1. Critically, Shapley approximation based on Monte-Carlo sampling has a polynomial-time complexity and has been proven to have a bounded error with a high probability Maleki et al. (2013).
While Monte-Carlo sampling can effectively reduce the complexity, the evaluation for each marginal contribution can still be costly due to the requirement of calculating the values of and . To further reduce the cost of evaluating the learning performance with different combinations of domain knowledge, we can train the model over a smaller number of epochs and also use a smaller dataset for testing. Moreover, if the size of domain knowledge set is very large, we can also train an predictor for the learning performance that takes the domain knowledge subset as input and directly estimates the resulting learning performance without training. Such performance predictors are also used for avoiding the cost of model training for each architecture candidate and speeding up neural architecture search Dai et al. (2019); Wang et al. (2020a).
5 Experiments
In this section, we run experiments to quantify the values of domain knowledge for two image classification tasks under a semi-supervised setting for which domain knowledge can significantly improve the learning performance.
5.1 DNN with Symbolic Domain Knowledge
In our experiments, the domain knowledge comes from the symbolic label knowledge which can be expressed as a logical sentence Xu et al. (2018); Diligenti et al. (2017). Specifically, a logical sentence is a set of assertions combined by relational expressions such as (and), (or) and (not). For example, for states , a logical sentence can be . Assume that for a logical sentence and a state , means that the state satisfies the logical knowledge , or in other words, the logical sentence can be written as . In the previous example, we have and .
A common approach to injecting symbolic knowledge in DNNs is adding a knowledge-based loss in the training objective Xu et al. (2018). Consider a classification task where is the input and is the label. We use a neural network as our classification model. Denote the output of as which is usually interpreted as the estimated probability that the corresponding label is the correct label. We include an additional loss term based on the symbolic knowledge as proposed in Xu et al. (2018), which is the negative logarithm of the estimated probability that the knowledge is satisfied. If represents a state of the sparse label space , given an domain knowledge in the logical sentence form and an input , the knowledge-based loss can be expressed as
(9) |
For example, if the domain knowledge means that only one label in is the correct output, then the knowledge-based loss is
(10) |
Similarly, if the knowledge means that the correct label with respect to input is one of the labels in the set , the knowledge-based loss can be expressed as
(11) |
The above domain knowledge injection method can be used to improve the learning performance for semi-supervised classification tasks. Specifically, we are given a training dataset with input-label pairs as well as an unlabelled dataset . Denote as the cross-entropy loss of an instance . The semi-supervised training objective with a set of knowledge injected is expressed as
(12) |
Given a test input , the label is usually predicted as . While the performance of prediction can be evaluated by various metrics, we consider the widely-used top-1 accuracy, which can be formally expressed as . In practice, however, a test dataset is usually used to empirically evaluate the accuracy, which is called the test accuracy:
(13) |
5.2 Results on MNIST
We begin our experiment of semi-supervised learning on a simple dataset — MNIST, which includes 60000 handwritten digit images with labels for training and 10000 images with labels for testing MNI . We use only 100 images (including 10 images for each digit) and their corresponding labels from the training dataset as labeled samples, and use the rest of images in the training dataset as unlabeled samples. The test accuracy is calculated on the entire testing dataset of 10000 images.
Knowledge (Accurate) | Test Accuracy | Knowledge (Imperfect C-II) | Test Accuracy |
---|---|---|---|
No Knowledge | 0.8372 | No Knowledge | 0.8372 |
One Hot | 0.8725 | One Hot | 0.8725 |
C-I | 0.9224 | C-I | 0.9224 |
C-II | 0.9654 | C-II | 0.9425 |
One Hot & C-I | 0.9396 | One Hot & C-I | 0.9396 |
One Hot & C-II | 0.9686 | One Hot & C-II | 0.9590 |
C-I & C-II | 0.9733 | C-I & C-II | 0.9678 |
One Hot & C-I & C-II | 0.9775 | One Hot & C-I & C-II | 0.9687 |



We consider a domain knowledge set {One Hot, Constraint-I, Constraint-II}. The knowledge One Hot as considered in Xu et al. (2018) means that each input has only one correct label, whose knowledge-based loss is expressed in Eqn. (9). The knowledge Constraint-I specifies whether or not the digit of an image belongs to the interval , while the knowledge Constraint-II informs us of whether or not the digit of an image is in the interval . Additionally, we quantify the knowledge values under the setting where Constraint-II in the knowledge set is imperfect. Specifically, we consider that 10% of the unlabelled samples have wrong information about Constraint-II: if the true label of a sample with corrupted knowledge Constraint-II is in the interval , the corrupted knowledge Constraint-II believes it is out of the interval and vice versa. The loss based on knowledge Constraint-I and Constraint-II can both be calculated according to Eqn. (10).
Our training setting follows Xu et al. (2018). Specifically, we consider a DNN with four fully-connected hidden layers having 1000, 500, 250, 250 hidden neurons respectively, a batch normalization layer and a dropout layer with dropout probability 0.5. The weight to balance the cross-entropy loss and the knowledge-based loss is 0.1. During training, each batch has 16 images with corresponding labels and 16 images without labels (100 labeled images are reused to balance the number of labeled and unlabeled examples). Adam optimizer with learning rate is used for updating weights followed by a tuning phase with learning rates and . Note that since we only have three different pieces of domain knowledge and our DNN has a reasonably low training cost, we re-train the DNN given each combination of domain knowledge without Monte-Carlo sampling.
The test accuracies given each combination of domain knowledge are shown in Table 1. Based on the accuracy results, we calculate the value of each domain knowledge in Fig. 1(a). We see that the knowledge Constraint-II has the largest contribution, followed by the knowledge Constraint-I, in terms of the test accuracy. The knowledge One Hot contributes the least, since it is simply common knowledge. Our study provides a rigours quantification of values of different knowledge.
In addition, we show in Table 1 the test accuracies given different domain knowledge combinations under the setting where Constraint-II is imperfect. Accordingly, the quantified contribution of knowledge in terms of test accuracy is shown in Fig. 1(b). We can see in Fig. 1(a) that the contribution of imperfect knowledge Constraint-II becomes naturally less than that of accurate knowledge Constraint-II, while the contribution of knowledge One Hot and Constraint-I increases slightly. This implies that the performance degradation is mostly caused by the imperfect knowledge Constraint-II, which is consistent with our intuition. Interestingly, although the knowledge Constraint-I is more accurate than Constraint-II, it does not provide as much contribution to the test accuracy improvement as Constraint-II.
5.3 Results on CIFAR10
Knowledge Combinations | Test Accuracy |
---|---|
No Knowledge | 0.7558 |
One Hot | 0.7716 |
Animal | 0.7948 |
Mammal | 0.8024 |
One Hot & Animal | 0.7983 |
One Hot & Mammal | 0.8105 |
Animal & Mammal | 0.8177 |
One Hot & Animal & Mammal | 0.8208 |
We also run experiments on the CIFAR10 dataset, which includes images with pixels and 10 classes (airplane, automobile, bird, cat, deer, dog, frog, horse, ship, truck) in the label space CIF . It has 50000 training samples and 10000 test examples. We use only 10% of the training images with labels (5000 labeled images), and another 15000 training images without labels for training. For testing, we use all the 10000 test images in the test dataset.
Our domain knowledge set has three different types of symbolic knowledge {One Hot, Animal, Mammal}. The knowledge One Hot provides the information that each image has only one correct label, whose knowledge-based loss is expressed in Eqn. (9). The knowledge Animal provides the information whether or not the label of an image belongs to animals. With the knowledge Animal , we can decide if the label of an input is in the label subset , even though the exact label is unknown. The knowledge Mammal provides information of whether or not the label of a training image belongs to mammals. Similarly, the knowledge Mammal specifies whether the label of a training image is in the label subset {cat, deer, dog, horse}. The loss based on knowledge Animal and Mammal can be calculated by Eqn. (10).
In our experiment, Ladder Net Pitelis et al. (2014); Xu et al. (2018) is used as the neural architecture. The weight to balance the cross entropy loss and knowledge-based loss is set as 0.1. A batch of 512 inputs consist of both labeled images and unlabeled images are used for each weight update. To balance the number of labeled and unlabeled inputs, the labeled inputs are reused. Training is performed by the momentum algorithm with a learning rate of 0.1, followed by a tuning phase with learning rates 0.05, 0.01 and 0.0001.
After training the neural network with different combinations of domain knowledge, we evaluate the test accuracy as shown in Table 2. The domain knowledge Shapley in terms of the test accuracy improvement is calculated and shown in Fig. 1(c). We can see that the knowledge Mammal is the most informative one and hence has the largest contribution to the test accuracy, while the common-sense knowledge One Hot contributes the least to the test accuracy improvement. The knowledge Animal has a contribution between the other two knowledge.
6 Conclusion
In this paper, we study valuation of domain knowledge and provide a first step towards quantitative understanding of contribution of domain knowledge injected into informed machine learning. Specifically, we propose domain knowledge Shapley to quantify the values of domain knowledge in terms of the learning performance improvement for machine learning tasks in a fair manner. We run experiments on both MNIST and CIFAR10 datasets using semi-supervised learning and quantitatively measure the fair value of different domain knowledge. Our results also quantify the extent to which imperfect domain knowledge can affect the learning performance (for the MNIST dataset).
References
- Alvarez Melis and Jaakkola [2018] David Alvarez Melis and Tommi Jaakkola. Towards robust interpretability with self-explaining neural networks. Advances in Neural Information Processing Systems, 31:7775–7784, 2018.
- Ancona et al. [2019] Marco Ancona, Cengiz Oztireli, and Markus Gross. Explaining deep neural networks with a polynomial time algorithm for shapley value approximation. In International Conference on Machine Learning, pages 272–281, 2019.
- Battaglia et al. [2016] Peter Battaglia, Razvan Pascanu, Matthew Lai, Danilo Jimenez Rezende, et al. Interaction networks for learning about objects, relations and physics. In Advances in neural information processing systems, pages 4502–4510, 2016.
- Benaim and Wolf [2018] Sagie Benaim and Lior Wolf. One-shot unsupervised cross domain translation. In Advances in Neural Information Processing Systems, pages 2104–2114, 2018.
- Borghesi et al. [2020] Andrea Borghesi, Federico Baldo, and Michela Milano. Improving deep learning models via constraint-based domain knowledge: a brief survey. arXiv preprint arXiv:2005.10691, 2020.
- Camburu [2020] Oana-Maria Camburu. Explaining deep neural networks. arXiv preprint arXiv:2010.01496, 2020.
- Castro et al. [2009] Javier Castro, Daniel Gómez, and Juan Tejada. Polynomial calculation of the shapley value based on sampling. Computers & Operations Research, 36(5):1726–1730, 2009.
- Chen et al. [2020] Yaran Chen, Ruiyuan Gao, Fenggang Liu, and Dongbin Zhao. Modulenet: Knowledge-inherited neural architecture search. arXiv preprint arXiv:2004.05020, 2020.
- Choi et al. [2017] Edward Choi, Mohammad Taha Bahadori, Le Song, Walter F Stewart, and Jimeng Sun. Gram: graph-based attention model for healthcare representation learning. In ACM International Conference on Knowledge Discovery and Data Mining, pages 787–795, 2017.
- [10] The cifar-10 dataset (https://www.cs.toronto.edu/~kriz/cifar.html).
- Dai et al. [2019] Xiaoliang Dai, Peizhao Zhang, Bichen Wu, Hongxu Yin, Fei Sun, Yanghan Wang, Marat Dukhan, Yunqing Hu, Yiming Wu, Yangqing Jia, et al. ChamNet: Towards efficient network design through platform-aware model adaptation. In CVPR, 2019.
- Deist et al. [2019] Timo M Deist, Andrew Patti, Zhaoqi Wang, David Krane, Taylor Sorenson, and David Craft. Simulation-assisted machine learning. Bioinformatics, 35(20):4072–4080, 2019.
- Diligenti et al. [2017] Michelangelo Diligenti, Marco Gori, and Claudio Sacca. Semantic-based regularization for learning and inference. Artificial Intelligence, 244:143–165, 2017.
- Gao et al. [2018a] Hang Gao, Zheng Shou, Alireza Zareian, Hanwang Zhang, and Shih-Fu Chang. Low-shot learning via covariance-preserving adversarial augmentation networks. In Advances in Neural Information Processing Systems, pages 975–985, 2018.
- Gao et al. [2018b] Yang Gao, Huazhe Xu, Ji Lin, Fisher Yu, Sergey Levine, and Trevor Darrell. Reinforcement learning from imperfect demonstrations. arXiv preprint arXiv:1802.05313, 2018.
- Ghorbani and Zou [2019] Amirata Ghorbani and James Zou. Data shapley: Equitable valuation of data for machine learning. arXiv preprint arXiv:1904.02868, 2019.
- Ghorbani and Zou [2020] Amirata Ghorbani and James Zou. Neuron shapley: Discovering the responsible neurons. arXiv preprint arXiv:2002.09815, 2020.
- Goodfellow et al. [2014] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in neural information processing systems, pages 2672–2680, 2014.
- Goodman and Flaxman [2017] Bryce Goodman and Seth Flaxman. European union regulations on algorithmic decision-making and a “right to explanation”. AI Magazine, 38(3):50–57, Oct. 2017.
- Greenwell et al. [2018] Brandon M Greenwell, Bradley C Boehmke, and Andrew J McCarthy. A simple and effective model-based variable importance measure. arXiv preprint arXiv:1805.04755, 2018.
- Guedj and Shawe-Taylor [2019] Benjamin Guedj and John Shawe-Taylor. A primer on pac-bayesian learning. International Conference on Machine Learning, 2019.
- Heo et al. [2018] Jay Heo, Hae Beom Lee, Saehoon Kim, Juho Lee, Kwang Joon Kim, Eunho Yang, and Sung Ju Hwang. Uncertainty-aware attention for reliable interpretation and prediction. In Advances in Neural Information Processing Systems, pages 909–918, 2018.
- Hester et al. [2017] Todd Hester, Matej Vecerik, Olivier Pietquin, Marc Lanctot, Tom Schaul, Bilal Piot, Dan Horgan, John Quan, Andrew Sendonaris, Gabriel Dulac-Arnold, et al. Deep q-learning from demonstrations. arXiv preprint arXiv:1704.03732, 2017.
- Hinton et al. [2015] Geoffrey Hinton, Oriol Vinyals, and Jeffrey Dean. Distilling the knowledge in a neural network. In NIPS Deep Learning and Representation Learning Workshop, 2015.
- Hu et al. [2016] Zhiting Hu, Xuezhe Ma, Zhengzhong Liu, Eduard Hovy, and Eric Xing. Harnessing deep neural networks with logic rules. Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2410–2420, 2016.
- Jiang et al. [2018] Chenhan Jiang, Hang Xu, Xiaodan Liang, and Liang Lin. Hybrid knowledge routed modules for large-scale object detection. In Advances in Neural Information Processing Systems, pages 1552–1563, 2018.
- Karpatne et al. [2017] Anuj Karpatne, William Watkins, Jordan Read, and Vipin Kumar. Physics-guided neural networks (pgnn): An application in lake temperature modeling. arXiv preprint arXiv:1710.11431, 2017.
- Liang et al. [2018] Xiaodan Liang, Zhiting Hu, Hao Zhang, Liang Lin, and Eric P Xing. Symbolic graph reasoning meets convolutions. In Advances in Neural Information Processing Systems, pages 1853–1863, 2018.
- Lipton [2018] Zachary C Lipton. The mythos of model interpretability. Queue, 16(3):31–57, 2018.
- Maleki et al. [2013] Sasan Maleki, Long Tran-Thanh, Greg Hines, Talal Rahwan, and Alex Rogers. Bounding the estimation error of sampling-based shapley value approximation. arXiv preprint arXiv:1306.4265, 2013.
- Mendoza et al. [2016] Hector Mendoza, Aaron Klein, Matthias Feurer, Jost Tobias Springenberg, and Frank Hutter. Towards automatically-tuned neural networks. In Workshop on Automatic Machine Learning, pages 58–65, 2016.
- [32] The mnist database of handwritten digits (http://yann.lecun.com/exdb/mnist/).
- Molnar [2020] Christoph Molnar. Interpretable Machine Learning. Lulu. com, 2020.
- Muralidhar et al. [2018] Nikhil Muralidhar, Mohammad Raihanul Islam, Manish Marwah, Anuj Karpatne, and Naren Ramakrishnan. Incorporating prior domain knowledge into deep neural networks. In IEEE International Conference on Big Data, pages 36–45, 2018.
- Pitelis et al. [2014] Nikolaos Pitelis, Chris Russell, and Lourdes Agapito. Semi-supervised learning using an unsupervised atlas. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 565–580, 2014.
- Qi et al. [2018] Hang Qi, Matthew Brown, and David G Lowe. Low-shot learning with imprinted weights. In IEEE conference on computer vision and pattern recognition, pages 5822–5830, 2018.
- Rai et al. [2019] Akshara Rai, Rika Antonova, Franziska Meier, and Christopher G Atkeson. Using simulation to improve sample-efficiency of bayesian optimization for bipedal robots. Journal of Machine Learning Research, 20:49–1, 2019.
- Ribeiro et al. [2016] Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. Model-agnostic interpretability of machine learning. ICML Workshop on Human Interpretability in Machine Learning, 2016.
- Ribeiro et al. [2018] Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. Anchors: High-precision model-agnostic explanations. In AAAI Conference on Artificial Intelligence, volume 18, pages 1527–1535, 2018.
- Shapley [1953] Lloyd S Shapley. A value for n-person games. Contributions to the Theory of Games, 2(28):307–317, 1953.
- Shen et al. [2019] Shiwen Shen, Simon X Han, Denise R Aberle, Alex A Bui, and William Hsu. An interpretable deep hierarchical semantic convolutional neural network for lung nodule malignancy classification. Expert systems with applications, 128:84–95, 2019.
- Shyam et al. [2017] Pranav Shyam, Shubham Gupta, and Ambedkar Dukkipati. Attentive recurrent comparators. In International Conference on Machine Learning, pages 3173–3181, 2017.
- Silvestri et al. [2020] Mattia Silvestri, Michele Lombardi, and Michela Milano. Injecting domain knowledge in neural networks: a controlled experiment on a constrained problem. arXiv preprint arXiv:2002.10742, 2020.
- Štrumbelj and Kononenko [2014] Erik Štrumbelj and Igor Kononenko. Explaining prediction models and individual predictions with feature contributions. Knowledge and information systems, 41(3):647–665, 2014.
- Towell and Shavlik [1994] Geoffrey G Towell and Jude W Shavlik. Knowledge-based artificial neural networks. Artificial intelligence, 70(1-2):119–165, 1994.
- von Rueden et al. [2019] Laura von Rueden, Sebastian Mayer, Katharina Beckh, Bogdan Georgiev, Sven Giesselbach, Raoul Heese, Birgit Kirsch, Julius Pfrommer, Annika Pick, Rajkumar Ramamurthy, et al. Informed machine learning–a taxonomy and survey of integrating knowledge into learning systems. arXiv preprint arXiv:1903.12394, 2019.
- Wang et al. [2020a] Tianzhe Wang, Kuan Wang, Han Cai, Ji Lin, Zhijian Liu, Hanrui Wang, Yujun Lin, and Song Han. APQ: Joint search for network architecture, pruning and quantization policy. In CVPR, 2020.
- Wang et al. [2020b] Yaqing Wang, Quanming Yao, James T Kwok, and Lionel M Ni. Generalizing from a few examples: A survey on few-shot learning. ACM Computing Surveys, 53(3):1–34, 2020.
- Xu et al. [2018] Jingyi Xu, Zilu Zhang, Tal Friedman, Yitao Liang, and Guy Broeck. A semantic loss function for deep learning with symbolic knowledge. In International Conference on Machine Learning, pages 5502–5511, 2018.
- Yona et al. [2019] Gal Yona, Amirata Ghorbani, and James Zou. Who’s responsible? jointly quantifying the contribution of the learning algorithm and training data. arXiv preprint arXiv:1910.04214, 2019.
- You et al. [2017] Shan You, Chang Xu, Chao Xu, and Dacheng Tao. Learning from multiple teacher networks. In ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2017.
- Zhang et al. [2018] Yabin Zhang, Hui Tang, and Kui Jia. Fine-grained visual categorization using meta-learning optimization with sample selection of auxiliary data. In European conference on computer vision, pages 233–248, 2018.
- Zhang et al. [2019] Quanshi Zhang, Yu Yang, Haotian Ma, and Ying Nian Wu. Interpreting cnns via decision trees. In IEEE Conference on Computer Vision and Pattern Recognition, pages 6261–6270, 2019.