Learning Accurate and Interpretable Decision Rule Sets from Neural Networks
Abstract
This paper proposes a new paradigm for learning a set of independent logical rules in disjunctive normal form as an interpretable model for classification. We consider the problem of learning an interpretable decision rule set as training a neural network in a specific, yet very simple two-layer architecture. Each neuron in the first layer directly maps to an interpretable if-then rule after training, and the output neuron in the second layer directly maps to a disjunction of the first-layer rules to form the decision rule set. Our representation of neurons in this first rules layer enables us to encode both the positive and the negative association of features in a decision rule. State-of-the-art neural net training approaches can be leveraged for learning highly accurate classification models. Moreover, we propose a sparsity-based regularization approach to balance between classification accuracy and the simplicity of the derived rules. Our experimental results show that our method can generate more accurate decision rule sets than other state-of-the-art rule-learning algorithms with better accuracy-simplicity trade-offs. Further, when compared with uninterpretable black-box machine learning approaches such as random forests and full-precision deep neural networks, our approach can easily find interpretable decision rule sets that have comparable predictive performance.
Introduction
Machine learning is finding its way to impact every sector of our society, including healthcare, information technology, transportation, entertainment, business, and criminal justice. In recent years, machine learning using neural networks have made tremendous advances in solving perceptual tasks like computer vision and natural language processing, with breakthrough performance in classification accuracy and generalization capability. However, neural network methods have generally produced black box models that are difficult or impossible for humans to understand. Their lack of interpretability makes it difficult to gain public trust for their use in high-stakes human-centered applications like medical-diagnosis and criminal justice, where decisions can have serious consequences on human lives (Rudin 2019).
Indeed, interpretability is a well-recognized goal in the machine learning community.
One popular approach to interpretable models is the use of decision rule sets (Cohen 1995; Su et al. 2015; Lakkaraju, Bach, and Leskovec 2016; Wang et al. 2017; Dash, Günlük, and Wei 2018), where the model comprises an unordered set of independent logical rules in disjunctive normal form (DNF).
Decision rule sets are inherently interpretable because the rules are expressed in simple IF-THEN sentences that correspond to logical combinations of input conditions that must be satisfied for a classification.
An example of a decision rule set with three clauses is as follows:
IF
(age 50) OR
(NOT smoker) OR
(cholesterol 130 AND blood pressure 120)
THEN
low heart disease risk.
In this example, the model would predict someone to have a low risk for heart disease if the person’s cholesterol level and blood pressure are below the specified thresholds.
The model not only provides a prediction, but the corresponding matching rule also provides an explanation that humans can easily understand.
In particular, the explanations are stated directly in terms of the input features, which can be categorical (e.g., color equal to red, blue, or green) or numerical (e.g., age 50) attributes, where the binary encoding of categorical and numerical attributes is well-studied (Wang et al. 2017; Dash, Günlük, and Wei 2018).
In this paper, we propose a new paradigm for learning accurate and interpretable decision rule sets as a neural network training problem. In particular, we consider the problem of learning an interpretable decision rule set as training a neural network in a simple two-layer fully-connected neural network architecture called a Decision Rules Network (DR-Net). In the first layer, called the Rules Layer, each trainable neuron with binary activation directly maps to a logical IF-THEN rule after training, where a positive input weight corresponds to a positive association of the input feature, a negative input weight corresponds to a negative association of the input feature, and a zero weight corresponds to an exclusion of the input feature. In the second layer, called the OR Layer, the trainable output neuron with binary activation directly maps to a disjunction of the first-layer rules to form the decision rule set.
By formulating the interpretable rules learning problem as a neural net training problem, state-of-the-art training approaches (including recent advances) can be harnessed for learning highly accurate classification models, including well-developed stochastic gradient descent algorithms for effective training. We are also able to leverage well-developed regularization concepts developed in the neural net community to trade off accuracy and model complexity in the training process. In particular, we propose a sparsity-based regularization approach in which the model complexity in terms of the length of the rules and the number of rules are captured in a regularization loss function. Minimizing the number of decision rules makes it easier for a user to understand all the conditions that correspond to a classification, and minimizing the lengths of the decision rules makes it easier for a user to interpret the explanations. This regularization loss function can be combined with a binary cross-entropy loss function that measures training accuracy, so that the training process can balance between classification accuracy and the simplicity of the derived rule set.
Other benefits of a neural net based formulation is the availability of sophisticated development frameworks (Abadi et al. 2016; Paszke et al. 2017) for model development, powerful computing platforms (e.g., GPUs and deep learning accelerators) for efficient learning and inference, and other developments like federated learning (Konečný et al. 2017) that enables multiple entities to collaboratively learn a common, robust model without sharing data, which addresses critical data privacy and security concerns.
In comparison with previous rule-learning approaches, our approach has several notable advantages. In (Lakkaraju, Bach, and Leskovec 2016; Wang et al. 2017), the pre-mining of frequent rule patterns is first used to produce a set of candidate rules, from which various algorithmic approaches are used to select a set of rules from these candidates. However, the requirement for pre-mining frequent rules limits the overall search space, thus hindering the algorithms from obtaining a globally optimized model. In (Su et al. 2015; Dash, Günlük, and Wei 2018), the problem is formulated as an integer-programming problem in which the pre-mining of rules is not required, but approximations are required to solve large scale problems. In contrast, our neural net based approach does not require rules mining and can take advantage of well-developed neural net training techniques to derive better interpretable models. By connecting interpretable rule-based learning to a neural network based formulation, we hope to open a new line of research that will lead to further fruitful results in the future.
Our experimental results show that our method can generate more accurate decision rule sets than other state-of-the-art rule-learning algorithms with better accuracy-simplicity trade-offs. Further, when compared with uninterpretable black box machine learning approaches such as random forests and full-precision deep neural networks, our approach can easily find interpretable decision rule sets that have comparable predictive performance.
Decision Rules Network
Given a classification dataset with binarized input features, our goal is to train a classifier in the form of a Boolean logic function in disjunctive normal form (OR-of-ANDs). In particular, each of the lower level conjunctive clauses (logical ANDs), which consists of a subset of input features and their negations, individually serves as a decision rule. An instance satisfies a conjunctive clause if all conditions specified in the clause are true in the instance. In the upper level of the function, all conjunctive clauses are unified by a disjunction (logical OR). Thus, a negative final prediction is produced only if none of the conjunctive clauses are satisfied. Otherwise, a positive final prediction will be made.
Mathematically, the training set contains data samples , , where comprises binarized features , , and . The final decision rule set learned from our method comprises parallel rules that we denote as clauses: . We define a clause to be a conjunction of predicates where and a predicate to be either an input feature or the negation of an input feature . If an input feature or the negation of an input feature is not present in clause , then we say that feature is excluded from clause , i.e. whether is or has no effect to the prediction of clause . Under this definition, an instance satisfies a clause only if all predicates in the clause are true in the instance i.e. for and for .
In this section, we introduce the architecture of our Decision Rules Network (DR-Net), which is a simple two-layer fully-connected neural network. The first layer, called the Rules Layer, consists of trainable neurons that map to logical IF-THEN rules, and the second layer, called the OR Layer, contains a trainable output neuron that maps to a disjunction of the first-layer rules to form the decision rule set. The goal of the design of this network is to simulate the logical formula in disjunctive normal form so that a trained DR-Net can be directly mapped to a set of interpretable decision rules.
Handling of Categorical and Numerical Attributes
Common tabular datasets generally comprise binary, categorical and numerical features. While our method is based on binary encoded input vectors, we employ the following pre-processing procedures, which are well established and studied in the machine learning literature, to binarize the input features. In particular, the values of binary features are left as what they are, whereas we apply standard one-hot encoding to transform categorical attributes to vectors of binary values. As for numerical features, we adopt quantile discretization to get a set of thresholds for each feature, where the original numerical value is one-hot-encoded into a binary vector by comparing with the thresholds (e.g., age 25, age 50, age 75) and encoded as if less than the threshold or otherwise. For example, considering a dataset that consists of the categorical feature “color” chosen from {red, green, blue} and a numerical feature “age” with thresholds {25, 50, 75}, our pre-processing approach will encode an instance [color: red, age: 30] as [red, green, blue, age 25, age 50, age 75] [1, 0, 0, 0, 1, 1]. Most other rule-learning methods (Wang et al. 2017; Dash, Günlük, and Wei 2018) require to convert binary, categorical and numerical features into both positive conditions, e.g., (color = blue) and (age 50), and negative conditions, e.g., (color blue) and (age 50), of the binary vectors in their pre-processing procedures. On the other hand, our encoding approach only involves those positive conditions without separately having their negations included. Further explanations will be discussed in the next section.
Rules Layer
The essence of a fully-connected layer is the dot-product operation shifted by a bias term. In this context, we notice that with binarized input features, a neuron can be constructed such that it effectively performs a logical AND operation by dynamically adjusting the bias based on the weight values and applying a binary step activation function afterwards. Then, by interpreting the full precision weights in a certain way, each neuron is effectively a conjunction of input features and thus the whole layer can be mapped to a set of clauses that can be later combined with disjunction to form a DNF rule set.
Mathematically, given the input to the Rules Layer as and the output as , a neuron in the Rules Layer performs its operation as follows:
(1) |
In Equation 1, the dot product of the weights and inputs is added with a dynamic bias, which depends on the weights of the neuron. With the dynamic bias and binarized inputs, the range of the outputs of the neurons in the Rules Layer is within . Note that the output can only be achieved when all inputs match the sign of the corresponding weights: all positive weights should have the inputs of and all negative weights should have the inputs of . Just like the behavior of weights in regular neurons, the zero weights in the Rules Layer mean that the corresponding inputs will not have any effect on the output.
In order for the neuron in the Rules Layer to function as a proper logical AND operation, we need to apply a binary step activation function to its output:
(2) |
When applied at the Rules Layer, the binary step function defined in Equation 2 simply maps the range to 0, which ensures that the neuron is turned on only when Equation 1 evaluates to 1. With the dynamic bias and binary step function, each neuron in the Rules Layer encodes a rule that has predicates, where is the number of non-zero weights of that neuron. As discussed earlier, in effect neuron in the Rules Layer maps to a logical IF-THEN rule after training, where a positive input weight corresponds to a positive association of the input feature, a negative input weight corresponds to a negative association of the input feature, and a zero weight corresponds to an exclusion of the input feature.
However, as can be observed, the activations of the first layer are discretized into binary integers that are not naturally differentiable and the classic gradient computation approach doesn’t apply here. Therefore, we utilize the straight-through estimator discussed in (Bengio, Léonard, and Courville 2013) with the gradient clipping technique. Denoted by the binarized activation based on , we compute the gradient as follows:
(3) |
where and are the gradients of classification loss w.r.t. and , respectively. The condition simulates the backward computation of the ReLU function, which introduces non-linearity into the training process and empirically improves the performance; whereas our motivation of the second condition is to address the saturation effect: we suppress the update of the full-precision activations that are greater than and are still driven by the gradient to increase, since further raising activations does not produce any difference after binarization.
As discussed in above, the addition of the negative conditions in the input space is critical to the selection-based methods (Wang et al. 2017; Dash, Günlük, and Wei 2018) since they only consider the presence and absence of features and cannot deduce negative correlations unless they are explicitly provided in the input space. On the other hand, besides the presence of a positive association or an exclusion, our Rules Layer also learns the negation of an input feature by assigning a negative weight to it, and hence, DR-Net can directly derive negative conditions from the corresponding input features. Therefore, appending negative conditions in the input binary vector is redundant in DR-Net, and the input space of our DR-Net is reduced by half comparing with those selection-based method.
OR Layer
To produce the disjunction of the logical rules learned in the Rules Layer, the OR Layer contains only one output neuron, where the weights are binarized as follows:
(4) |
The output neuron performs a dot product with a negative bias as follows:
(5) |
where is a small value such that is positive when at least one input is activated. With a sigmoid activation function and a binary cross-entropy loss, this particular neuron behaves as an OR gate: the output is by default turned off because of the negative bias, while it produces a positive value if at least one rule is activated with a corresponding , which exactly mimics the behavior of the logical OR function. The binarized weights act as rule selectors that filter out rules that do not contribute to the model’s predictive performance. An example of our complete network structure is shown in Figure 1. We practically use in our implementations.

Sparsity-Based Regularization
The neural network structure proposed above outlines a way to derive a set of decision rules using stochastic gradient descent. As discussed above, a zero weight for a Rules Layer neuron corresponds to the exclusion of the corresponding input feature. Similarly, a zero weight for the OR Layer output neuron corresponds to the exclusion of the corresponding rule from the rule set. Thus, it should be clear that maximizing the sparsity of the Rules Layer neurons corresponds to simplifying the corresponding rules, and maximizing the sparsity of the OR Layer neuron corresponds to minimizing the number of rules.
However, to eliminate an input feature from a logical rule or a logical rule from the complete rule set, the corresponding weight has to be exactly zero, which is difficult to achieve in the typical network training process. To achieve a high degree of sparsity with exact zero weights, we explicitly incorporate a sparsity-based regularization mechanism into the training process using an approach akin to regularization by explicitly training mask variables.
As discussed in (Louizos, Welling, and Kingma 2017) as a way to achieve network sparsity through regularization, a binary random variable is attached to each weight of the model to indicate whether the corresponding weight is kept or removed. With this, we can reparameterize each weight as the product of a weight and the corresponding binary random variable :
(6) |
Assuming each is subject to a Bernoulli distribution with parameter , i.e. , the probability that is 1 is just . In (Louizos, Welling, and Kingma 2017), regularization is implemented by summing all parameters as the penalty term in the loss function111As explained below, we do not use normalization as the regularization term. Instead, we explicitly capture the model complexity with Equation 7 as the regularization term.. In order to train the binary random variables with stochastic gradient descent, two different gradient estimators have been proposed in (Louizos, Welling, and Kingma 2017) and (Li and Ji 2019), respectively, to approximate the Bernoulli distribution.
Applying the above regularization method to the Rules Layer is straightforward: all weight parameters are replaced by their product with the corresponding mask variables. For the OR Layer, since the weights will ultimately be binarized, we can just directly substitute the mask variables for the weights to simplify the process. That is, we can simply treat , with no need for a separate variable222Since is already binarized, there is no need to further binarize to derive with Equation 4. i.e., we can just use ..
We then incorporate a sparsity-based regularization term in the loss function to model the complexity of the rule set represented by the neural network. We denote by and the penalty of the non-zero mask variables of the Rules Layer and the OR Layer, respectively, where is the feature index, and is the index to the -th neuron (rule). Then the regularization loss is defined as follows:
(7) |
which explicitly captures the model complexity, as similarly defined in (Dash, Günlük, and Wei 2018). In particular, the model complexity of a rule set is defined as the sum of the number of rules and the total number of predicates in all rules. Following this definition, the first and the second terms of Equation 7 quantify the losses for the number of rules and the total number of predicates, respectively. Note that, according to the second term in Equation 7, the loss for the number of predicates in the -th rule will be effectively removed if is at or near zero: i.e., the -th neuron in the Rules Layer is disconnected from the OR Layer.
With the above sparsity-based regularization applied to DR-Net, the overall loss function we optimize for can be expressed as follows:
(8) |
where is the binary cross-entropy loss, is the regularization penalty that is specified by Equation 7, and is the regularization coefficient that balances the classification accuracy and rule set complexity.
Alternating Two-Phase Training Strategy
As previously discussed, each neuron in the first layer (the Rules Layer) of our proposed network architecture encodes an interpretable decision rule, whereas the output neuron in the second layer (the OR Layer) chooses some of the rules to be included in the set of decision rules. Empirically, we noticed that it is more effective to train our DR-Net with gradient-based optimizers (e.g., SGD) in an alternating manner, potentially due to the reduced search space and simpler optimization goals. In particular, our “alternating training strategy” consists of two training phases. We first freeze the OR Layer and only update the parameters in the Rules Layer to learn plausible rules. In the second phase, the Rules Layer is then fixed and we optimize the OR Layer such that redundant rules are eliminated while necessary inactive rules can also be re-enabled. The whole network is trained by alternating between the two training phases until convergence.
In addition, since the sparsity of the Rules Layer is directly related to the simplicity of rules, whereas the second layer is more focused on the selection of these derived rules, we further allow the flexible weighting of the sparsity-based regularization loss of the two layers. Specifically, as illustrated in Equation 8, the balance between classification loss and regularization loss is implemented via the regularization coefficient , where we can practically use different values for the two phases. In other words, the Rules Layer and the OR Layer are optimized over and , respectively, where and are defined as follows:
(9) |
In this way, the trade-off between model simplicity and accuracy in our experiments can be modulated by the adjustments of and .
Experimental Evaluation
The numerical experiments were evaluated on 4 publicly available binary classification datasets, which all have more than 10,000 instances and more than 10 attributes for each instance before binarization. The first two selected datasets are from UCI Machine Learning Repository (Dua and Graff 2017): MAGIC gamma telescope (magic) and adult census (adult), which are also used in recent works on rule set classifiers (Dash, Malioutov, and Varshney 2014; Wang et al. 2017; Dash, Günlük, and Wei 2018). The magic dataset is a dataset with pure numerical attributes while the adult dataset has a mix of both categorical and numerical attributes. The other two datasets are relatively recent datasets: the FICO HELOC dataset (heloc) and the home price prediction dataset (house), which have all numerical attributes. In all datasets, pre-processing is performed to encode categorical and numerical attributes into binary variables, as discussed earlier in the paper. Also, we append negative conditions for all other models except DR-Net.
Our goal is to learn a set of decision rules using our DR-Net and compare our model with other state-of-the-art rule learners and machine learning models. The results include model accuracies and complexities. Apart from the model complexity defined earlier (the number of rules plus the total number of conditions in the rule set), we also define the rule complexity, which is the average number of conditions in each rule of the model. We consider three other rule learners to directly compare with our work in terms of both accuracy and interpretability: the RIPPER algorithm (Cohen 1995), Bayesian Rule Sets (BRS) (Wang et al. 2017), and the Column Generation (CG) algorithm from (Dash, Günlük, and Wei 2018). The first one is an old rule set learning algorithm that is a variant of the Sequential Covering algorithm, while the other two are representatives of recent works in rule learning classifiers. We used open-source implementations on GitHub for all three algorithms, where the CG implementation (Arya et al. 2019) is slightly modified from the original paper. Other models used for comparison are the scikit-learn (Pedregosa et al. 2011) implementations of the decision tree learner CART (Breiman et al. 1984) and Random Forests (RF) (Breiman 2001). We also include a full-precision deep neural network (DNN) model with 6 layers, 50 neurons per hidden layer and ReLU activations. The last two models are uninterpretable models intended to provide baselines for typical performances that black-box models can achieve on these datasets. These uninterpretable baseline results serve as benchmarks for accuracy comparisons.
For DR-Net, we used the Adam optimizer with a fixed learning rate of and no weight decay across all experiments. There are 50 neurons in the Rules layer to ensure there is an efficient search space for all datasets. The alternating two-phase training strategy discussed earlier is employed with 10,000 total number of training epochs and 1,000 epochs for each layer. For simplicity, the batch size is fixed at 2,000 and the weights are uniformly initialized within the range between and . The parameters that are related to sparsity-based regularization are set the same as in the original paper (Louizos, Welling, and Kingma 2017).
Classification Performance
We evaluated the predictive performance of DR-Net by comparing both test accuracy and complexity with other state-of-the-art machine learning models. 5-fold nested cross validation was employed to select the parameters for all rule learners that explicitly trade-off between accuracy and interpretability to maximize the training set accuracies. To ensure that the final rule learner models are interpretable, we constrained the possible parameters for nested cross validation to a range that results in a low model complexity. For DR-Net, We fixed the to be in Equation 9 and only was varied in the experiment. Although there are many parameters in BRS to control the rule complexity, we followed the procedure used in (Dash, Günlük, and Wei 2018) and only varied the multiplier in prior hyper-parameter to save running time. For RIPPER, we varied the maximum number of conditions and the maximum number of rules as hyper-parameters of the implementation, which are directly related to the complexity of the model. The CG implementation in (Arya et al. 2019) doesn’t have the complexity bound parameter as specified in (Dash, Günlük, and Wei 2018) but instead provides two hyper-parameters to specify the costs of each clause and of each condition, which were used in our experiment to control the rule set complexity. We left all other parameters for these three algorithms (CG, BRS, RIPPER) as default. For CART and RF, we constrained the maximum depth of trees to be 100 for all datasets to achieve better generalization. For DNN, we used the same training parameters (number of epochs, batch size, learning rate, etc.) with a weight decay of . The test accuracy results of all models on all datasets are shown in Table 1 and the corresponding complexities are shown in Table 2. We omitted the results of the complexities of CART, RF and DNN because they have a different notion of model complexity and rule complexity.
dataset | magic | adult | heloc | house |
interpretable | ||||
DR-Net | 84.42 | 82.97 | 69.71 | 85.71 |
(0.53) | (0.51) | (1.05) | (0.40) | |
CG | 83.68 | 82.67 | 68.65 | 83.90 |
(0.87) | (0.48) | (3.48) | (0.18) | |
BRS | 81.44 | 79.35 | 69.42 | 83.04 |
(0.61) | (1.78) | (3.72) | (0.11) | |
RIPPER | 82.22 | 81.67 | 69.67 | 82.47 |
(0.51) | (1.05) | (2.09) | (1.84) | |
CART | 80.56 | 78.87 | 60.61 | 82.37 |
(0.86) | (0.12) | (2.83) | (0.29) | |
uninterpretable | ||||
RF | 86.47 | 82.64 | 70.30 | 88.70 |
(0.54) | (0.49) | (3.70) | (0.28) | |
DNN | 87.07 | 84.33 | 70.64 | 88.84 |
(0.71) | (0.42) | (3.37) | (0.26) |
It can be seen in Table 1 that our method outperforms other interpretable models on all datasets. For these better accuracy results, our method does not establish a similar superiority in the complexity comparison (Table 2). However, as shown Figure 2 and further discussed in the next section, our DR-Net approach can often achieve higher accuracy at comparable complexities. It is interesting to see that DR-Net maintains a relatively good model complexity compared with the corresponding rule complexity, which is exactly because our regularization loss function is designed specifically to minimize the model complexity instead of the rule complexity. Compared with RIPPER, which greedily mines good rules in each iteration to maximize the training accuracy, DR-Net is very competitive in the sense that it has similar or better test performance while consistently maintaining a lower model complexity. One advantage of the BRS algorithm over other models is that it consistently generates sparse models across all datasets, but at the expense of significantly inferior accuracies. The CART decision tree algorithm turned out to be the worst performing model in our experiments, which might result from overfitting. The results in Table 1 and Table 2 suggest that our DR-Net approach is very competitive as a machine learning model for interpretable classification. Finally, our DR-Net approach is able to achieve accuracies within only 3% of the uninterpretable models (RF and DNN) on the datasets evaluated.
dataset | magic | adult | heloc | house |
---|---|---|---|---|
DR-Net | 109.4 | 86.0 | 13.8 | 85.0 |
5.22 | 13.54 | 6.33 | 6.31 | |
CG | 112.8 | 120.0 | 3.4 | 28.6 |
3.72 | 3.77 | 1.90 | 5.15 | |
BRS | 40.0 | 16.8 | 16.6 | 31.2 |
3.00 | 3.00 | 2.96 | 3.00 | |
RIPPER | 189.4 | 117.6 | 72.8 | 328.0 |
6.01 | 4.66 | 5.24 | 7.01 |
Accuracy-Complexity Trade-off




In this experiment, we compared the accuracy-complexity trade-off of our DR-Net with other rule learning algorithms: CG, BRS and RIPPER. The parameters that were selected to be varied in this experiment are the same as ones in the first experiment. Instead of using nested cross validation to select best parameters on the validation set, we manually picked a set of values for each selected parameters for each algorithm to generate different sets of accuracy-complexity pairs. We ran the experiments on all datasets and the results with the average of the 5-fold cross validation are shown in Figure 2. Apart from model complexity and rule complexity, we included a third metric to show the average number of rules in each generated rule set versus the test accuracy. For each method compared, the dots connected by the line segments shown correspond to Pareto efficient models where all other points below the Pareto frontier have either lower accuracies or higher complexities.
The characteristic of being able to attain a high test accuracy with an acceptable model complexity for DR-Net in Table 1 and Table 2 is carried over to Figure 2. For the magic, adult and house datasets, DR-Net outperforms all other rule learners in terms of the accuracy by a substantial margin when the model complexity, the rule complexity or the number of rules exceeds a certain threshold. Although DR-Net does not dominate RIPPER on the heloc dataset, their accuracy comparison is very close if enough model complexity or number of rules is given. The only thing that DR-Net falls behind a little bit is in the rule complexity vs. accuracy comparison on the heloc dataset. In theory, DR-Net can achieve relatively low rule complexity with a different regularization loss function that can quantify the average number of conditions in the rule set, which we leave as future work. It is also interesting to note that the number of rules from DR-Net varies in a relatively narrower range compared with other approaches as shown in the third column of Figure 2, which is directly resulted by fixing in Equation 9. BRS does not demonstrate a clear accuracy-complexity trade-off as its results all group in a very narrow range, which is also noted and explained in (Dash, Günlük, and Wei 2018). This experiment shows that DR-Net can be preferred over other rule learners because of its potential for achieving a much higher test accuracy with a relatively moderate complexity sacrifice.
Related Work
The learning of Boolean rules and rule sets is well studied with different variants. While the learning of two-level Boolean decision rule set has an extensive history in different communities, most of them employ heuristic algorithms that optimize for certain criteria that are not directly related to classification accuracy or model simplicity. Representatives of these methods include logical analysis of data (Crama, Hammer, and Ibaraki 1988; Boros et al. 2000), association rule mining and classification (Clark and Niblett 1989; Liu, Hsu, and Ma 1998), and greedy set covering (Cohen 1995).
With the increasing interest in the field of explainable machine learning, researchers have in recent years added model complexity to the optimization objective so that accuracy and simplicity can be jointly optimized. Several approaches select rules from a pre-mined set of candidate rules (Wang et al. 2017; Lakkaraju, Bach, and Leskovec 2016). A Bayesian framework is presented in (Wang et al. 2017) for selecting pre-mined rules by approximately constructing a maximum a posteriori (MAP) solution. In (Lakkaraju, Bach, and Leskovec 2016), the joint optimization problem is approximately solved by a local search algorithm. In these methods, the requirement for rules pre-mining limits the overall search space, hindering their ability to find a globally optimized model. Other approaches based on integer-programming (IP) formulations (Su et al. 2015; Dash, Günlük, and Wei 2018) do not require rules pre-mining, but they rely on approximate solutions for large datasets. In (Dash, Günlük, and Wei 2018), the IP problem is approximately solved by relaxing it into a linear programming problem and applying the column generation algorithm, whereas (Su et al. 2015) utilizes various optimization approaches including block coordinate descent and alternating minimization algorithm.
Besides decision rule sets, decision lists (Rivest 1987; Bertsimas, Chang, and Rudin 2012; Letham et al. 2015) and decision trees (Breiman et al. 1984; Rokach and Maimon 2005) are also interpretable rule-based models. In decision lists, rules are ordered in an IF-THEN-ELSE sequence. However, the chaining of rules via an IF-THEN-ELSE sequence means that the interpretation of an activated rule requires an understanding of all preceding rules. This can make the explanation more difficult for humans to understand. In decision trees, rules are organized into a tree structure. However, they are often prone to overfitting.
Conclusion and Extensions
In this paper, we presented a simple two-layer neural network architecture, which can be directly mapped to a set of interpretable decision rules, along with a procedure to accurately train the network for classification. We described a sparsity-based regularization approach that can capture the complexity of the trained model in terms of the length of the rules and the number of rules. The incorporation of this regularization loss into the overall loss function enables the training process to balance between classification accuracy and model complexity. With our neural net formulation, we are able to leverage state-of-the-art neural net infrastructures to learn highly accurate and interpretable rule-based models. Our experimental results show that our method can generate more accurate decision rule sets than other state-of-the-art rule-learners with better accuracy-simplicity trade-offs. When compared with uninterpretable black box models such as random forests and full-precision deep neural networks, our approach can easily learn interpretable models that have comparable predictive performance.
We focus in this paper on the binary classification problem, but the approach can be easily extended to multi-class classification by deploying separate output neurons for each class and mapping each output neuron to a corresponding set of rules for the respective class. A default class and a tie-breaking function could be used in the event that no class or more than one class is activated, respectively (Lakkaraju, Bach, and Leskovec 2016), or these cases can be handled by error correcting output codes (Schapire 1997). We plan to investigate in future work potentially more powerful tie-breaking mechanisms that can be directly trained as part of the neural net formulation, for example by directly interpreting softmax results.
Acknowledgements
This research has been supported in part by the National Science Foundation (NSF IIS-1956339).
References
- Abadi et al. (2016) Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G. S.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Goodfellow, I.; Harp, A.; Irving, G.; Isard, M.; Jia, Y.; Jozefowicz, R.; Kaiser, L.; Kudlur, M.; Levenberg, J.; Mane, D.; Monga, R.; Moore, S.; Murray, D.; Olah, C.; Schuster, M.; Shlens, J.; Steiner, B.; Sutskever, I.; Talwar, K.; Tucker, P.; Vanhoucke, V.; Vasudevan, V.; Viegas, F.; Vinyals, O.; Warden, P.; Wattenberg, M.; Wicke, M.; Yu, Y.; and Zheng, X. 2016. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems.
- Arya et al. (2019) Arya, V.; Bellamy, R. K. E.; Chen, P.-Y.; Dhurandhar, A.; Hind, M.; Hoffman, S. C.; Houde, S.; Liao, Q. V.; Luss, R.; Mojsilović, A.; Mourad, S.; Pedemonte, P.; Raghavendra, R.; Richards, J.; Sattigeri, P.; Shanmugam, K.; Singh, M.; Varshney, K. R.; Wei, D.; and Zhang, Y. 2019. One Explanation Does Not Fit All: A Toolkit and Taxonomy of AI Explainability Techniques. URL https://arxiv.org/abs/1909.03012.
- Bengio, Léonard, and Courville (2013) Bengio, Y.; Léonard, N.; and Courville, A. 2013. Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation.
- Bertsimas, Chang, and Rudin (2012) Bertsimas, D.; Chang, A.; and Rudin, C. 2012. ORC: Ordered Rules for Classification A Discrete Optimization Approach to Associative Classification .
- Boros et al. (2000) Boros, E.; Hammer, P. L.; Ibaraki, T.; Kogan, A.; Mayoraz, E.; and Muchnik, I. 2000. An implementation of logical analysis of data. IEEE Transactions on Knowledge and Data Engineering 12(2): 292–306.
- Breiman (2001) Breiman, L. 2001. Random Forests. Mach. Learn. 45(1): 5–32. ISSN 0885-6125. doi:10.1023/A:1010933404324. URL https://doi.org/10.1023/A:1010933404324.
- Breiman et al. (1984) Breiman, L.; Friedman, J.; Stone, C.; and Olshen, R. 1984. Classification and Regression Trees. The Wadsworth and Brooks-Cole statistics-probability series. Taylor & Francis. ISBN 9780412048418. URL https://books.google.com/books?id=JwQx-WOmSyQC.
- Clark and Niblett (1989) Clark, P.; and Niblett, T. 1989. The CN2 Induction Algorithm. Mach. Learn. 3(4): 261–283. ISSN 0885-6125. doi:10.1023/A:1022641700528. URL https://doi.org/10.1023/A:1022641700528.
- Cohen (1995) Cohen, W. W. 1995. Fast Effective Rule Induction. In In Proceedings of the Twelfth International Conference on Machine Learning, 115–123. Morgan Kaufmann.
- Crama, Hammer, and Ibaraki (1988) Crama, Y.; Hammer, P. L.; and Ibaraki, T. 1988. Cause-Effect Relationships and Partially Defined Boolean Functions. Ann. Oper. Res. 16(1–4): 299–325. ISSN 0254-5330. doi:10.1007/BF02283750. URL https://doi.org/10.1007/BF02283750.
- Dash, Günlük, and Wei (2018) Dash, S.; Günlük, O.; and Wei, D. 2018. Boolean Decision Rules via Column Generation. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS’18, 4660–4670. Red Hook, NY, USA: Curran Associates Inc.
- Dash, Malioutov, and Varshney (2014) Dash, S.; Malioutov, D. M.; and Varshney, K. R. 2014. Screening for learning classification rules via Boolean compressed sensing. In 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 3360–3364.
- Dua and Graff (2017) Dua, D.; and Graff, C. 2017. UCI Machine Learning Repository. URL http://archive.ics.uci.edu/ml.
- Konečný et al. (2017) Konečný, J.; McMahan, H. B.; Yu, F. X.; Richtárik, P.; Suresh, A. T.; and Bacon, D. 2017. Federated Learning: Strategies for Improving Communication Efficiency.
- Lakkaraju, Bach, and Leskovec (2016) Lakkaraju, H.; Bach, S. H.; and Leskovec, J. 2016. Interpretable Decision Sets: A Joint Framework for Description and Prediction. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, 1675–1684. New York, NY, USA: Association for Computing Machinery. ISBN 9781450342322. doi:10.1145/2939672.2939874. URL https://doi.org/10.1145/2939672.2939874.
- Letham et al. (2015) Letham, B.; Rudin, C.; McCormick, T. H.; and Madigan, D. 2015. Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model.
- Li and Ji (2019) Li, Y.; and Ji, S. 2019. -ARM: Network Sparsification via Stochastic Binary Optimization.
- Liu, Hsu, and Ma (1998) Liu, B.; Hsu, W.; and Ma, Y. 1998. Integrating Classification and Association Rule Mining. In Proceedings of the Fourth International Conference on Knowledge Discovery and Data Mining, KDD’98, 80–86. AAAI Press.
- Louizos, Welling, and Kingma (2017) Louizos, C.; Welling, M.; and Kingma, D. P. 2017. Learning Sparse Neural Networks through Regularization.
- Paszke et al. (2017) Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; and Lerer, A. 2017. Automatic differentiation in PyTorch. In NIPS-W.
- Pedregosa et al. (2011) Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; Vanderplas, J.; Passos, A.; Cournapeau, D.; Brucher, M.; Perrot, M.; and Duchesnay, E. 2011. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 12: 2825–2830.
- Rivest (1987) Rivest, R. L. 1987. Learning Decision Lists. Mach. Learn. 2(3): 229–246. ISSN 0885-6125. doi:10.1023/A:1022607331053. URL https://doi.org/10.1023/A:1022607331053.
- Rokach and Maimon (2005) Rokach, L.; and Maimon, O. 2005. Top–Down Induction of Decision Trees Classifiers–A survey. Systems, Man, and Cybernetics, Part C: Applications and Reviews, IEEE Transactions on 487.
- Rudin (2019) Rudin, C. 2019. Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. Nature Machine Intelligence 1: 206–215.
- Schapire (1997) Schapire, R. E. 1997. Using Output Codes to Boost Multiclass Learning Problems. In Proceedings of the Fourteenth International Conference on Machine Learning, ICML ’97, 313–321. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc. ISBN 1558604863.
- Su et al. (2015) Su, G.; Wei, D.; Varshney, K. R.; and Malioutov, D. M. 2015. Interpretable Two-level Boolean Rule Learning for Classification.
- Wang et al. (2017) Wang, T.; Rudin, C.; Doshi-Velez, F.; Liu, Y.; Klampfl, E.; and MacNeille, P. 2017. A Bayesian Framework for Learning Rule Sets for Interpretable Classification. Journal of Machine Learning Research 18(70): 1–37. URL http://jmlr.org/papers/v18/16-003.html.