This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Learning Performance Maximizing Ensembles with Explainability Guarantees

Vincent Pisztora, Jia Li
Abstract

In this paper we propose a method for the optimal allocation of observations between an intrinsically explainable glass box model and a black box model. An optimal allocation being defined as one which, for any given explainability level (i.e. the proportion of observations for which the explainable model is the prediction function), maximizes the performance of the ensemble on the underlying task, and maximizes performance of the explainable model on the observations allocated to it, subject to the maximal ensemble performance condition. The proposed method is shown to produce such explainability optimal allocations on a benchmark suite of tabular datasets across a variety of explainable and black box model types. These learned allocations are found to consistently maintain ensemble performance at very high explainability levels (explaining 74%74\% of observations on average), and in some cases even outperform both the component explainable and black box models while improving explainability.

Introduction

In most high stakes settings, such as medical diagnosis (Gulum, Trombley, and Kantardzic 2021) and criminal justice (Rudin 2019), model predictions have two viability requirements. Firstly, they must exceed a given global performance threshold, thus ensuring an adequate understanding of the underlying process. Secondly, model predictions must be explainable.

Explainability, however defined (Linardatos, Papastefanopoulos, and Kotsiantis 2020), is a desirable characteristic in any prediction function. Intrinsically interpretable “glass box” models ((Agarwal et al. 2021), (Lemhadri, Ruan, and Tibshirani 2021), (Rymarczyk et al. 2020)), which are explainable by construction, are particularly advantageous as they require no additional post-hoc processing ((Ribeiro, Singh, and Guestrin 2016), (Lundberg and Lee 2017)) to achieve explainability, and thus also avoid complications arising from post-hoc explanation learning ((Rudin 2019), (Garreau and Luxburg 2020)). Due to these advantages, glass box models are uniquely suited to settings where faithful explanations of predictions are required.

However, using an approach of “complete explainability”, in which a glass box model is used as the prediction function across the entire feature space, may not be viable. It may be the case that, in a given setting, no glass box exists that can adequately model the relationship of interest in all regions of the feature space. Thus in some regions, the model’s predictions will fail to exceed the performance threshold required by the use-case. If, as a consequence, the model exceeds the application’s global error tolerance (e.g. a low accuracy in stroke prediction (Gage et al. 2001)), it may not be usable in practice.

An alternative, “partial explainability” approach requires instead that only a proportion of observations be provided intrinsically explainable predictions. We will refer to this proportion, which is the proportion of observations for which the explainable model is the prediction function, as the explainability level qq. Such approaches, including our proposed method, Ensembles with Explainability Guarantees (EEG), can provide high performance while maximizing explainability, and work especially well in cases where the explainable model can be paired with an alternate model with complementary strengths. As demonstrated in Fig. 1, by identifying the areas of expertise of the glass box and black box models, the EEG approach can allocate predictions accordingly to improve both performance and explainability.

Refer to caption
Figure 1: This figure shows a two-class classification task in which the areas of expertise (the diamond pattern for the glass box and the spiral pattern for the black box model) are complementary. In this example, the glass box achieves a 92.7%92.7\% accuracy, the black box reaches 95.0%95.0\% accuracy, and the allocated ensemble of the two exceeds both with a 95.8%95.8\% accuracy. Thus, the resulting EEG allocation improves performance over both component models while also providing explainability (for 20%20\% of observations in this case).

Generally, implementations of a partial explainability approach consist of an ensemble of models including at least one explainable model and alternate model (often a black box model), and an allocation scheme by which observations are distributed among the ensemble members for prediction. Individual methods are characterized by their heterogeneity in the following aspects.

Methods vary in the range of component models they can accommodate. Some are defined for only one set of glass box, black box, and allocator model types - for example LSP (Wang and Saligrama 2012) and OTSAM (Wang, Fujimaki, and Motohashi 2015), which use binary tree-type splitting to define regions, and linear models and sparse additive models respectively to predict within regions. Other methods are black box agnostic but still limited in glass box and allocator model type - for example HyRS (Wang 2019), HyPM (Wang and Lin 2021), CRL (Pan, Wang, and Hara 2020), and HybridCORELS (Ferry, Laberge, and Aïvodji 2023), which use rule-based models as both glass box and allocator. EEG is the only fully model-agnostic partial explainability method which can be implemented with any combination of glass box, black box, and allocator models.

Methods also vary in the approach used to learn each ensemble member model (i.e. glass box and black box). Most methods first learn the black box model globally (on the full dataset), and then learn the glass box model locally (on its allocated subset of the data), either simultaneously with the allocator (HyRS, HyPM, CRL, and HybridCORELS) or in an alternating EM-style (LSP, OTSAM, and AdaBudg (Nan and Saligrama 2017)). EEG on the other hand, learns both ensemble member models globally first before learning the allocations between them - similar to most general adaptive ensembling methods, e.g. (Gao et al. 2019), (Inoue 2019).

Finally, methods are characterized by their allocation criteria which commonly consist of an objective which combines one or more of the following - the explainability level, the underlying task performance of the ensemble, and the complexity of the glass box model. Most methods optimize a measure of post-allocation ensemble performance - LSP, HyRS, HyPM, and HybridCORELS minimize a 0/1 misclassification loss, AdaBudg uses a more flexible logistic loss, and CRL maximizes accuracy across a range of explainability levels. Several methods with rule-based glass box/allocator hybrid models (HyRS, HyPM, CRL, and HybridCORELS) also include a penalty on the complexity of these models. To control the explainability level, methods either include a reward term in the loss (HyRS, HyPM, and CRL), or directly restrict the model space to candidates which achieve the explainability level (HybridCORELS). In contrast, EEG optimizes an MSE loss between the predicted and actual “glass box allocation desirability” percentile of each observation.

More extensive reviews of the partial explainability approach and explainability methods in general are available in (Linardatos, Papastefanopoulos, and Kotsiantis 2020), (Nauta et al. 2022), and (Sahakyan, Aung, and Rahwan 2021).

As outlined above, our proposed method, Ensembles with Explainability Guarantees (EEG), differs from existing works in its approach to the partial explainability problem. The key novelties of this new approach, and their corresponding advantages are summarized below.

Independent and Global Component Models: The first key innovation of the EEG approach is the independent learning of each component model (i.e. the ensemble member models and allocator). As a result, EEG is agnostic to task, data, and component model type. Thus, the most powerful models can be used for each component as determined by the setting - in contrast with previous works which are more restricted.

Another important consequence of separate component model learning is that glass box predictions are independently explainable in the global context, and thus immune from “explainability collapse” - a scenario in which the allocator subsumes the glass box’s prediction role, diminishing the value of the explainable prediction, in the extreme case reducing the glass box to an uninformative constant function. On the other hand, methods which either learn glass box models locally, or jointly with the allocator, are vulnerable to this type of degeneration.

Allocation Desirability Ranking: The second novel aspect of the EEG approach is the concept of allocation desirability. Given an ensemble of models, allocation desirability quantifies how beneficial it is for a given observation to be allocated to the default ensemble member model, say the glass box. Thus, it induces a preference for glass box allocation between all pairs of observations and consequently also defines a ranking of allocation preference across all observations that is optimal irrespective of the desired explainability level.

A key advantage of such a ranking is that it is independent of the training criteria of the ensemble member models, and thus can be adapted to score allocation desirability using metrics that best fit the setting. Indeed, the EEG desirability metric builds a ranking using a combination of relative sufficient performance and absolute performance measures which can natively accommodate any underlying problem type (e.g. regression, classification). This particular desirability metric also offers several additional benefits including allocation desirability percentile and sufficiency category estimates for each observation.

Q-Complete Allocation Optimality: The final key point of novelty of the EEG approach is the optimality of allocation, as defined in Proposition 1 and Proposition 2, which is encoded in the allocation desirability ranking for any explainability level. Thus, the learned allocator, which estimates this ranking, is an explicit function of qq and provides the allocation solution to any explainability level after training only once. This capability is in contrast with previous works which provide, at most, several explainability level solutions with varying degrees of stability (Ferry, Laberge, and Aïvodji 2023).

These unique capabilities of the EEG method enable the following practical use cases:

  • Given a minimum performance requirement on the underlying task, the method can be used to obtain the allocation with the highest explainability level that achieves or exceeds the performance threshold.

  • Given a minimum explainability level requirement, the method can be used to obtain the allocation with the highest ensemble performance which meets or exceeds the required explainability level.

  • Given a minimal level of post-allocation glass box-specific performance, the allocation that achieves the highest explainability level while meeting or exceeding this requirement can be found.

  • Given a set of observations, sufficiency category estimates can be obtained for each, identifying which observations are likely to yield incorrect decisions and describing the likely failure mode for each such case to inform potential post-hoc remedies.

In the following sections, we first describe our method in detail and provide some theoretical assurances on the characteristics of the resulting allocator in the Methods section. Then, in the Experiments section, we describe the experimental settings and the estimation of the allocator, and demonstrate the method’s favorable performance.

Methodology

Setting

First, we define the underlying task as the estimation of the function f(x)=yf(x)=y where x𝒳,y𝒴x\in\mathcal{X},y\in\mathcal{Y}. We also define observations as z=(x,y)𝒳×𝒴=𝒵z=(x,y)\in\mathcal{X}\times\mathcal{Y}=\mathcal{Z}, the training dataset Dn={zi:i{1,,n},zi𝒵}D^{n}=\{z_{i}:i\in\{1,...,n\},z_{i}\in\mathcal{Z}\}, and loss function for the underlying task lU:𝒴×𝒴l_{U}:\mathcal{Y}\times\mathcal{Y}\rightarrow\mathbb{R}. Next, we define the ensemble component models - first, the intrinsically explainable glass box model as g:𝒳𝒴g:\mathcal{X}\rightarrow\mathcal{Y} and the alternate, black box model as b:𝒳𝒴b:\mathcal{X}\rightarrow\mathcal{Y}, both of which are learned independently on the full training dataset DnD^{n}.

Next, we define the allocation task. We define the class of all allocator functions as A={a:𝒱{0,1}}A=\{a:\mathcal{V}\rightarrow\{0,1\}\} and the class of all “proper” allocator functions as A^={a:𝒱^{0,1}}\hat{A}=\{a:\hat{\mathcal{V}}\rightarrow\{0,1\}\}, where 𝒱\mathcal{V} is a general space of inputs (typically 𝒵\mathcal{Z}) and 𝒱^𝒱\hat{\mathcal{V}}\subseteq\mathcal{V} containing only information available at allocation time. Next we define the class of qq-explainable allocators as Aq={aq:aqA,1ni=1naq(vi)=q}A_{q}=\{a_{q}:a_{q}\in A,\frac{1}{n}\sum^{n}_{i=1}a_{q}(v_{i})=q\} and the corresponding class of “proper” qq-explainable allocators as A^q={aq:aqA^,1ni=1naq(vi)=q}\hat{A}_{q}=\{a_{q}:a_{q}\in\hat{A},\frac{1}{n}\sum^{n}_{i=1}a_{q}(v_{i})=q\}, for q𝒬={in:i{1,,n}}[0,1]q\in\mathcal{Q}=\{\frac{i}{n}:i\in\{1,...,n\}\}\subseteq[0,1], with qq being the explainability level. Note, the set AA is used to define the optimal allocator, whereas the set A^\hat{A} is searched to obtain an estimator of this optimum.

We next define indicators of performance sufficiency. These functions s:𝒵{0,1}s:\mathcal{Z}\rightarrow\{0,1\} should be thought of as context-dependent indicators of whether performance within a region of the feature space is sufficiently high to use the model in question reliably for explanation. For classification tasks, we define performance sufficiency as sf(z)=𝕀{f(x)=y}s^{\prime}_{f}(z)=\mathbb{I}{\{f(x)=y\}}, and for regression tasks as sf(z)=𝕀{lU(f(x),y)<ϵ}s^{\prime}_{f}(z)=\mathbb{I}{\{l_{U}(f(x),y)<\epsilon\}}, with f:𝒳𝒴f:\mathcal{X}\rightarrow\mathcal{Y}. In practice ϵ\epsilon should be selected based on problem-specific context, however, lacking such context in the regression experiments conducted for this study, ϵ\epsilon was selected to be the lower of the average validation losses of gg and bb, as a reasonable threshold for prediction correctness. These sufficiency indicators generate the following partition of the data: Z0={z:z𝒵,sg(z)+sb(z)=0},Z2={z:z𝒵,sg(z)+sb(z)=2},Zg={z:z𝒵,sg(z)=1,sb(z)=0}Z^{\prime}_{0}=\{z:z\in\mathcal{Z},s^{\prime}_{g}(z)+s^{\prime}_{b}(z)=0\},Z^{\prime}_{2}=\{z:z\in\mathcal{Z},s^{\prime}_{g}(z)+s^{\prime}_{b}(z)=2\},Z^{\prime}_{g}=\{z:z\in\mathcal{Z},s^{\prime}_{g}(z)=1,s^{\prime}_{b}(z)=0\}, and Zb={z:z𝒵,sg(z)=0,sb(z)=1}Z^{\prime}_{b}=\{z:z\in\mathcal{Z},s^{\prime}_{g}(z)=0,s^{\prime}_{b}(z)=1\}, with n0=|Z0|,n2=|Z2|,ng=|Zg|,nb=|Zb|n_{0}=|Z^{\prime}_{0}|,n_{2}=|Z^{\prime}_{2}|,n_{g}=|Z^{\prime}_{g}|,n_{b}=|Z^{\prime}_{b}|, and nq=nqn_{q}=nq.

Next we motivate the use of the sufficiency perspective. Sufficiency functions are critical for defining coherent allocations when, as is often the case, the absolute performance measures used to learn ensemble component models do not match allocation preference (e.g. loss minimization vs accuracy maximization). Consider the following constrained allocation decision in which only one observation can be allocated to gg:

Obs lU(g)l_{U}(g) sgs_{g} lU(b)l_{U}(b) sbs_{b}
z1z_{1} 0 1 2 1
z2z_{2} 3 1 4 0

In this case, loss minimization dictates an allocation of z1z_{1} to gg and z2z_{2} to bb, which would allocate z2z_{2} to an insufficient prediction. Sufficiency maximization would however yield a more satisfactory allocation of z2z_{2} to gg and z1z_{1} to bb. This example demonstrates the utility of sufficiency allocation - distinguishing between a case where the user is willing to sacrifice “a bit of performance” (as quantified by sufficiency) for explanation (z1z_{1}), and a case were even a small performance drop results in an explanation that is not sufficiently trustworthy to use (z2z_{2}).

In the next section, we define the objective of the allocation task and introduce our proposed approach for addressing it.

Optimal Allocation

In the allocation task, the objective is to construct an allocator aqa_{q} that will determine which model, either the explainable gg or the black box bb, is used for prediction on any given observation zz, in a manner that is optimal relative to the following criteria. Firstly, for any given explainability level qq, the allocator should distribute observations in a way that maximizes sufficient ensemble performance, defined as t¯(aq)=1ni=1nsg(zi)aq(vi)+sb(zi)(1aq(vi))\bar{t}^{\prime}(a_{q})=\frac{1}{n}\sum_{i=1}^{n}s_{g}^{\prime}(z_{i})a_{q}(v_{i})+s_{b}^{\prime}(z_{i})(1-a_{q}(v_{i})). Secondly, and again for any qq, the allocator should maximize sufficient explainable prediction t¯g(aq)=1ni=1nsg(zi)aq(vi)\bar{t}^{\prime}_{g}(a_{q})=\frac{1}{n}\sum_{i=1}^{n}s_{g}^{\prime}(z_{i})a_{q}(v_{i}), i.e. the performance of the model gg on the subset of observations it has been allocated, subject to maintaining maximal t¯(aq)\bar{t}^{\prime}(a_{q}). Finally, the allocator should also be consistent in its allocations across the values of qq, meaning that if an observation is allocated to gg for a given qq, it should remain allocated to the glass box for all higher explainability levels as well. Next, we define our allocator, and show that it meets the three criteria introduced above.

Our proposed allocation function is defined as aq(z)=𝕀{r(z)>1q}a^{\prime}_{q}(z)=\mathbb{I}\{r^{\prime}(z)>1-q\}, where rescaled ranking r(z)=rankDn(r~(z))nr^{\prime}(z)=\frac{\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z))}{n}, and ranking r~(z)=2sg(z)sb(z)σ(lU(b(x),y)lU(g(x),y))\tilde{r}^{\prime}(z)=2s^{\prime}_{g}(z)-s^{\prime}_{b}(z)-\sigma(l_{U}(b(x),y)-l_{U}(g(x),y)), with σ(x)=11+ex\sigma(x)=\frac{1}{1+e^{-x}}.

The intuition behind the allocator is as follows. First, all observations are sorted in sufficient performance maximizing order, i.e. allocation of observations in ZgZ_{g}^{\prime} to gg is prioritized over allocation of observations in Z2Z_{2}^{\prime} and Z0Z_{0}^{\prime}, which in turn are prioritized over ZbZ_{b}^{\prime}. Next, observations are sorted in explainable sufficient performance maximizing order, i.e. Z2Z_{2}^{\prime} is prioritized ahead of Z0Z_{0}^{\prime} for allocation to gg. Then, within each sufficiency category, observations are ordered in absolute performance maximizing order, i.e. observations with large relative performance of gg over bb are prioritized for allocation to gg. Next, this ranking is normalized, yielding the glass box allocation desirability percentile rr^{\prime}. An important feature of this percentile is that it is constant with respect to qq, thus the optimal observations to allocate to gg, for any level of qq, are simply the nqn_{q} most highly ranked, resulting in the allocator aqa^{\prime}_{q}.

Note that in the described methodology, sufficiency based allocation can be viewed as a generalization of allocation via absolute performance, and can thus be reduced to the latter by selecting either sf(z)=0s_{f}(z)=0 or sf(z)=1s_{f}(z)=1, z,f\forall z,f.

Next, we state the optimality properties of the proposed allocator aqa^{\prime}_{q}. The proofs are available in the long form paper on arxiv.org in the Theoretical Results section of the Appendix.

Proposition 1.

(Maximal Sufficient Performance) q𝒬,aqAq\forall q\in\mathcal{Q},a^{\prime}_{q}\in A^{*}_{q} where Aq={aq:aq=argmaxaqAqt¯(aq)}A^{*}_{q}=\{a^{*}_{q}\colon a^{*}_{q}=\arg\max_{a_{q}\in A_{q}}\bar{t}^{\prime}(a_{q})\} and t¯(aq)=1ni=1nt(aq,zi)=1ni=1nsg(zi)aq(zi)+sb(zi)(1aq(zi))\bar{t}^{\prime}(a_{q})=\frac{1}{n}\sum_{i=1}^{n}t^{\prime}(a_{q},z_{i})=\frac{1}{n}\sum_{i=1}^{n}s^{\prime}_{g}(z_{i})a_{q}(z_{i})+s^{\prime}_{b}(z_{i})(1-a_{q}(z_{i}))

Proposition 2.

(Maximal Sufficient Explainable Performance) q𝒬,aqAq|g\forall q\in\mathcal{Q},a^{\prime}_{q}\in A^{*}_{q|g} where Aq|g={aq|g:aq|g=argmaxaqAqt¯g(aq)}A^{*}_{q|g}=\{a^{*}_{q|g}\colon a^{*}_{q|g}=\arg\max_{a^{*}_{q}\in A^{*}_{q}}\bar{t}^{\prime}_{g}(a^{*}_{q})\} and t¯g(aq)=1ni=1ntg(aq,zi)=1ni=1nsg(zi)aq(zi)\bar{t}^{\prime}_{g}(a^{*}_{q})=\frac{1}{n}\sum_{i=1}^{n}t^{\prime}_{g}(a^{*}_{q},z_{i})=\frac{1}{n}\sum_{i=1}^{n}s^{\prime}_{g}(z_{i})a^{*}_{q}(z_{i})

Proposition 3.

(Monotone Allocation) qi<qj𝒬\forall q_{i}<q_{j}\in\mathcal{Q}, {z:z𝒵,aqi(z)=1}{z:z𝒵,aqj(z)=1}\{z:z\in\mathcal{Z},a^{\prime}_{q_{i}}(z)=1\}\subseteq\{z:z\in\mathcal{Z},a^{\prime}_{q_{j}}(z)=1\}

Experiments

In this section we describe the data, model training procedures, performance evaluation metrics, and results of our experiments.

Datasets

Tabular data is used to evaluate the proposed methodology as it the setting for which the required intrinsically explainable glass box models are most readily available. Following the tabular data benchmarking framework proposed by (Grinsztajn, Oyallon, and Varoquaux 2022), we conduct experiments on a set of 31 datasets (13 classification, 18 regression). These datasets represent the full set of provided datasets with quantitative features less the four largest scale datasets (omitted due to computational limitations). These datasets are summarized in Table 4.

Each dataset is split (70%, 9%, 21%) into training, validation, and test sets respectively, following (Grinsztajn, Oyallon, and Varoquaux 2022). All features and regression response variables are rescaled to the range [-1,1].

Models

Both glass box and black box models are learned on the full training dataset for each underlying task. For classification datasets, two types of glass box model are fitted, a logistic regression and a classification tree, as well as two types of black box model, a gradient boosting trees classifier and a neural network classifier. Analogously, for regression datasets, two types of glass box model are fitted, a linear regression and a regression tree, as well as two types of black box model, a gradient boosting trees regressor and a neural network regressor. In all cases, the architecture of the neural networks is the “Wide ResNet-28” model (Zagoruyko and Komodakis 2016) adapted to tabular data with the replacement of convolutional layers with fully connected layers.

An allocator is subsequently also learned on the full training dataset. Both gradient boosting trees regressors and neural networks are fitted as allocators for each allocation task.
For allocator training, the features xx are augmented with four additional constructed features, the predictions g(x)g(x) and b(x)b(x), and two distance measures d(g(x),b(x))d(g(x),b(x)) between them, the cross-entropy and MSE. In our experiments, inclusion of these features improved allocator learning - likely by removing the need for the allocator to attempt to learn these quantities on its own. Allocation performance is further improved by ensembling the feature-dependent learned allocator aqa^{\prime}_{q} with a strong feature-independent allocator aq′′a^{\prime\prime}_{q}, where aq′′(d(g(z),b(z)))=𝕀{rankDnd(g(z),b(z))n>1q}a^{\prime\prime}_{q}(d(g(z),b(z)))=\mathbb{I}\{\frac{\mathrm{rank}_{D^{n}}d(g(z),b(z))}{n}>1-q\}. aq′′a^{\prime\prime}_{q} can be viewed as an “assume the black box is correct” allocation rule which is more likely to assign an observation to gg if the distance between the predictions of gg and bb is high. Which of the two allocators is used for a given qq is determined by their respective performances on the validation set.

Hyperparameter Tuning

Hyperparameter tuning for all models is done using 4-fold cross-validation, with the exception of the neural network tuning which is done using the validation set. A grid search is done to select the best hyperparameters for each model with search values available in Tables 5, 6, and 7.
Each glass box and black box model is tuned on the full set of hyperparameters each time it is replicated. The gradient boosting trees allocator models are retuned on the full hyperparameter set each time as well. The neural network allocator is not retuned however, and instead uses the optimal settings found in the fitting of the black box on each dataset.

Metrics

We define the following metrics which are used to measure performance of our method. First we define the Percentage Performance Captured over Random (PPCR) for a given allocator as follows: PPCR(aq)=AUC(aq)AUC(rq)AUC(oq)AUC(rq)PPCR(a_{q})=\frac{AUC(a_{q})-AUC(r_{q})}{AUC(o_{q})-AUC(r_{q})} where AUC(fq)AUC(f_{q}) is the area under the curve of function fqf_{q} over all values of qq in its domain, oqo_{q} is the oracle allocator which has perfect information on the whole dataset, and rqr_{q} is the random allocator which selects a subset from the data being allocated uniformly at random. The PPCR metric is a percentage and represents the proportion of the oracle AUC, in excess of that also covered by random allocation, that the learned allocator is able to capture. Thus a value of zero indicates performance on par with rqr_{q} and a value of one represents perfect allocation.

Next, we define the Percent Q Equal or Over Max (PQEOM) as the percentage of qq values for which the allocator is performing at least as well as the most accurate ensemble member model (i.e. gg or bb) and Percent Q Over Max (PQOM) as the percentage of q values for which the allocator is performing better than the most accurate ensemble member model (i.e. gg or bb).

Next, we define the Percent Contribution of Feature-dependent Allocator (PCFA) as the percentage of qq values for which the feature dependent allocator aqa^{\prime}_{q} is used for allocation decisions as opposed to the feature-independent allocator aq′′a^{\prime\prime}_{q}. A value close to one indicates that aqa^{\prime}_{q} is used often, while a value close to zero indicates it is aq′′a^{\prime\prime}_{q} instead.

Next, we define the 95%95\% Threshold Q Max (95TQM) as the highest value of qq for which the ensemble performance meets or exceeds 95%95\% of the performance of the better of gg and bb. Thus this is a measure of how much explainability can be utilized before the performance price becomes material.

Next, we define the maximum accuracy achieved by the allocator across all qq (Max Acc), and the highest value of qq for which this accuracy is maintained (Argmax qq). The Max Acc can be benchmarked against the AUC, interpretable as the average accuracy across qq. Each of these metrics is a percentage and higher values correspond with higher performance and higher explainability at this maximum performance level, respectively.

Finally, we define the accuracy with which the four sufficiency categories (ZgZ^{\prime}_{g}, ZbZ^{\prime}_{b}, Z2Z^{\prime}_{2}, and Z0Z^{\prime}_{0}) can be estimated as the sufficiency accuracy (ss Acc). The higher this accuracy, the better able the allocator is to inform the user of which category a given observation is likely to be a member of.

Results

Evaluation of allocator performance using the metrics defined previously as well as visual inspection of the performance vs explainability trade-off curves (Fig. 2) revealed both the benefits and some of the limitations of learned allocation in the tabular data setting.

Firstly, performance was found to consistently and significantly outperform random allocation, as quantified by a cross-dataset PPCR of 37%37\% (Table 1), indicating that the learned allocation captured close to 40%40\% of the area under the curve available and in excess of random allocation. It was also found that on some datasets in particular, learned allocation performed close to oracle allocation (e.g. 89%89\% and 71%71\% on the IsoletR and BrazilianHousesR datasets).

Learned allocation was also found to perform at least at the level of the best ensemble member model across an average of 74%74\% of the explainability range (PQEOM in Table 1). This indicates that for many datasets, there is a substantial explainability “free lunch” to be taken advantage of without performance loss. On a few datasets, performance of the allocated ensemble was found to outperform both gg and bb for a majority (93%93\%) of the qq range (PolR and FifaR PQOM). The 95TQM metric also supported these conclusions, with a cross-dataset average value of 94%94\% indicating that allocation performance was within 5%5\% of maximal individual model performance across approximately all values of qq.

Assessing the PCFA metric suggests some limits to the upside of learned, feature dependent allocation - at least in the tested tabular data setting. A cross-dataset average value of 35%±34%35\%\pm 34\% indicates that on average, the range for which the feature dependent allocator is used over the feature-independent one is indistinguishable from zero. This is consistent with a visual inspection of the representative performance-explainability curve e.g. Fig. 2 (b) where there is no improvement to be had in excess of aq′′a^{\prime\prime}_{q}. However, it is noted that the only possibility for “homerun” allocations is through the feature dependent aqa^{\prime}_{q} as seen in Fig 2 (a) with the PolR dataset and also in Table 1 for datasets SulfurR, BikeSharingR, and FifaR. Thus the ensembled allocation scheme offers this upside without downside risk of low performance in either aqa^{\prime}_{q} or aq′′a^{\prime\prime}_{q}.

Evaluation of the case in which a single allocation is needed is also positive. On a cross-dataset average, the 84%84\% maximal accuracy achieved is quite high, and is also achieved at a high average explainability level (64%64\%). Particularly strong individual results can be seen in the Pol and SulfurR datasets (Table 1). We also find that on a observation-level, the allocation is an accurate estimator of sufficiency category, with a cross-dataset average of 76%76\% and with few datasets with accuracy under 60%60\%.

Refer to caption
(a) PolR Dataset
Refer to caption
(b) SuperconductR Dataset
Figure 2: This figure shows two examples of the explainability vs performance trade-off, comparing the random, oracle, and learned allocation curves. The PolR dataset is an example of complementary gg and bb models, resulting in an allocated ensemble that outperforms both component models across most of the qq range. The SuperconductR dataset is an example of an explainability “free lunch” in which the bb accuracy is maintained while increasing explainability using the allocator. Curves for all datasets are available in the Appendix of the long form paper available on arxiv.org.
Dataset AUC PPCR PQEOM PQOM PCFA 95TQM Max Acc Argmax qq ss Acc
Wine 79 ±\pm 0 21 ±\pm 0 71 ±\pm 0 0 ±\pm 0 7 ±\pm 0 98 ±\pm 0 80 ±\pm 0 70 ±\pm 0 78 ±\pm 0
Phoneme 87 ±\pm 0 12 ±\pm 1 78 ±\pm 4 7 ±\pm 2 6 ±\pm 1 100 ±\pm 0 87 ±\pm 0 50 ±\pm 34 81 ±\pm 2
KDDIPUMS 88 ±\pm 0 17 ±\pm 1 65 ±\pm 3 34 ±\pm 5 17 ±\pm 7 100 ±\pm 0 88 ±\pm 0 66 ±\pm 9 80 ±\pm 1
EyeMovements 66 ±\pm 0 33 ±\pm 0 55 ±\pm 1 7 ±\pm 6 15 ±\pm 2 70 ±\pm 0 68 ±\pm 0 31 ±\pm 24 52 ±\pm 0
Pol 98 ±\pm 0 49 ±\pm 0 98 ±\pm 0 2 ±\pm 0 0 ±\pm 0 100 ±\pm 0 99 ±\pm 0 98 ±\pm 0 96 ±\pm 0
Bank 76 ±\pm 0 -19 ±\pm 0 4 ±\pm 1 0 ±\pm 0 0 ±\pm 0 100 ±\pm 0 79 ±\pm 0 100 ±\pm 0 71 ±\pm 0
MagicTelescope 86 ±\pm 0 39 ±\pm 0 87 ±\pm 2 12 ±\pm 11 10 ±\pm 0 100 ±\pm 0 86 ±\pm 0 47 ±\pm 28 82 ±\pm 1
House16H 89 ±\pm 0 40 ±\pm 0 84 ±\pm 4 9 ±\pm 6 6 ±\pm 2 98 ±\pm 0 89 ±\pm 0 82 ±\pm 8 86 ±\pm 0
Credit 78 ±\pm 0 5 ±\pm 1 56 ±\pm 4 14 ±\pm 19 95 ±\pm 0 100 ±\pm 0 78 ±\pm 0 76 ±\pm 6 72 ±\pm 0
California 90 ±\pm 0 52 ±\pm 0 88 ±\pm 0 0 ±\pm 0 7 ±\pm 0 98 ±\pm 0 91 ±\pm 0 88 ±\pm 0 90 ±\pm 0
Electricity 92 ±\pm 0 58 ±\pm 0 88 ±\pm 0 0 ±\pm 0 7 ±\pm 0 98 ±\pm 0 93 ±\pm 0 88 ±\pm 0 92 ±\pm 0
Jannis 79 ±\pm 0 30 ±\pm 0 53 ±\pm 5 21 ±\pm 5 14 ±\pm 1 98 ±\pm 0 79 ±\pm 0 35 ±\pm 6 76 ±\pm 0
MiniBooNE 94 ±\pm 0 54 ±\pm 0 90 ±\pm 0 0 ±\pm 0 5 ±\pm 0 100 ±\pm 0 94 ±\pm 0 90 ±\pm 0 92 ±\pm 0
WineR 73 ±\pm 0 43 ±\pm 0 85 ±\pm 0 10 ±\pm 0 20 ±\pm 0 90 ±\pm 0 74 ±\pm 0 78 ±\pm 0 67 ±\pm 0
IsoletR 91 ±\pm 0 89 ±\pm 0 68 ±\pm 0 0 ±\pm 0 49 ±\pm 10 73 ±\pm 0 95 ±\pm 0 68 ±\pm 0 87 ±\pm 1
CPUR 75 ±\pm 0 40 ±\pm 2 52 ±\pm 18 0 ±\pm 0 29 ±\pm 11 80 ±\pm 0 77 ±\pm 0 70 ±\pm 0 59 ±\pm 1
SulfurR 98 ±\pm 0 63 ±\pm 2 73 ±\pm 9 1 ±\pm 2 79 ±\pm 4 100 ±\pm 0 98 ±\pm 0 84 ±\pm 22 96 ±\pm 0
BrazilianHousesR 96 ±\pm 0 71 ±\pm 1 83 ±\pm 0 64 ±\pm 6 1 ±\pm 2 93 ±\pm 0 98 ±\pm 0 8 ±\pm 3 88 ±\pm 0
AileronsR 75 ±\pm 0 13 ±\pm 0 70 ±\pm 10 5 ±\pm 7 4 ±\pm 2 100 ±\pm 0 76 ±\pm 0 40 ±\pm 35 64 ±\pm 0
MiamiHousingR 76 ±\pm 0 44 ±\pm 0 76 ±\pm 0 0 ±\pm 0 67 ±\pm 10 80 ±\pm 0 78 ±\pm 0 75 ±\pm 0 65 ±\pm 0
PolR 88 ±\pm 0 44 ±\pm 1 96 ±\pm 1 93 ±\pm 1 88 ±\pm 0 100 ±\pm 0 88 ±\pm 0 81 ±\pm 2 84 ±\pm 0
ElevatorsR 75 ±\pm 0 25 ±\pm 0 64 ±\pm 4 2 ±\pm 2 17 ±\pm 0 88 ±\pm 0 75 ±\pm 0 51 ±\pm 29 63 ±\pm 0
BikeSharingR 77 ±\pm 0 26 ±\pm 1 88 ±\pm 4 22 ±\pm 16 82 ±\pm 1 100 ±\pm 0 78 ±\pm 0 21 ±\pm 13 72 ±\pm 0
FifaR 77 ±\pm 0 33 ±\pm 0 95 ±\pm 0 93 ±\pm 0 78 ±\pm 0 100 ±\pm 0 77 ±\pm 0 50 ±\pm 0 69 ±\pm 0
CaliforniaR 78 ±\pm 0 38 ±\pm 0 73 ±\pm 0 68 ±\pm 0 68 ±\pm 0 93 ±\pm 0 79 ±\pm 0 42 ±\pm 0 63 ±\pm 0
HousesR 78 ±\pm 0 41 ±\pm 0 73 ±\pm 0 0 ±\pm 0 48 ±\pm 1 80 ±\pm 0 79 ±\pm 0 73 ±\pm 0 62 ±\pm 0
SuperconductR 83 ±\pm 0 42 ±\pm 0 60 ±\pm 13 24 ±\pm 11 95 ±\pm 0 95 ±\pm 0 83 ±\pm 0 57 ±\pm 1 76 ±\pm 0
HouseSalesR 76 ±\pm 0 50 ±\pm 1 64 ±\pm 12 0 ±\pm 0 56 ±\pm 6 85 ±\pm 0 78 ±\pm 0 78 ±\pm 0 63 ±\pm 0
House16HR 92 ±\pm 0 55 ±\pm 0 90 ±\pm 0 0 ±\pm 0 6 ±\pm 0 98 ±\pm 0 92 ±\pm 0 90 ±\pm 0 86 ±\pm 0
DiamondsR 70 ±\pm 0 19 ±\pm 0 83 ±\pm 0 34 ±\pm 6 73 ±\pm 0 100 ±\pm 0 71 ±\pm 0 20 ±\pm 6 65 ±\pm 0
MedicalChargesR 86 ±\pm 0 24 ±\pm 0 91 ±\pm 1 85 ±\pm 1 0 ±\pm 0 100 ±\pm 0 86 ±\pm 0 67 ±\pm 2 83 ±\pm 0
Average 83 ±\pm 9 37 ±\pm 21 74 ±\pm 19 20 ±\pm 29 35 ±\pm 34 94 ±\pm 9 84 ±\pm 8 64 ±\pm 24 76 ±\pm 12
Table 1: This table summarizes allocation performance across several metrics on each dataset and, in the bottom row, across the datasets (all metrics are reported as percentages). Averages and standard deviations are reported over 5 replicates. Metric definitions can be found in the Metrics section, and discussion of results can be found in the Results section.

Ablation Studies

Allocator Feature Set Selection

In addition to the features xx used to learn the glass box and black box models, the allocation task also has access to their predictions g(x)g(x) and b(x)b(x), and any functions of the two - since the allocator is learned subsequent to the training of these models. To obtain the optimal feature set for allocation, standard tuning procedures (e.g. cross validation) can be employed to evaluate all feature sets of interest. However, as each candidate feature set requires the training of a corresponding allocator for evaluation, this approach can be prohibitively costly.

Thus, the following study was conducted to determine whether a consistently best feature set exists for the tabular data context used in the experiments. First, the universe of candidate features was selected to be the original features xx used as inputs for the ensemble component models, the predictions of both of these models g(x)g(x) and b(x)b(x), and finally two measures of discrepancy between the predictions, the cross-entropy dce(g(x),b(x))d_{ce}(g(x),b(x)) and the mean squared error dmse(g(x),b(x))d_{mse}(g(x),b(x)). The measures of disagreement were included as features as they translate to the “feature independent” strategy of allocation to model bb for low values of d(a(x),b(x))d(a(x),b(x)) - in other words the optimal allocation strategy assuming aa is always correct.

The candidate features were grouped into the sets listed in Table 2 and then used to train allocators on a subset of the benchmark datasets (Wine, WineR, Phoneme, SulfurR, Bank, BrazilianHousesR, FifaR, KDDIPUMS) with 6 replicates per model.

Next each feature set was evaluated as follows. First, within each dataset, each feature set’s performance (defined as the AUC) was compared to the performance of the best alternative set of features. Then, the proportion of datasets for which the feature set being evaluated was not significantly worse (i.e. either significantly better or not significantly different) than the best alternative was recorded and reported in Table 2 for three significance levels (10%10\%, 5%5\%, 1%1\%).

The results support the following three conclusions. First, no one feature set proved universally best across the tested datasets and thus a full search across feature sets would be advised in settings without resource constraint. Second, although no universally best feature set was found, the “kitchen sink” set of all candidate features (xx, gg, bb, dced_{ce}, dmsed_{mse}) was found to be best most consistently and was thus used to train all allocators reported in Table 1. Finally, allocators trained on just the original features xx were found to be consistently worst among all alternatives thus supporting the augmentation of the original features in some form. This finding is consistent with the intuition that the predictions of the component models would be very useful to learning the optimal allocation and would be either very difficult or impossible to learn from rr^{\prime}, the optimal allocation ranking response, alone during training.

Feature Set α:0.01\alpha:0.01 α:0.05\alpha:0.05 α:0.1\alpha:0.1
xx 18.75% 18.75% 12.50%
gg, bb 43.75% 43.75% 31.25%
dced_{ce} 31.25% 18.75% 18.75%
dmsed_{mse} 50.00% 43.75% 37.50%
xx, dced_{ce} 37.50% 25.00% 18.75%
xx, dmsed_{mse} 56.25% 37.50% 25.00%
gg, bb, dced_{ce} 31.25% 25.00% 25.00%
gg, bb, dmsed_{mse} 50.00% 50.00% 37.50%
xx, gg, bb 56.25% 43.75% 37.50%
xx, gg, bb, dced_{ce} 50.00% 43.75% 37.50%
xx, gg, bb, dmsed_{mse} 56.25% 37.50% 37.50%
xx, gg, bb, dced_{ce}, dmsed_{mse} 75.00% 56.25% 43.75%
Table 2: This table reports the percentage of datasets for which the allocator learned on the corresponding feature set is significantly better than or not significantly different from the best alternative feature set trained allocator. Results across three significance levels are reported and show that the “kitchen sink” x,g,b,dce,dmsex,g,b,d_{ce},d_{mse} feature set is most consistently best (bolded) while the “unaugmented” original feature set of xx is consistently the worst across all α\alpha.

Ensemble Component Model Selection

The performance of any allocated ensemble is highly dependent not only on the individual performance of its component models (i.e. gg and bb) but on their level of synergy as well. In particular, it may be the case that the component model pair in (g0,b0,a0)(g_{0},b_{0},a_{0}) individually outperforms the respective component models in (g1,b1,a1)(g_{1},b_{1},a_{1}) but that the allocator a0a_{0} trained with (g0,b0)(g_{0},b_{0}) underperforms a1a_{1}. In this case, the high relative advantages of (g1,b1)(g_{1},b_{1}) in different segments of the feature space overcome their global performance disadvantages as individual models compared to their counterparts in (g0,b0)(g_{0},b_{0}) to yield a stronger ensemble.

Thus, to determine how often high relative advantage overcomes superior individual performance in allocator training, the following study was conducted. For each dataset, an allocator was trained on each combination of available glass box (tree and regression) and black box (gradient boosting trees and neural network) models (i.e. four allocators per dataset). Then the allocator aIa_{I}, trained using the pair of component models (gI,bI)(g_{I},b_{I}) with the highest individual validation performance, was identified along with the allocator aCa_{C}, trained using the pair of component models (gC,bC)(g_{C},b_{C}) resulting in the highest ensemble validation performance. Finally the difference in test performance was measured between aCa_{C} and aIa_{I} (AUCΔ=AUC(aC)AUC(aI)AUC\Delta=AUC(a_{C})-AUC(a_{I})).

The resulting AUCΔAUC\Delta values support the following two conclusions. Firstly, while a relatively high proportion (41.9%41.9\%) of datasets yield different allocators depending on which of the two different component model selection processes (individual vs. combined performance) they utilize, the cross-dataset average difference in allocator performance is not significantly different from zero (0.01±0.030.01\pm 0.03). This result suggests that the glass box and black box model types used for the experiments did not exhibit high relative expertise in different parts of the feature space, indicating that it may be beneficial to use a more diverse set of component models in this setting. However, in rare cases (e.g. IsoletR, BrazilianHousesR) the combined performance selection method yields as much as 15%15\% in additional performance. Thus, in resource constrained settings, or in cases in which many glass box and black box model types are under consideration, the individual performance selection method appears relatively low risk, although a full search across all component model combinations (the method used for Table 1) is recommended when feasible.

Dataset gIg_{I} bIb_{I} gCg_{C} bCb_{C} Match? AUC Δ\Delta
Wine Tree DNN Tree DNN Yes 0.0
Phoneme Tree GBT Tree DNN No (0.01)
KDDIPUMS Tree GBT Tree GBT Yes 0.0
EyeMovements Regr. GBT Regr. GBT Yes 0.0
Pol Tree DNN Tree GBT No (0.0)
Bank Tree DNN Tree DNN Yes 0.0
MagicTelescope Tree GBT Tree GBT Yes 0.0
House16H Tree GBT Tree GBT Yes 0.0
Credit Tree GBT Tree GBT Yes 0.0
California Tree GBT Tree GBT Yes 0.0
Electricity Tree GBT Tree GBT Yes 0.0
Jannis Tree GBT Tree GBT Yes 0.0
MiniBooNE Tree GBT Tree GBT Yes 0.0
WineR Regr. GBT Regr. GBT Yes 0.0
IsoletR Regr. DNN Tree DNN No 0.15
CPUR Tree GBT Tree DNN No 0.0
SulfurR Tree DNN Tree GBT No 0.04
BrazilianHousesR Tree GBT Tree DNN No 0.08
AileronsR Regr. GBT Regr. DNN No 0.0
MiamiHousingR Tree GBT Tree GBT Yes 0.0
PolR Tree DNN Tree GBT No (0.0)
ElevatorsR Regr. DNN Regr. GBT No 0.01
BikeSharingR Tree GBT Tree GBT Yes 0.0
FifaR Tree GBT Tree DNN No 0.0
CaliforniaR Tree GBT Tree DNN No 0.01
HousesR Tree GBT Tree DNN No 0.01
SuperconductR Tree GBT Tree GBT Yes 0.0
HouseSalesR Tree GBT Tree GBT Yes 0.0
House16HR Tree GBT Tree GBT Yes 0.0
DiamondsR Tree GBT Tree GBT Yes 0.0
MedicalChargesR Tree GBT Tree DNN No 0.0
Median/Average Tree (83.9%) GBT (77.4%) Tree (87.1%) GBT (64.5%) Yes (58.1%) 0.01 ±\pm 0.03
Table 3: For each dataset, this table reports the glass box (gg) and black box (bb) models which have the highest individual validation performance (Indiv. gg and Indiv. bb) as well as the (gg, bb) pair that produces the highest allocator validation set performance (Combi. gg and Combi. bb). Also shown is the difference (AUCIndiv.AUCCombi.AUC_{Indiv.}-AUC_{Combi.}) in the test performance between allocators trained on the individually selected component models and paired component models - a positive value indicating that the paired component models result in a better allocator.

Acknowledgements / Thanks

We would like to thank Jonathan Siegel for valuable discussion of the theoretical aspects of this work.

This research is supported by the National Science Foundation under grant number CCF-2205004. Computations for this research were performed on the Pennsylvania State University’s Institute for Computational and Data Sciences’ Roar supercomputer.


References

  • Agarwal et al. (2021) Agarwal, R.; Melnick, L.; Frosst, N.; Zhang, X.; Lengerich, B.; Caruana, R.; and Hinton, G. E. 2021. Neural additive models: Interpretable machine learning with neural nets. Advances in neural information processing systems, 34: 4699–4711.
  • Ferry, Laberge, and Aïvodji (2023) Ferry, J.; Laberge, G.; and Aïvodji, U. 2023. Learning Hybrid Interpretable Models: Theory, Taxonomy, and Methods. arXiv preprint arXiv:2303.04437.
  • Gage et al. (2001) Gage, B. F.; Waterman, A. D.; Shannon, W.; Boechler, M.; Rich, M. W.; and Radford, M. J. 2001. Validation of clinical classification schemes for predicting stroke: results from the National Registry of Atrial Fibrillation. Jama, 285(22): 2864–2870.
  • Gao et al. (2019) Gao, X.; Shan, C.; Hu, C.; Niu, Z.; and Liu, Z. 2019. An adaptive ensemble machine learning model for intrusion detection. Ieee Access, 7: 82512–82521.
  • Garreau and Luxburg (2020) Garreau, D.; and Luxburg, U. 2020. Explaining the explainer: A first theoretical analysis of LIME. In International conference on artificial intelligence and statistics, 1287–1296. PMLR.
  • Grinsztajn, Oyallon, and Varoquaux (2022) Grinsztajn, L.; Oyallon, E.; and Varoquaux, G. 2022. Why do tree-based models still outperform deep learning on typical tabular data? Advances in Neural Information Processing Systems, 35: 507–520.
  • Gulum, Trombley, and Kantardzic (2021) Gulum, M. A.; Trombley, C. M.; and Kantardzic, M. 2021. A review of explainable deep learning cancer detection models in medical imaging. Applied Sciences, 11(10): 4573.
  • Inoue (2019) Inoue, H. 2019. Adaptive ensemble prediction for deep neural networks based on confidence level. In The 22nd International Conference on Artificial Intelligence and Statistics, 1284–1293. PMLR.
  • Lemhadri, Ruan, and Tibshirani (2021) Lemhadri, I.; Ruan, F.; and Tibshirani, R. 2021. LassoNet: Neural Networks with Feature Sparsity. In Banerjee, A.; and Fukumizu, K., eds., Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, volume 130 of Proceedings of Machine Learning Research, 10–18. PMLR.
  • Linardatos, Papastefanopoulos, and Kotsiantis (2020) Linardatos, P.; Papastefanopoulos, V.; and Kotsiantis, S. 2020. Explainable ai: A review of machine learning interpretability methods. Entropy, 23(1): 18.
  • Lundberg and Lee (2017) Lundberg, S. M.; and Lee, S.-I. 2017. A unified approach to interpreting model predictions. Advances in neural information processing systems, 30.
  • Nan and Saligrama (2017) Nan, F.; and Saligrama, V. 2017. Adaptive classification for prediction under a budget. Advances in neural information processing systems, 30.
  • Nauta et al. (2022) Nauta, M.; Trienes, J.; Pathak, S.; Nguyen, E.; Peters, M.; Schmitt, Y.; Schlötterer, J.; van Keulen, M.; and Seifert, C. 2022. From anecdotal evidence to quantitative evaluation methods: A systematic review on evaluating explainable ai. ACM Computing Surveys.
  • Pan, Wang, and Hara (2020) Pan, D.; Wang, T.; and Hara, S. 2020. Interpretable companions for black-box models. In International conference on artificial intelligence and statistics, 2444–2454. PMLR.
  • Ribeiro, Singh, and Guestrin (2016) Ribeiro, M. T.; Singh, S.; and Guestrin, C. 2016. ” Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, 1135–1144.
  • Rudin (2019) Rudin, C. 2019. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature machine intelligence, 1(5): 206–215.
  • Rymarczyk et al. (2020) Rymarczyk, D.; Struski, Ł.; Tabor, J.; and Zieliński, B. 2020. Protopshare: Prototype sharing for interpretable image classification and similarity discovery. arXiv preprint arXiv:2011.14340.
  • Sahakyan, Aung, and Rahwan (2021) Sahakyan, M.; Aung, Z.; and Rahwan, T. 2021. Explainable Artificial Intelligence for Tabular Data: A Survey. IEEE Access, 9: 135392–135422.
  • Wang, Fujimaki, and Motohashi (2015) Wang, J.; Fujimaki, R.; and Motohashi, Y. 2015. Trading interpretability for accuracy: Oblique treed sparse additive models. In Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, 1245–1254.
  • Wang and Saligrama (2012) Wang, J.; and Saligrama, V. 2012. Local supervised learning through space partitioning. Advances in Neural Information Processing Systems, 25.
  • Wang (2019) Wang, T. 2019. Gaining free or low-cost interpretability with interpretable partial substitute. In International Conference on Machine Learning, 6505–6514. PMLR.
  • Wang and Lin (2021) Wang, T.; and Lin, Q. 2021. Hybrid predictive models: When an interpretable model collaborates with a black-box model. The Journal of Machine Learning Research, 22(1): 6085–6122.
  • Zagoruyko and Komodakis (2016) Zagoruyko, S.; and Komodakis, N. 2016. Wide residual networks. arXiv preprint arXiv:1605.07146.

Appendix A Appendix

Theoretical Results

Definition 1.

(Z Sets)
Z=ZZZ^{\prime}_{\bigtriangleup\square}=Z^{\prime}_{\bigtriangleup}\cup Z^{\prime}_{\square} and Z,=Z×ZZ^{\prime}_{\bigtriangleup,\square}=Z^{\prime}_{\bigtriangleup}\times Z^{\prime}_{\square}

Definition 2.

(C Sets)
C,(a)={(zi,zj):ziZ,zjZ,a(zi)<a(zj)}C_{\bigtriangleup,\square}(a)=\{(z_{i},z_{j}):z_{i}\in Z^{\prime}_{\bigtriangleup},z_{j}\in Z^{\prime}_{\square},a(z_{i})<a(z_{j})\}

Lemma 1.

(Increasing in r~\tilde{r}) (zb,z0,z2,zg)(Zb×Z0×Z2×Zg),r~(zb)<r~(z0)<r~(z2)<r~(zg)\forall(z_{b},z_{0},z_{2},z_{g})\in(Z^{\prime}_{b}\times Z^{\prime}_{0}\times Z^{\prime}_{2}\times Z^{\prime}_{g}),\tilde{r}^{\prime}(z_{b})<\tilde{r}^{\prime}(z_{0})<\tilde{r}^{\prime}(z_{2})<\tilde{r}^{\prime}(z_{g})

Proof.

Notice {r~(zb):zbZb}(2,1),{r~(z0):z0Z0}(1,0),{r~(z2):z2Z2}(0,1),{r~(zg):zgZg}(1,2)\{\tilde{r}^{\prime}(z_{b})\colon z_{b}\in Z^{\prime}_{b}\}\subseteq(-2,-1),\{\tilde{r}^{\prime}(z_{0})\colon z_{0}\in Z^{\prime}_{0}\}\subseteq(-1,0),\{\tilde{r}^{\prime}(z_{2})\colon z_{2}\in Z^{\prime}_{2}\}\subseteq(0,1),\{\tilde{r}^{\prime}(z_{g})\colon z_{g}\in Z^{\prime}_{g}\}\subseteq(1,2), statement follows. ∎

Lemma 2.

(aqa^{\prime}_{q} to r~\tilde{r}) |Zq|=nq,Zq=R~q|Z^{\prime}_{q}|=n_{q},Z^{\prime}_{q}=\tilde{R}^{\prime}_{q} where Zq={z:aq(z)=1},R~q={z:rank(r~(z))>n1q}Z^{\prime}_{q}=\{z\colon a^{\prime}_{q}(z)=1\},\tilde{R}^{\prime}_{q}=\{z\colon\mathrm{rank}(\tilde{r}^{\prime}(z))>n_{1-q}\}

Proof.

Part 1: Zq={z:aq(z)=1}={z:r(z)>1q}={z:rank(r~(z))n>1q}={z:rank(r~(z))>n(1q)}={z:rank(r~(z))>n1q}=R~qZ^{\prime}_{q}=\{z\colon a^{\prime}_{q}(z)=1\}=\{z\colon r^{\prime}(z)>1-q\}=\{z\colon\frac{\mathrm{rank}(\tilde{r}^{\prime}(z))}{n}>1-q\}=\{z:\mathrm{rank}(\tilde{r}^{\prime}(z))>n(1-q)\}=\{z:\mathrm{rank}(\tilde{r}^{\prime}(z))>n_{1-q}\}=\tilde{R}^{\prime}_{q}
Part 2: Observe that, by construction, r~(z)\tilde{r}^{\prime}(z) is the r~(z)th\tilde{r}^{\prime}(z)^{th} percentile. Thus, |Zq|=|{z:r(z)>1q}|=n(1(1q))=nq|Z^{\prime}_{q}|=|\{z\colon r^{\prime}(z)>1-q\}|=n(1-(1-q))=n_{q}. ∎

Lemma 3.

(Set Equivalence 1) Aq=Bq,q𝒬A^{*}_{q}=B^{*}_{q},\forall q\in\mathcal{Q}, where Aq={aq:aq=argmaxaqAq1ni=1nsg(zi)aq(zi)+sb(zi)(1aq(zi))}A^{*}_{q}=\{a^{*}_{q}\colon a^{*}_{q}=\arg\max_{a_{q}\in A_{q}}\frac{1}{n}\sum_{i=1}^{n}s^{\prime}_{g}(z_{i})a_{q}(z_{i})+s^{\prime}_{b}(z_{i})(1-a_{q}(z_{i}))\} and Bq={bq:bqAq,Cg,02(bq)=C02,b(bq)=Cg,b(bq)=}B^{*}_{q}=\{b^{*}_{q}:b^{*}_{q}\in A_{q},C_{g,02}(b^{*}_{q})=C_{02,b}(b^{*}_{q})=C_{g,b}(b^{*}_{q})=\emptyset\}

Proof.

Part 1: We show that bqBqbqAqb^{*}_{q}\in B^{*}_{q}\implies b^{*}_{q}\in A^{*}_{q} for given q𝒬q\in\mathcal{Q}. Assume for contradiction that bqAqb^{*}_{q}\notin A^{*}_{q}. Then (zi,zj)𝒵×𝒵\exists(z_{i}^{\circ},z_{j}^{\circ})\in\mathcal{Z}\times\mathcal{Z} s.t. [sg(zi)bq(zi)+sb(zi)(1bq(zi))]+[sg(zj)bq(zj)+sb(zj)(1bq(zj))]<[sg(zi)bq(zj)+sb(zi)(1bq(zj))]+[sg(zj)bq(zi)+sb(zj)(1bq(zi))][s^{\prime}_{g}(z_{i}^{\circ})b^{*}_{q}(z_{i}^{\circ})+s^{\prime}_{b}(z_{i}^{\circ})(1-b^{*}_{q}(z_{i}^{\circ}))]+[s^{\prime}_{g}(z_{j}^{\circ})b^{*}_{q}(z_{j}^{\circ})+s^{\prime}_{b}(z_{j}^{\circ})(1-b^{*}_{q}(z_{j}^{\circ}))]<[s^{\prime}_{g}(z_{i}^{\circ})b^{*}_{q}(z_{j}^{\circ})+s^{\prime}_{b}(z_{i}^{\circ})(1-b^{*}_{q}(z_{j}^{\circ}))]+[s^{\prime}_{g}(z_{j}^{\circ})b^{*}_{q}(z_{i}^{\circ})+s^{\prime}_{b}(z_{j}^{\circ})(1-b^{*}_{q}(z_{i}^{\circ}))] i.e. there exists a pair of observations for which swapping allocation decisions increases the objective function.
Case 1 (L.3.1.1): (zi,zj)Zg,gZb,bZ02,02(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{g,g}\cup Z^{\prime}_{b,b}\cup Z^{\prime}_{02,02}. But, (zi,zj)Zg,gZb,bZ02,02\forall(z_{i},z_{j})\in Z^{\prime}_{g,g}\cup Z^{\prime}_{b,b}\cup Z^{\prime}_{02,02}, we have that sg(zi)bq(zi)+sb(zi)(1bq(zi))]+[sg(zj)bq(zj)+sb(zj)(1bq(zj))]=[sg(zi)bq(zj)+sb(zi)(1bq(zj))]+[sg(zj)bq(zi)+sb(zj)(1bq(zi))]s^{\prime}_{g}(z_{i})b^{*}_{q}(z_{i})+s^{\prime}_{b}(z_{i})(1-b^{*}_{q}(z_{i}))]+[s^{\prime}_{g}(z_{j})b^{*}_{q}(z_{j})+s^{\prime}_{b}(z_{j})(1-b^{*}_{q}(z_{j}))]=[s^{\prime}_{g}(z_{i})b^{*}_{q}(z_{j})+s^{\prime}_{b}(z_{i})(1-b^{*}_{q}(z_{j}))]+[s^{\prime}_{g}(z_{j})b^{*}_{q}(z_{i})+s^{\prime}_{b}(z_{j})(1-b^{*}_{q}(z_{i}))] which contradicts the definition of (zi,zj)(z_{i}^{\circ},z_{j}^{\circ}), thus (zi,zj)Zg,gZb,bZ02,02(z_{i}^{\circ},z_{j}^{\circ})\notin Z^{\prime}_{g,g}\cup Z^{\prime}_{b,b}\cup Z^{\prime}_{02,02}.
Case 2 (L.3.1.2): (zi,zj)Zg,02(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{g,02}. Then, t(bq,zi)=bq(zi),t(bq,zj)=c={1,zj𝒵20,zj𝒵0t^{\prime}(b^{*}_{q},z_{i}^{\circ})=b^{*}_{q}(z_{i}^{\circ}),t^{\prime}(b^{*}_{q},z_{j}^{\circ})=c=\begin{cases}1,z_{j}^{\circ}\in\mathcal{Z}^{\prime}_{2}\\ 0,z_{j}^{\circ}\in\mathcal{Z}^{\prime}_{0}\end{cases}. Thus, t(bq,zi)+t(bq,zj)=bq(zi)+ct^{\prime}(b^{*}_{q},z_{i}^{\circ})+t^{\prime}(b^{*}_{q},z_{j}^{\circ})=b^{*}_{q}(z_{i}^{\circ})+c which by assumption on zi,zjbq(zi)=0<bq(zj)=1Cg,02(bq)z_{i}^{\circ},z_{j}^{\circ}\implies b^{*}_{q}(z_{i}^{\circ})=0<b^{*}_{q}(z_{j}^{\circ})=1\implies C_{g,02}(b^{*}_{q})\neq\emptyset which is a contradiction with bqBqb^{*}_{q}\in B^{*}_{q}, so (zi,zj)Zg,02(z_{i}^{\circ},z_{j}^{\circ})\notin Z^{\prime}_{g,02}.
Case 3 (L.3.1.3): (zi,zj)Z02,b(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{02,b}. Using the same argument as in Case 2 (L.3.1.2), contradiction and thus (zi,zj)Z02,b(z_{i}^{\circ},z_{j}^{\circ})\notin Z^{\prime}_{02,b} follows.
Case 4 (L.3.1.4): (zi,zj)Zg,b(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{g,b}. Using the same argument as in Case 2 (L.3.1.2), contradiction and thus (zi,zj)Zg,b(z_{i}^{\circ},z_{j}^{\circ})\notin Z^{\prime}_{g,b} follows.
Thus we have shown that q\forall q, no such zi,zjz_{i}^{\circ},z_{j}^{\circ} exist, and so bqAqb^{*}_{q}\in A^{*}_{q}.
Part 2: We show aqAqaqBqa^{*}_{q}\in A^{*}_{q}\implies a^{*}_{q}\in B^{*}_{q} for given q𝒬q\in\mathcal{Q}. Assume for contradiction that aqBqa^{*}_{q}\notin B^{*}_{q}. Then zi,zj\exists z_{i}^{\circ},z_{j}^{\circ} s.t. (zi,zj)Cg,02C02,bCg,b(z^{\circ}_{i},z^{\circ}_{j})\in C_{g,02}\cup C_{02,b}\cup C_{g,b}.
Case 1 (L.3.2.1): (zi,zj)Cg,02(z_{i}^{\circ},z_{j}^{\circ})\in C_{g,02}. Then, t(aq,zi)=aq(zi)t^{\prime}(a^{*}_{q},z_{i}^{\circ})=a_{q}(z_{i}^{\circ}) and t(aq,zj)=c={1,zj𝒵20,zj𝒵0t^{\prime}(a^{*}_{q},z_{j}^{\circ})=c=\begin{cases}1,z_{j}^{\circ}\in\mathcal{Z}^{\prime}_{2}\\ 0,z_{j}^{\circ}\in\mathcal{Z}^{\prime}_{0}\end{cases} and aq(zi)=0<1=aq(zj)a^{*}_{q}(z_{i}^{\circ})=0<1=a^{*}_{q}(z_{j}^{\circ}). But then, t(aq,zi)+t(aq,zj)=c<1+c=[sg(zi)aq(zj)+s(zi)(1aq(zj))]+[sg(zj)aq(zi)+s(zj)(1aq(zi))]t^{\prime}(a^{*}_{q},z_{i}^{\circ})+t^{\prime}(a^{*}_{q},z_{j}^{\circ})=c<1+c=[s^{\prime}_{g}(z_{i}^{\circ})a^{*}_{q}(z_{j}^{\circ})+s^{\prime}(z_{i}^{\circ})(1-a^{*}_{q}(z_{j}^{\circ}))]+[s^{\prime}_{g}(z_{j}^{\circ})a^{*}_{q}(z_{i}^{\circ})+s^{\prime}(z_{j}^{\circ})(1-a^{*}_{q}(z_{i}^{\circ}))] i.e. swapping the allocation decisions of ziz_{i} and zjz_{j} increases the objective function aqAq\implies a^{*}_{q}\notin A^{*}_{q} which is a contradiction. Thus, (zi,zj)Cg,02(z_{i}^{\circ},z_{j}^{\circ})\notin C_{g,02}.
Case 2 (L.3.2.2): (zi,zj)C02,b(z_{i}^{\circ},z_{j}^{\circ})\in C_{02,b}. Using the same argument as in Case 1 (L.3.2.1), zi,zjC02,bz_{i}^{\circ},z_{j}^{\circ}\notin C_{02,b} follows.
Case 3 (L.3.2.3): (zi,zj)Cg,b(z_{i}^{\circ},z_{j}^{\circ})\in C_{g,b}. Using the same argument as in Case 1 (L.3.2.1), zi,zjCg,bz_{i}^{\circ},z_{j}^{\circ}\notin C_{g,b} follows.
Thus, (zi,zj)Cg,02C02,bCg,baqBq(z^{\circ}_{i},z^{\circ}_{j})\notin C_{g,02}\cup C_{02,b}\cup C_{g,b}\implies a^{*}_{q}\in B^{*}_{q}
Having shown both directions of implication, we conclude Aq=BqA^{*}_{q}=B^{*}_{q}

Proposition 1.

(Maximal Sufficient Performance) q𝒬,aqAq\forall q\in\mathcal{Q},a^{\prime}_{q}\in A^{*}_{q} where Aq={aq:aq=argmaxaqAqt¯(aq)}A^{*}_{q}=\{a^{*}_{q}\colon a^{*}_{q}=\arg\max_{a_{q}\in A_{q}}\bar{t}^{\prime}(a_{q})\} and t¯(aq)=1ni=1nt(aq,zi)=1ni=1nsg(zi)aq(zi)+sb(zi)(1aq(zi))\bar{t}^{\prime}(a_{q})=\frac{1}{n}\sum_{i=1}^{n}t^{\prime}(a_{q},z_{i})=\frac{1}{n}\sum_{i=1}^{n}s^{\prime}_{g}(z_{i})a_{q}(z_{i})+s^{\prime}_{b}(z_{i})(1-a_{q}(z_{i}))

Proof.

From Lemma 3, it is enough to show that aqBqa^{\prime}_{q}\in B^{*}_{q}. First, from Lemma 2, it follows that aqAqa^{\prime}_{q}\in A_{q}. Next, we show that Cg,02(aq)=C02,b(aq)=Cg,b(aq)=C_{g,02}(a^{\prime}_{q})=C_{02,b}(a^{\prime}_{q})=C_{g,b}(a^{\prime}_{q})=\emptyset.
Case 1 (P.1.1): (zi,zj)Zg,02(z^{\circ}_{i},z^{\circ}_{j})\in Z^{\prime}_{g,02}. From Lemma 1, r~(zi)>r~(zj)rankDn(r~(zi))>rankDn(r~(zj))𝕀{rankDn(r~(zi))>n1q}𝕀{rankDn(r~(zj))>n1q}aq(zi)aq(zj)\tilde{r}^{\prime}(z^{\circ}_{i})>\tilde{r}^{\prime}(z^{\circ}_{j})\iff\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z^{\circ}_{i}))>\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z^{\circ}_{j}))\implies\mathbb{I}\{\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z^{\circ}_{i}))>n_{1-q}\}\geq\mathbb{I}\{\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z^{\circ}_{j}))>n_{1-q}\}\iff a^{\prime}_{q}(z^{\circ}_{i})\geq a^{\prime}_{q}(z^{\circ}_{j}) and thus (zi,zj)Cg,02(aq)(z^{\circ}_{i},z^{\circ}_{j})\notin C_{g,02}(a^{\prime}_{q}).
Case 2 (P.1.2): (zi,zj)Z02,b(z^{\circ}_{i},z^{\circ}_{j})\in Z^{\prime}_{02,b}. Using the same argument as in Case 1 (P.1.1), (zi,zj)C02,b(aq)(z^{\circ}_{i},z^{\circ}_{j})\notin C_{02,b}(a^{\prime}_{q}) follows.
Case 3 (P.1.3): (zi,zj)Zg,b(z^{\circ}_{i},z^{\circ}_{j})\in Z^{\prime}_{g,b}. Using the same argument as in Case 1 (P.1.1), (zi,zj)Cg,b(aq)(z^{\circ}_{i},z^{\circ}_{j})\notin C_{g,b}(a^{\prime}_{q}) follows.
Case 4 (P.1.4): (zi,zj)(𝒵×𝒵)(Zg,02Z02,bZg,b)(zi,zj)Cg,02(aq)C02,b(aq)Cg,b(aq)(z^{\circ}_{i},z^{\circ}_{j})\in(\mathcal{Z}\times\mathcal{Z})-(Z^{\prime}_{g,02}\cup Z^{\prime}_{02,b}\cup Z^{\prime}_{g,b})\implies(z^{\circ}_{i},z^{\circ}_{j})\notin C_{g,02}(a^{\prime}_{q})\cup C_{02,b}(a^{\prime}_{q})\cup C_{g,b}(a^{\prime}_{q}) by definition.
We have shown that aqAqa^{\prime}_{q}\in A_{q} and Cg,02(aq)=C02,b(aq)=Cg,b(aq)=C_{g,02}(a^{\prime}_{q})=C_{02,b}(a^{\prime}_{q})=C_{g,b}(a^{\prime}_{q})=\emptyset, thus aqBq=Aqa^{\prime}_{q}\in B^{*}_{q}=A^{*}_{q}. ∎

Lemma 4.

(Set Equivalence 2) Aq|g=Bq|gA^{*}_{q|g}=B^{*}_{q|g} where q𝒬,Aq|g={aq|g:aq|g=argmaxaqAq1ni=1nsg(zi)aq(zi)}q\in\mathcal{Q},A^{*}_{q|g}=\{a^{*}_{q|g}\colon a^{*}_{q|g}=\arg\max_{a^{*}_{q}\in A^{*}_{q}}\frac{1}{n}\sum_{i=1}^{n}s^{\prime}_{g}(z_{i})a^{*}_{q}(z_{i})\} and Bq|g={bq|g:bq|gAq,C2,0(bq|g)=}B^{*}_{q|g}=\{b^{*}_{q|g}:b^{*}_{q|g}\in A^{*}_{q},C_{2,0}(b^{*}_{q|g})=\emptyset\}

Proof.

Part 1: First we show that bq|gBq|gbq|gAq|gb^{*}_{q|g}\in B^{*}_{q|g}\implies b^{*}_{q|g}\in A^{*}_{q|g}. Assume for contradiction that bq|gAq|gb^{*}_{q|g}\notin A^{*}_{q|g}, then (zi,zj)𝒵×𝒵\exists(z_{i}^{\circ},z_{j}^{\circ})\in\mathcal{Z}\times\mathcal{Z} s.t. α=sg(zi)bq|g(zi)+sg(zj)bq|g(zj)<sg(zi)bq|g(zj)+sg(zj)bq|g(zi)=β\alpha=s^{\prime}_{g}(z_{i}^{\circ})b^{*}_{q|g}(z_{i}^{\circ})+s^{\prime}_{g}(z_{j}^{\circ})b^{*}_{q|g}(z_{j}^{\circ})<s^{\prime}_{g}(z_{i}^{\circ})b^{*}_{q|g}(z_{j}^{\circ})+s^{\prime}_{g}(z_{j}^{\circ})b^{*}_{q|g}(z_{i}^{\circ})=\beta i.e. there exists a pair of observations for which swapping allocation decisions increases the gg-specific objective function. Next, bq|gBq|gbq|gAq(zi,zj)𝒵×𝒵b^{*}_{q|g}\in B^{*}_{q|g}\implies b^{*}_{q|g}\in A^{*}_{q}\implies\forall(z_{i},z_{j})\in\mathcal{Z}\times\mathcal{Z}, and in particular for (zi,zj)(z_{i}^{\circ},z_{j}^{\circ}), [sg(zi)bq|g(zi)+sb(zi)(1bq|g(zi))]+[sg(zj)bq|g(zj)+sb(zj)(1bq|g(zj))][sg(zi)bq|g(zj)+sb(zi)(1bq|g(zj))]+[sg(zj)bq|g(zi)+sb(zj)(1bq|g(zi))]αβ+(bq|g(zi)bq|g(zj))(sb(zi)sb(zj))[s^{\prime}_{g}(z_{i}^{\circ})b^{*}_{q|g}(z_{i}^{\circ})+s^{\prime}_{b}(z_{i}^{\circ})(1-b^{*}_{q|g}(z_{i}^{\circ}))]+[s^{\prime}_{g}(z_{j}^{\circ})b^{*}_{q|g}(z_{j}^{\circ})+s^{\prime}_{b}(z_{j}^{\circ})(1-b^{*}_{q|g}(z_{j}^{\circ}))]\geq[s^{\prime}_{g}(z_{i}^{\circ})b^{*}_{q|g}(z_{j}^{\circ})+s^{\prime}_{b}(z_{i}^{\circ})(1-b^{*}_{q|g}(z_{j}^{\circ}))]+[s^{\prime}_{g}(z_{j}^{\circ})b^{*}_{q|g}(z_{i}^{\circ})+s^{\prime}_{b}(z_{j}^{\circ})(1-b^{*}_{q|g}(z_{i}^{\circ}))]\iff\alpha\geq\beta+(b^{*}_{q|g}(z_{i}^{\circ})-b^{*}_{q|g}(z_{j}^{\circ}))(s^{\prime}_{b}(z_{i}^{\circ})-s^{\prime}_{b}(z_{j}^{\circ})). Observe, (Z2,0)c=[Zg,gZb,bZg,bZg,02Z02,bZ0,0Z2,2][Zg,bZ0,bZg,2](Z^{\prime}_{2,0})^{c}=[Z^{\prime}_{g,g}\cup Z^{\prime}_{b,b}\cup Z^{\prime}_{g,b}\cup Z^{\prime}_{g,02}\cup Z^{\prime}_{02,b}\cup Z^{\prime}_{0,0}\cup Z^{\prime}_{2,2}]\cup[Z^{\prime}_{g,b}\cup Z^{\prime}_{0,b}\cup Z^{\prime}_{g,2}] Case 1 (L.4.1.1): (zi,zj)Zg,gZb,bZg,bZg,02Z02,bZ0,0Z2,2(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{g,g}\cup Z^{\prime}_{b,b}\cup Z^{\prime}_{g,b}\cup Z^{\prime}_{g,02}\cup Z^{\prime}_{02,b}\cup Z^{\prime}_{0,0}\cup Z^{\prime}_{2,2}. Then sb(zi)sb(zj)=0αβs^{\prime}_{b}(z_{i}^{\circ})-s^{\prime}_{b}(z_{j}^{\circ})=0\implies\alpha\geq\beta. But this is a contradiction with the assumed α<β\alpha<\beta, thus (zi,zj)Zg,gZb,bZg,bZg,02Z02,bZ0,0Z2,2(z_{i}^{\circ},z_{j}^{\circ})\notin Z^{\prime}_{g,g}\cup Z^{\prime}_{b,b}\cup Z^{\prime}_{g,b}\cup Z^{\prime}_{g,02}\cup Z^{\prime}_{02,b}\cup Z^{\prime}_{0,0}\cup Z^{\prime}_{2,2}.
Case 2 (L.4.1.2): (zi,zj)Zg,bZ0,bZg,2(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{g,b}\cup Z^{\prime}_{0,b}\cup Z^{\prime}_{g,2}. Then, αβ+(bq|g(zi)bq|g(zj))(sb(zi)sb(zj))bq|g(zi)bq|g(zj)bq|g(zi)(sg(zi)sg(zj))bq|g(zj)(sg(zi)sg(zj))αβ\alpha\geq\beta+(b^{*}_{q|g}(z_{i}^{\circ})-b^{*}_{q|g}(z_{j}^{\circ}))(s^{\prime}_{b}(z_{i}^{\circ})-s^{\prime}_{b}(z_{j}^{\circ}))\iff b^{*}_{q|g}(z_{i}^{\circ})\geq b^{*}_{q|g}(z_{j}^{\circ})\implies b^{*}_{q|g}(z_{i}^{\circ})(s^{\prime}_{g}(z_{i}^{\circ})-s^{\prime}_{g}(z_{j}^{\circ}))\geq b^{*}_{q|g}(z_{j}^{\circ})(s^{\prime}_{g}(z_{i}^{\circ})-s^{\prime}_{g}(z_{j}^{\circ}))\iff\alpha\geq\beta since (sg(zi)sg(zj))0(s^{\prime}_{g}(z_{i}^{\circ})-s^{\prime}_{g}(z_{j}^{\circ}))\geq 0 for (zi,zj)Zg,bZ0,bZg,2(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{g,b}\cup Z^{\prime}_{0,b}\cup Z^{\prime}_{g,2}. But this is a contradiction with α<β\alpha<\beta, thus (zi,zj)Zg,bZ0,bZg,2(z_{i}^{\circ},z_{j}^{\circ})\notin Z^{\prime}_{g,b}\cup Z^{\prime}_{0,b}\cup Z^{\prime}_{g,2}
Case 3 (L.4.1.3): (zi,zj)Z2,0(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{2,0}. Then, (sb(zi)sb(zj))(bq|g(zi)bq|g(zj))=bq|g(zi))bq|g(zj))0(s^{\prime}_{b}(z_{i}^{\circ})-s^{\prime}_{b}(z_{j}^{\circ}))(b^{*}_{q|g}(z_{i}^{\circ})-b^{*}_{q|g}(z_{j}^{\circ}))=b^{*}_{q|g}(z_{i}^{\circ}))-b^{*}_{q|g}(z_{j}^{\circ}))\geq 0 since C2,0(bq|g)=C_{2,0}(b^{*}_{q|g})=\emptyset. Equivalently, αβ\alpha\geq\beta, which is a contradiction with α<β\alpha<\beta, thus (zi,zj)Z2,0(z_{i}^{\circ},z_{j}^{\circ})\notin Z^{\prime}_{2,0}. Thus, we have shown that (zi,zj)𝒵×𝒵bq|gAq|g(z_{i}^{\circ},z_{j}^{\circ})\notin\mathcal{Z}\times\mathcal{Z}\implies b^{*}_{q|g}\notin A^{*}_{q|g}.
Part 2: Next we show that bq|gAq|gbq|gBq|gb^{*}_{q|g}\in A^{*}_{q|g}\implies b^{*}_{q|g}\in B^{*}_{q|g}. bq|gAq|g(zi,zj)𝒵×𝒵,α=sg(zi)bq|g(zi)+sg(zj)bq|g(zj)sg(zi)bq|g(zj)+sg(zj)bq|g(zi)=βb^{*}_{q|g}\in A^{*}_{q|g}\implies\forall(z_{i},z_{j})\in\mathcal{Z}\times\mathcal{Z},\alpha=s^{\prime}_{g}(z_{i})b^{*}_{q|g}(z_{i})+s^{\prime}_{g}(z_{j})b^{*}_{q|g}(z_{j})\geq s^{\prime}_{g}(z_{i})b^{*}_{q|g}(z_{j})+s^{\prime}_{g}(z_{j})b^{*}_{q|g}(z_{i})=\beta. Thus, (zi,zj)Z2,0\forall(z_{i}^{\circ},z_{j}^{\circ})\in Z^{\prime}_{2,0}, α=bq|g(zi)bq|g(zj)=βC2,0(bq|g)=bq|gBq|g\alpha=b^{*}_{q|g}(z_{i}^{\circ})\geq b^{*}_{q|g}(z_{j}^{\circ})=\beta\implies C_{2,0}(b^{*}_{q|g})=\emptyset\implies b^{*}_{q|g}\in B^{*}_{q|g}.
Having shown both sides of the implication, we conclude Aq|g=Bq|gA^{*}_{q|g}=B^{*}_{q|g}.

Proposition 2.

(Maximal Sufficient Explainable Performance) q𝒬,aqAq|g\forall q\in\mathcal{Q},a^{\prime}_{q}\in A^{*}_{q|g} where Aq|g={aq|g:aq|g=argmaxaqAqt¯g(aq)}A^{*}_{q|g}=\{a^{*}_{q|g}\colon a^{*}_{q|g}=\arg\max_{a^{*}_{q}\in A^{*}_{q}}\bar{t}^{\prime}_{g}(a^{*}_{q})\} and t¯g(aq)=1ni=1ntg(aq,zi)=1ni=1nsg(zi)aq(zi)\bar{t}^{\prime}_{g}(a^{*}_{q})=\frac{1}{n}\sum_{i=1}^{n}t^{\prime}_{g}(a^{*}_{q},z_{i})=\frac{1}{n}\sum_{i=1}^{n}s^{\prime}_{g}(z_{i})a^{*}_{q}(z_{i})

Proof.

From Lemma 4 it is sufficient to show that aqBq|ga^{\prime}_{q}\in B^{*}_{q|g} or equivalently that aqAqa^{\prime}_{q}\in A^{*}_{q} and that (zi,zj)Z2×Z0,aq(zi)aq(zj)\forall(z_{i},z_{j})\in Z^{\prime}_{2}\times Z^{\prime}_{0},a^{\prime}_{q}(z_{i})\geq a^{\prime}_{q}(z_{j}).
From Proposition 1, aqAqa^{\prime}_{q}\in A^{*}_{q}.
From Lemma 1, r~(zi)>r~(zj)rankDn(r~(zi))>rankDn(r~(zj))𝕀{rankDn(r~(zi))>n1q}𝕀{rankDn(r~(zj))>n1q}aq(zi)aq(zj)\tilde{r}^{\prime}(z_{i})>\tilde{r}^{\prime}(z_{j})\iff\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z_{i}))>\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z_{j}))\implies\mathbb{I}\{\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z_{i}))>n_{1-q}\}\geq\mathbb{I}\{\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z_{j}))>n_{1-q}\}\iff a^{\prime}_{q}(z_{i})\geq a^{\prime}_{q}(z_{j}). Thus, aqAqa^{\prime}_{q}\in A^{**}_{q}. ∎

Proposition 3.

(Monotone Allocation) qi<qj𝒬\forall q_{i}<q_{j}\in\mathcal{Q}, {z:z𝒵,aqi(z)=1}{z:z𝒵,aqj(z)=1}\{z:z\in\mathcal{Z},a^{\prime}_{q_{i}}(z)=1\}\subseteq\{z:z\in\mathcal{Z},a^{\prime}_{q_{j}}(z)=1\}

Proof.

From Lemma 2, {z:z𝒵,aqi(z)=1}={z:z𝒵,rankDn(r~(z))>n1qi}\{z:z\in\mathcal{Z},a^{\prime}_{q_{i}}(z)=1\}=\{z:z\in\mathcal{Z},\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z))>n_{1-q_{i}}\}. Since qi<qj,n1qi>n1qjq_{i}<q_{j},n_{1-q_{i}}>n_{1-q_{j}}, thus {z:z𝒵,aqi(z)=1}={z:z𝒵,rankDn(r~(z))>n1qi}{z:z𝒵,rankDn(r~(z))>n1qj}={z:z𝒵,aqj(z)=1}\{z:z\in\mathcal{Z},a^{\prime}_{q_{i}}(z)=1\}=\{z:z\in\mathcal{Z},\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z))>n_{1-q_{i}}\}\subseteq\{z:z\in\mathcal{Z},\mathrm{rank}_{D^{n}}(\tilde{r}^{\prime}(z))>n_{1-q_{j}}\}=\{z:z\in\mathcal{Z},a^{\prime}_{q_{j}}(z)=1\}. ∎

Dataset Task Observations Features Epochs Tree Regr. GBT DNN
Wine classification 2,554 11 800 73.9 72.1 80.5 79.5
Phoneme classification 3,172 5 800 85.3 73.3 88.5 87.1
KDDIPUMS classification 5,188 20 800 88.3 85.8 88.0 83.6
EyeMovements classification 7,608 20 800 58.5 53.7 68.2 56.6
Pol classification 10,082 26 400 97.1 86.9 98.4 98.5
Bank classification 10,578 7 400 79.0 73.6 79.7 75.6
MagicTelescope classification 13,376 10 400 81.7 77.6 85.9 84.4
House16H classification 13,488 16 400 84.7 83.2 89.2 86.2
Credit classification 16,714 10 400 77.1 73.9 77.7 71.4
California classification 20,634 8 400 85.5 82.4 90.6 87.8
Electricity classification 38,474 7 400 86.8 74.6 92.6 81.5
Jannis classification 57,580 54 400 74.6 72.4 79.3 77.6
MiniBooNE classification 72,998 50 400 89.8 84.6 93.9 93.0
WineR regression 6,497 11 800 0.25 0.25 0.20 0.23
IsoletR regression 7,797 613 800 0.39 0.38 0.25 0.16
CPUR regression 8,192 21 800 0.06 0.19 0.04 0.05
SulfurR regression 10,081 6 400 0.06 0.08 0.05 0.03
BrazilianHousesR regression 10,692 8 400 0.02 0.09 0.01 0.01
AileronsR regression 13,750 33 400 0.11 0.10 0.09 0.09
MiamiHousingR regression 13,932 14 400 0.12 0.17 0.08 0.08
PolR regression 15,000 26 400 0.15 0.61 0.09 0.07
ElevatorsR regression 16,599 16 400 0.10 0.09 0.07 0.06
BikeSharingR regression 17,379 6 400 0.22 0.30 0.21 0.21
FifaR regression 18,063 5 400 0.25 0.34 0.23 0.24
CaliforniaR regression 20,640 8 400 0.23 0.27 0.16 0.18
HousesR regression 20,640 8 400 0.17 0.20 0.13 0.13
SuperconductR regression 21,263 79 400 0.14 0.21 0.10 0.11
HouseSalesR regression 21,613 15 400 0.10 0.12 0.08 0.08
House16HR regression 22,784 16 400 0.10 0.11 0.08 0.10
DiamondsR regression 53,940 6 400 0.12 0.14 0.12 0.11
MedicalChargesR regression 163,065 5 100 0.04 0.12 0.04 0.04
Table 4: This table summarizes the datasets used in the experiments - their number of observations and features and the number of training epochs used to learn the neural networks for each. Also included are the test accuracies (for classification tasks) and RMSE (for regression tasks) for each ensemble component model with best performance bolded.
Model L1 penalty β\beta min split max leaf max depth
Logistic regression 2i2^{i} for ii in range(-10,10) - - -
Linear regression 2i2^{i} for ii in range(-10,10) - - -
Classification tree - 2i2^{i} for ii in range(1,12) 2i2^{i} for ii in range(1,12) 2i2^{i} for ii in range(1,12)
Regression tree - 2i2^{i} for ii in range(1,12) 2i2^{i} for ii in range(1,12) 2i2^{i} for ii in range(1,12)
Table 5: This table summarizes the hyperparameter values searched by model type for all glass box models.
Model learning rate n estimators max depth subsample
GBT classifier 0.001, 0.01, 0.1 2i2^{i}, ii in range(4,10) 2i2^{i}, ii in range(3,7) (2,4,6,8,10)*0.1
GBT regressor 0.001, 0.01, 0.1 2i2^{i}, ii in range(4,10) 2i2^{i}, ii in range(3,7) (2,4,6,8,10)*0.1
GBT allocator 0.01, 0.1 2i2^{i}, ii in range(2,11,2) 2i2^{i}, ii in range(2,6) (25,50,75,100)*0.01
Table 6: This table summarizes the hyperparameter values searched by model type for all gradient boosting trees models.
Model L2 penalty β\beta learning rate lr schedule optimizer
WRN regressor 0, 1e-7, 1e-5, 1e-3 (1e-5, 1e-4, 1e-3), (1e-4, 1e-3, 1e-2) constant, cosine Adam, SGD+Mtm
WRN allocator 0, 1e-7, 1e-5, 1e-3 (1e-5, 1e-4, 1e-3), (1e-4, 1e-3, 1e-2) constant, cosine Adam, SGD+Mtm
Table 7: This table summarizes the hyperparameter values searched by model type for all neural network models.
Refer to caption
Figure 3: This figure provides the sufficient accuracy values of the allocated EEG ensemble for each explainability level and every dataset using the neural network allocator.
Refer to caption
Figure 4: This figure provides the sufficient accuracy values of the allocated EEG ensemble for each explainability level and every dataset using the gradient boosting trees allocator.