This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\savesymbol

Bbbk \savesymbolopenbox \restoresymbolNEWBbbk \restoresymbolNEWopenbox

Provably Robust and Plausible Counterfactual Explanations for
Neural Networks via Robust Optimisation

Junqi Jiang1, Jianglin Lan2, Francesco Leofante1, Antonio Rago1, Francesca Toni1
1 Department of Computing, Imperial College London
2 James Watt School of Engineering, University of Glasgow
{junqi.jiang, f.leofante, a.rago, f.toni}@imperial.ac.uk, [email protected]
Abstract

Counterfactual Explanations (CEs) have received increasing interest as a major methodology for explaining neural network classifiers. Usually, CEs for an input-output pair are defined as data points with minimum distance to the input that are classified with a different label than the output. To tackle the established problem that CEs are easily invalidated when model parameters are updated (e.g. retrained), studies have proposed ways to certify the robustness of CEs under model parameter changes bounded by a norm ball. However, existing methods targeting this form of robustness are not sound or complete, and they may generate implausible CEs, i.e., outliers wrt the training dataset. In fact, no existing method simultaneously optimises for proximity and plausibility while preserving robustness guarantees. In this work, we propose Provably RObust and PLAusible Counterfactual Explanations (PROPLACE)111The implementation is available at https://github.com/junqi-jiang/proplace, a method leveraging on robust optimisation techniques to address the aforementioned limitations in the literature. We formulate an iterative algorithm to compute provably robust CEs and prove its convergence, soundness and completeness. Through a comparative experiment involving six baselines, five of which target robustness, we show that PROPLACE achieves state-of-the-art performances against metrics on three evaluation aspects.

1 Introduction

Counterfactual Explanations (CEs) have become a major methodology to explain NNs due to their simplicity, compliance with the regulations Wachter et al. (2017), and alignment with human thinking Celar and Byrne (2023). Given an input point to a classifier, a CE is a modified input classified with another, often more desirable, label. Consider a customer that is denied a loan by the machine learning system of a bank. A CE the bank provided for this customer could be, the loan application would have been approved, had you raised your annual salary by $ 6000. Several desired properties of CEs have been identified in the literature, the most fundamental of which is validity, requiring that the CE needs to be correctly classified with a specified label Tolomei et al. (2017). Proximity refers to the closeness between the CE and the input measured by some distance metric, which translates to a measure of the effort the end user has to make to achieve the prescribed changes Wachter et al. (2017). The CEs should also lie on the data manifold of the training dataset and not be an outlier, which is assessed via plausibility Poyiadzi et al. (2020). Most recently, the robustness of CEs, amounting to their validity under various types of uncertainty, has drawn increasing attention due to its real-world importance. In this work, we consider robustness to the model parameter changes occurring in the classifier on which the CE was generated. Continuing the loan example, assume the bank’s machine learning model is retrained with new data, while, in the meantime, the customer has achieved a raise in salary (as prescribed by the CE). The customer may then return to the bank only to find that the previously specified CE is now invalidated by the new model. In this case, the bank could be seen as being responsible by the user and could potentially be legally liable, risking financial and reputational damage to the organisation. The quality of such unreliable CE is also questionable: Rawal et al. (2020); Dutta et al. (2022) have shown that CEs found by existing non-robust methods are prone to such invalidation due to their closeness to the decision boundary.

Various methods have been proposed to tackle this issue. Nguyen et al. (2022); Dutta et al. (2022); Hamman et al. (2023) focus on building heuristic methods using model confidence, Lipschitz continuity, and quantities related to the data distribution. Upadhyay et al. (2021); Black et al. (2022); Jiang et al. (2023) consider optimising the validity of CEs under bounded model parameter changes, which are also empirically shown to be robust to the unbounded parameter changes scenarios. Among the existing methods, only Jiang et al. (2023) provides robustness guarantees in a formal approach, which are known to be lacking in the explainable AI (XAI) literature in general, aside from some notable examples, e.g. as introduced in Marques-Silva and Ignatiev (2022). Their method generates such provably robust CEs via iteratively tuning the hyperparameters of an arbitrary non-robust CEs method and testing for robustness. However, this method cannot always guarantee soundness and is not complete, which is also the case for the method in Upadhyay et al. (2021). Another limitation in the current literature is that the methods targeting this form of robustness guarantee do not find plausible CEs, limiting their practical applicability.

Such limitations have motivated this work. After discussing relevant studies in Section 2, we introduce the robust optimisation problem for computing CEs with proximity property as the objective, and robustness and plausibility properties as constraints (Section 3). In Section 4, we then present Provably RObust and PLAusible CEs (PROPLACE), a method leveraging on robust optimisation techniques to address the limitation in the literature that no method optimises for proximity and plausibility while providing formal robustness guarantees. We show the (conditional) soundness and completeness of our method, and give a bi-level optimisation procedure that will converge and terminate. Finally, in our experiments, we compare PROPLACE with six existing CE methods, five of which target robustness, on four benchmark datasets. The results show that our method achieves the best robustness and plausibility, while demonstrating superior proximity among the most robust baselines.

2 Related Work

As increasing interest has been focused on XAI, a plethora of CE generation methods have been proposed (see Guidotti (2022) for a recent overview). Given our focus on neural networks, we cover those explaining the outputs of these models. Wachter et al. (2017) proposed a gradient-based optimisation method targeting the validity and proximity of CEs. Similarly, using the mixed integer programming (MILP) representation of neural networks, Mohammadi et al. (2021) formulated the CEs search into a constrained optimisation problem such that the resulting CEs are guaranteed to be valid. Mothilal et al. (2020) advocated generating a diverse set of CEs for each input to enrich the information provided to the explainee. Several works also addressed actionability constraints Ustun et al. (2019); Verma et al. (2022); Vo et al. (2023), only allowing changes in the actionable features of real users. Poyiadzi et al. (2020) proposed a graph-based method to find a path of CEs that are all lying within the data manifold. Several other works have proposed to use (variational) auto-encoders or nearest neighbours to induce plausibility Dhurandhar et al. (2018); Pawelczyk et al. (2020a); Van Looveren and Klaise (2021). Among these properties, actionability and plausibility are two orthogonal considerations which make the CEs realistic in practice, and trade-offs have been identified between plausibility and proximity Pawelczyk et al. (2020b).

In this work, our focus is on the property of robustness to changes in the model parameters, i.e. weights and biases in the underlying classifier. Several studies looked at CEs under bounded model parameter changes of a neural network: Upadhyay et al. (2021) formulated a novel loss function and solved using gradient-based methods. Black et al. (2022) proposed a heuristic based on the classifier’s Lipschitz constant and the model confidence to search for robust CEs. Jiang et al. (2023) used interval abstractions Prabhakar and Afzal (2019) to certify the robustness against bounded parameter changes, and embed the certification process into existing CE methods. Differently to our approach, these methods do not generate plausible CEs or guarantee that provably robust CEs are found. Other relevant works place their focus on the robustness of CEs against unbounded model changes. Ferrario and Loi (2022) took the approach of augmenting the training data with previously generated CEs. Nguyen et al. (2022) focused on the data distribution and formulated the problem as posterior probability ratio minimisation to generate robust and plausible CEs. By using first- and second-moment information, Bui et al. (2022) proposed lower and upper bounds on the CEs’ validity under random parameter updates and generated robust CEs using gradient descent. Dutta et al. (2022) defined a novel robustness measure based on the model confidences over the neighbourhood of the CE, and used dataset points that satisfy some robustness test to find close and plausible CEs. Their notion is then further re-calibrated for neural networks with probabilistic robustness guarantees in Hamman et al. (2023). Trade-offs between robustness and proximity were discussed by Pawelczyk et al. (2022) and Upadhyay et al. (2021).

Other forms of CEs’ robustness have also been investigated, for example, robustness against: input perturbations Alvarez-Melis and Jaakkola (2018); Sharma et al. (2020); Bajaj et al. (2021); Dominguez-Olmedo et al. (2022); Huai et al. (2022); Virgolin and Fracaros (2023); Zhang et al. (2023); noise in the execution of CEs Pawelczyk et al. (2022); Leofante and Lomuscio (2023b, a); Maragno et al. (2023); and model multiplicity Pawelczyk et al. (2020b); Leofante et al. (2023).

3 Preliminaries and Problem Statement

Notation.

Given an integer kk, we use [k][k] to denote the set {1,,k}\{1,\ldots,k\}. We use |S|\lvert S\rvert to denote the cardinality of a set SS.

Neural Network (NN).

We denote a NN as Θ:𝒳d𝒴\mathcal{M}_{\Theta}:\mathcal{X}\subseteq\mathbb{R}^{d}\rightarrow\mathcal{Y}\subseteq\mathbb{N}, where the inputs are dd-dimensional vectors and the outputs are discrete class labels. Θ\Theta represents the collection of parameters that characterise the NN. Throughout the paper, we will illustrate our method using the binary classification case (i.e. 𝒴={0,1}\mathcal{Y}=\{0,1\}), though the method is readily applicable to multi-class classification. Let Θ(x)\mathcal{M}_{\Theta}(x) also (with an abuse of notation) refer to the pre-sigmoid (logit) value in the NN. Then, for an input x𝒳x\in\mathcal{X}, we say Θ\mathcal{M}_{\Theta} classifies xx as class 11 if Θ(x)0\mathcal{M}_{\Theta}(x)\geq 0, otherwise Θ\mathcal{M}_{\Theta} classifies xx as class 0.

Counterfactual Explanation (CE).

For an input x𝒳x\in\mathcal{X} that is classified to the unwanted class 0 (assumed throughout the paper), a CE x𝒳x^{\prime}\in\mathcal{X} is some other data point “similar” to the input, e.g. by some distance measure, but classified to the desired class 11.

Definition 1.

(CE) Given a NN Θ\mathcal{M}_{\Theta}, an input x𝒳x\in\mathcal{X} such that Θ(x)<0\mathcal{M}_{\Theta}(x)<0, and a distance metric dist:d×d+dist:\mathbb{R}^{d}\times\mathbb{R}^{d}\rightarrow\mathbb{R}^{+}, a CE x𝒳x^{\prime}\in\mathcal{X} is such that:

argminx\displaystyle\operatorname*{arg\,min}_{x^{\prime}} dist(x,x)\displaystyle\text{ }dist(x,x^{\prime})
subject to Θ(x)0\displaystyle\quad\mathcal{M}_{\Theta}(x^{\prime})\geq 0

The minimum distance objective targets the minimum effort by the end user to achieve a change, which corresponds to the basic requirement of proximity mentioned in Section 1. In the literature, normalised L1L_{1} distance is often adopted as the distance metric because it induces changes in fewer features in the CE Wachter et al. (2017). However, methods that find such plain CEs usually result in unrealistic combinations of features, or outliers to the underlying data distribution of the training dataset. A plausible CE avoids these issues and is formally defined as follows:

Definition 2.

(Plausible CE) Given a NN Θ\mathcal{M}_{\Theta} and an input x𝒳x\in\mathcal{X} such that Θ(x)<0\mathcal{M}_{\Theta}(x)<0, a distance metric dist:d×d+dist:\mathbb{R}^{d}\times\mathbb{R}^{d}\rightarrow\mathbb{R}^{+} and some plausible region 𝒳plausd\mathcal{X}_{plaus}\subseteq\mathbb{R}^{d}, a plausible CE is an xx^{\prime} such that:

argminx\displaystyle\operatorname*{arg\,min}_{x^{\prime}} dist(x,x)\displaystyle\text{ }dist(x,x^{\prime})
subject to Θ(x)0,x𝒳plaus\displaystyle\quad\mathcal{M}_{\Theta}(x^{\prime})\geq 0,\quad x^{\prime}\in\mathcal{X}_{plaus}

The plausible region 𝒳plaus\mathcal{X}_{plaus} may be used to eliminate any unrealistic feature values (e.g. a value of 0.95 for a discrete feature), or to indicate a densely populated region that is close to the data manifold of the training dataset. Additionally, it may also include some actionability considerations, such as restricting immutable attributes (e.g. avoiding suggesting changes in gender) or specifying some relations between input features (e.g. obtaining a doctoral degree should also cost the user at least 4 years).

Robustness of Counterfactual Explanations.

Studies have shown that CEs found by the above formulations are readily invalidated when small changes occur in the model parameters of the NNs. We formalise this in the following and begin by introducing a distance measure between two NNs and a definition of model shift. Note that Definitions 3 to 7 are adapted from Jiang et al. (2023).

Definition 3.

(Distance between two NNs) Consider two NNs Θ\mathcal{M}_{\Theta}, Θ\mathcal{M}_{\Theta^{\prime}} of the same architecture characterised by parameters Θ\Theta and Θ\Theta^{\prime}. For 0p0\leq p\leq\infty, the p-distance between Θ\mathcal{M}_{\Theta} and Θ\mathcal{M}_{\Theta^{\prime}} is dp(Θ,Θ)=ΘΘpd_{p}(\mathcal{M}_{\Theta},\mathcal{M}_{\Theta^{\prime}})=\lVert\Theta-\Theta^{\prime}\rVert_{p}.

Definition 4.

(Bounded model shifts) Given a NN Θ\mathcal{M}_{\Theta}, δ>0\delta\in\mathbb{R}_{>0} and 0p0\leq p\leq\infty, the set of bounded model shifts is defined as Δ={Θdp(Θ,Θ)δ}\Delta=\{\mathcal{M}_{\Theta^{\prime}}\mid d_{p}(\mathcal{M}_{\Theta},\mathcal{M}_{\Theta^{\prime}})\leq\delta\}.

Certifying Robustness.

Having presented the definitions required to formalise the optimisation problem for finding provably robust and plausible CEs, we now introduce another relevant technique that uses interval abstractions to certify the robustness of CEs. We refer to the certification process as the Δ\Delta-robustness test; this will be used for parts of our method and also as an evaluation metric in the experiments. We assume p=p=\infty for bounded model shifts Δ\Delta throughout the paper.

Definition 5.

(Interval abstraction of NN) Consider a NN Θ\mathcal{M}_{\Theta} with Θ=[θ0,,θd]\Theta=[\theta_{0},\ldots,\theta_{d}]. Given a set of bounded model shifts Δ\Delta, we define the interval abstraction of Θ\mathcal{M}_{\Theta} under Δ\Delta as the model (Θ,Δ):𝒳\powerset\mathcal{I}_{(\Theta,\Delta)}:\mathcal{X}\rightarrow\powerset{\mathbb{R}} (for \powerset\powerset{\mathbb{R}} the set of all closed intervals over \mathbb{R}) such that:

  • Θ\mathcal{M}_{\Theta} and (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} have the same architecture;

  • (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} is parameterised by an interval-valued vector 𝚯=[𝜽0,,𝜽d]\bm{\Theta}\!\!=\!\![\bm{\theta}_{0},\!\ldots,\!\bm{\theta}_{d}] such that, for i{0,,d}i\in\{0,\ldots,d\}, 𝜽i=[θiδ,θi+δ]\bm{\theta}_{i}=[\theta_{i}-\delta,\theta_{i}+\delta], where δ\delta is the bound in Δ\Delta.

When p=p=\infty, 𝜽i\bm{\theta}_{i} encodes the range of possible model parameter changes by the application of Δ\Delta to Θ\mathcal{M}_{\Theta}. Given a fixed input, by propagating the weight and bias intervals, the output range of (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} exactly represents the possible output range for the input by applying Δ\Delta to Θ\mathcal{M}_{\Theta} Jiang et al. (2023).

Definition 6.

(Interval abstraction of NN classification) Let (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} be the interval abstraction of a NN Θ\mathcal{M}_{\Theta} under Δ\Delta. Given an input x𝒳x\in\mathcal{X}, let (Θ,Δ)(x)=[l,u]\mathcal{I}_{(\Theta,\Delta)}(x)=[l,u]. Then, we say that (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} classifies xx as class 11 if l0l\geq 0 (denoted, with an abuse of notation, (Θ,Δ)(x)0\mathcal{I}_{(\Theta,\Delta)}(x)\geq 0), and as class 0 if u<0u<0 (denoted, with an abuse of notation, (Θ,Δ)(x)<0\mathcal{I}_{(\Theta,\Delta)}(x)<0).

Indeed, for an input, if the lower bound ll of pre-sigmoid output node interval [l,u][l,u] of (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} satisfies l0l\geq 0, then it means all shifted models in Δ\Delta would predict the input with a pre-sigmoid value that is greater than or equal to 0, all resulting in predicted label 1. We apply this intuition to the CE context:

Definition 7.

(Δ\Delta-robust CE) Consider an input x𝒳x\in\mathcal{X} and a model Θ\mathcal{M}_{\Theta} such that Θ(x)<0\mathcal{M}_{\Theta}{(x)}<0. Let (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} be the interval abstraction of Θ\mathcal{M}_{\Theta} under Δ\Delta. We say that a CE xx^{\prime} is Δ\Delta-robust iff (Θ,Δ)(x)0\mathcal{I}_{(\Theta,\Delta)}(x^{\prime})\geq 0.

Checking whether a CE xx^{\prime} is Δ\Delta-robust requires the calculation of the lower bound ll of the pre-sigmoid output node interval [l,u][l,u] of (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)}. This process can be encoded as a MILP program (see Appendix B in Jiang et al. (2023)).

Optimising for Robustness and Plausibility.

Now we introduce the targeted provably robust and plausible optimisation problem based on Definitions 2 and 7, by taking inspiration from the robust optimisation technique Ben-Tal et al. (2009).

Definition 8.

(Provably robust and plausible CE) Given a NN Θ\mathcal{M}_{\Theta}, an input x𝒳x\in\mathcal{X} such that Θ(x)<0\mathcal{M}_{\Theta}(x)<0, a distance metric dist:d×d+dist:\mathbb{R}^{d}\times\mathbb{R}^{d}\rightarrow\mathbb{R}^{+} and some plausible region 𝒳plausd\mathcal{X}_{plaus}\subseteq\mathbb{R}^{d}, let (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} be the interval abstraction of Θ\mathcal{M}_{\Theta} under the bounded model shifts Δ\Delta. Then, a provably robust and plausible CE x𝒳x^{\prime}\in\mathcal{X} is such that:

argminx\displaystyle\operatorname*{arg\,min}_{x^{\prime}} dist(x,x)\displaystyle\text{ }dist(x,x^{\prime}) (1a)
subject to (Θ,Δ)(x)0,\displaystyle\quad\mathcal{I}_{(\Theta,\Delta)}(x^{\prime})\geq 0, (1b)
x𝒳plaus\displaystyle\quad x^{\prime}\in\mathcal{X}_{plaus} (1c)

The optimisation problem (1) can be equivalently rewritten as follows:

argminx\displaystyle\operatorname*{arg\,min}_{x^{\prime}} dist(x,x)\displaystyle\text{ }dist(x,x^{\prime}) (2a)
subject to maxΘΔ[Θ(x)]0,\displaystyle\quad\max_{\mathcal{M}_{\Theta^{\prime}}\in\Delta}[-\mathcal{M}_{\Theta^{\prime}}(x^{\prime})]\leq 0, (2b)
x𝒳plaus\displaystyle\quad x^{\prime}\in\mathcal{X}_{plaus} (2c)

We show next a novel approach for solving this robust optimisation problem (2).

4 PROPLACE

Algorithm 1 PROPLACE
1:input xx, model Θ\mathcal{M}_{\Theta},
2:    training dataset 𝒟={(x1,y1),,(xn,yn)}\mathcal{D}=\{(x_{1},y_{1}),\ldots,(x_{n},y_{n})\},
3:    set of bounded model shifts Δ\Delta,
4:    plausible region to be used as CEs search space 𝒳plaus\mathcal{X}_{plaus}.
5:Init: xx^{\prime}\leftarrow\emptyset; Δ{Θ}\Delta^{\prime}\leftarrow{\{\mathcal{M}_{\Theta}\}}
6:Repeat until (Θ(x))0(-\mathcal{M}_{\Theta^{\prime}}(x^{\prime}))\leq 0xOuter_minimisation(Θ,x,𝒳plaus,Δ)x^{\prime}\leftarrow\texttt{Outer\_minimisation}(\mathcal{M}_{\Theta},x,\mathcal{X}_{plaus},\Delta^{\prime})ΘInner_maximisation(x,Δ,Δ)\mathcal{M}_{\Theta^{\prime}}\leftarrow\texttt{Inner\_maximisation}(x^{\prime},\Delta^{\prime},\Delta)ΔΔ{Θ}\Delta^{\prime}\leftarrow\Delta^{\prime}\cup\{\mathcal{M}_{\Theta^{\prime}}\}
7:return xx^{\prime}

The procedure for computing robust and plausible CEs, solving the optimisation problem (2) , is summarised in Algorithm 1. We will first introduce how the plausible region 𝒳plaus\mathcal{X}_{plaus} is constructed in Section 4.1 (corresponding to Line 3, Algorithm 1). Then, in Section 4.2 we will present the bi-level optimisation method (corresponding to Lines 4-5, Algorithm 1) to solve the robust optimisation problem (2). In Section 4.2 we will also instantiate the complete bi-level optimisation formulations (in MILP form) of our method for NNs with ReLU activation functions. Finally, in Section 4.3 we discuss the soundness and completeness of Algorithm 1 and prove its convergence.

4.1 Identifying Search Space 𝒳plaus\mathcal{X}_{plaus}

As mentioned in Section 2, points from the training dataset (especially kk-nearest-neighbours) are frequently utilised in the literature to induce plausibility. In this work, we propose to use a more specialised kind of dataset point, kk Δ\Delta-robust nearest-neighbours, to construct the search space for CEs that is both plausible and robust.

Definition 9.

(k Δ\Delta-robust nearest-neighbours) Given a NN Θ\mathcal{M}_{\Theta} and an input x𝒳x\in\mathcal{X} such that Θ(x)<0\mathcal{M}_{\Theta}(x)<0, a distance metric dist:d×d+dist:\mathbb{R}^{d}\times\mathbb{R}^{d}\rightarrow\mathbb{R}^{+}, a dataset 𝒟d\mathcal{D}\subseteq\mathbb{R}^{d} on which Θ\mathcal{M}_{\Theta} is trained, and a set of bounded model shifts of interest Δ\Delta, let (Θ,Δ)\mathcal{I}_{(\Theta,\Delta)} be the interval abstraction of Θ\mathcal{M}_{\Theta} under Δ\Delta. Then, the k Δ\Delta-robust nearest-neighbours of xx is a set Sk,Δ𝒟S_{k,\Delta}\subseteq\mathcal{D} with cardinality |Sk,Δ|=k|S_{k,\Delta}|=k such that:

  • xSk,Δ\forall x^{\prime}\in S_{k,\Delta}, xx^{\prime} is Δ\Delta-robust, i.e. (Θ,Δ)(x)0\mathcal{I}_{(\Theta,\Delta)}(x^{\prime})\geq 0,

  • x′′𝒟Sk,Δ\forall x^{{}^{\prime\prime}}\in\mathcal{D}\setminus S_{k,\Delta}, if x′′x^{{}^{\prime\prime}} is Δ\Delta-robust, dist(x,x′′)maxxSk,Δ dist(x,x)dist(x,x^{{}^{\prime\prime}})\geq\underset{x^{\prime}\in S_{k,\Delta}}{\emph{max}}\text{ }dist(x,x^{\prime}).

The first constraint enforces the Δ\Delta-robustness, and the second states that the points contained in the set are the kk nearest points to the input xx amongst all the Δ\Delta-robust dataset points. In practice, in order to compute the kk Δ\Delta-robust nearest-neighbours, we fit a k-d tree on the dataset points that are classified to the desired class, then iteratively query the k-d tree for the nearest neighbour of an input, until the result satisfies the Δ\Delta-robustness test (Definition 7).

Restricting the CE search space within the convex hull of these robust neighbours will likely induce high plausibility (and robustness). However, because these points are deep within parts of the training dataset that are classified to another class, they may be far from the model’s decision boundary, therefore resulting in large distances to the inputs. In fact, Dutta et al. (2022); Hamman et al. (2023) adopted similar robust nearest neighbours (using other notions of robustness tests) as the final CEs, and poor proximity was observed in their experiment results. They have also shown that finding CEs using line search between proximal CEs and these robust neighbours can slightly improve proximity.

In our case, since the validity of the CEs can be guaranteed from the optimisation procedures (Section 4.2), we expand the plausible search space across the decision boundary by taking the input into consideration, which is assumed to also be inside the data distribution.

Definition 10.

(Plausible region) Given an input xdx\in\mathbb{R}^{d} and its k Δ\Delta-robust nearest neighbours Sk,ΔS_{k,\Delta}, the plausible region 𝒳plaus\mathcal{X}_{plaus} is the convex hull of Sk,Δ{x}S_{k,\Delta}\cup\{x\}.

By restricting the CE search space to such convex hull, the method has the flexibility to find close CEs (with xx as a vertex), or robust and plausible CEs (with the robust neighbours as other vertices). This 𝒳plaus\mathcal{X}_{plaus} ensures the soundness and completeness of our method (Section 4.3).

4.2 Bi-level Optimisation Method with MILP

4.2.1 Outer and Inner Optimisation problems

We separate the robust optimisation problem (2) to solve into outer minimisation and inner maximisation problems, as specified in Definitions 3 and 12.

Definition 11.

Given a NN Θ\mathcal{M}_{\Theta} and an input x𝒳x\in\mathcal{X} such that Θ(x)<0\mathcal{M}_{\Theta}(x)<0, a distance metric dist:d×d+dist:\mathbb{R}^{d}\times\mathbb{R}^{d}\rightarrow\mathbb{R}^{+} and some plausible region 𝒳plausd\mathcal{X}_{plaus}\subseteq\mathbb{R}^{d}, let Δ\Delta^{\prime} be a set of shifted models. Then, the outer minimisation problem finds a CE xx^{\prime} such that:

argminx\displaystyle\operatorname*{arg\,min}_{x^{\prime}} dist(x,x)\displaystyle\text{ }dist(x,x^{\prime}) (3a)
subject to Θ(x)0, for each ΘΔ,\displaystyle\quad-\mathcal{M}_{\Theta^{\prime}}(x^{\prime})\leq 0,\text{ for each }\mathcal{M}_{\Theta^{\prime}}\in\Delta^{\prime}, (3b)
x𝒳plaus\displaystyle\quad x^{\prime}\in\mathcal{X}_{plaus} (3c)
Definition 12.

Given a CE x𝒳x^{\prime}\in\mathcal{X} found by the outer minimisation problem, the set of bounded model shifts Δ\Delta, the inner maximisation problem finds a shifted model Θ\mathcal{M}_{\Theta^{\prime}} such that:

arg maxΘ\displaystyle\underset{\mathcal{M}_{\Theta^{\prime}}}{\emph{\text{arg max}}} Θ(x)\displaystyle\text{ }-\mathcal{M}_{\Theta^{\prime}}(x^{\prime}) (4a)
subject to ΘΔ\displaystyle\quad\mathcal{M}_{\Theta^{\prime}}\in\Delta (4b)

The outer minimisation problem relaxes the constraint that a CE should be robust to all possible model shifts in the set Δ\Delta; instead, it requires robustness wrt a subset of the model changes ΔΔ\Delta^{\prime}\subset\Delta. Δ\Delta^{\prime} is initialised with the original classification model Θ\mathcal{M}_{\Theta}. At the first execution, the outer minimisation finds the closest CE xx^{\prime} valid for that model. Then, xx^{\prime} is passed to the inner maximisation problem to compute the model shift S(Θ)S({\mathcal{M}_{\Theta})} that produces the lowest model output score. This model shift is considered to be the worst-case perturbation on the model parameters in the set Δ\Delta, and is added to Δ\Delta^{\prime}. In the next iterations, xx^{\prime} is updated to the closest CE valid for all the models in Δ\Delta^{\prime} (outer), which is being expanded (inner), until convergence.

4.2.2 MILP Formulations

The proposed bi-level optimisation method in Section 4.2.1 is independent of specific NN structures. In this section, we take NNs with ReLU activation functions as an example to further elaborate the method. We denote the total number of hidden layers in an NN Θ\mathcal{M}_{\Theta} as hh. We call N0N_{0}, Nh+1N_{h+1}, and NiN_{i} the sets of input, output, and hidden layer nodes for i[h]i\in[h], and their node values are V0V_{0}, Vh+1V_{h+1}, and ViV_{i}. For hidden layer nodes Vi=ReLU(WiVi1+Bi)V_{i}=\text{ReLU}(W_{i}V_{i-1}+B_{i}), and for output layer nodes Vh+1=Wh+1Vh+Bh+1V_{h+1}=W_{h+1}V_{h}+B_{h+1}, where WiW_{i} is the weight matrix connecting nodes at layers i1i-1 and ii, and BiB_{i} is the bias vector of nodes NiN_{i}. We instantiate the formulations using normalised L1L_{1}, while our method PROPLACE can accommodate arbitrary distance metrics.

The outer minimisation problem is equivalent to the following MILP program, where the superscripts jj on weight matrices and bias vectors indicate they are model parameters of the jj-th model ΘjΔ\mathcal{M}_{\Theta}^{j}\in\Delta^{\prime}:

minx,γ,λ\displaystyle\underset{x^{\prime},\gamma,\lambda}{\min} xx1\displaystyle\quad\|x-x^{\prime}\|_{1} (5a)
s.t. V0j=x,\displaystyle V_{0}^{j}=x^{\prime}, (5b)
γij{0,1}|Ni|,i[h],j[|Δ|]\displaystyle\gamma_{i}^{j}\in\{0,1\}^{|N_{i}|},~{}i\in[h],j\in[|\Delta^{\prime}|] (5c)
0VijMγij,i[h],j[|Δ|]\displaystyle 0\leq V_{i}^{j}\leq M\gamma_{i}^{j},~{}i\in[h],j\in[|\Delta^{\prime}|] (5d)
WijVi1j+BijVij(WijVi1j+Bij)+M(1γij),\displaystyle W_{i}^{j}V_{i-1}^{j}+B_{i}^{j}\leq V_{i}^{j}\leq(W_{i}^{j}V_{i-1}^{j}+B_{i}^{j})+M(1-\gamma_{i}^{j}), (5e)
i[h],j[|Δ|]\displaystyle i\in[h],j\in[|\Delta^{\prime}|]
Wh+1jVhj+Bh+1j0,j[|Δ|]\displaystyle W_{h+1}^{j}V_{h}^{j}+B_{h+1}^{j}\geq 0,j\in[|\Delta^{\prime}|] (5f)
λl[0,1],l[|Sk,Δ{x}|],l=1|Sk,Δ{x}|λl=1\displaystyle\lambda_{l}\in[0,1],~{}l\in[|S_{k,\Delta}\cup\{x\}|],~{}\sum_{l=1}^{|S_{k,\Delta}\cup\{x\}|}\lambda_{l}=1
x=l=1|Sk,Δ{x}|λlxl,xlSk,Δ{x}\displaystyle x^{\prime}=\sum_{l=1}^{|S_{k,\Delta}\cup\{x\}|}\lambda_{l}x^{\prime}_{l},x^{\prime}_{l}\in S_{k,\Delta}\cup\{x\} (5g)

Constraints (5b) - (5f) and constraint (5g) correspond respectively to the robustness and plausibility requirement in (3b) - (3c).

The inner maximisation program can be formulated as the following MILP program, where the superscripts 0 on weight matrices and biases indicate they are model parameters of the original model Θ\mathcal{M}_{\Theta}, and δ\delta is the bound of model magnitude change specified in Δ\Delta:

maxV,W,B,γ\displaystyle\underset{V,W,B,\gamma}{\max} Vh+1\displaystyle\quad-V_{h+1} (6a)
s.t. V0=x,\displaystyle V_{0}=x^{\prime}, (6b)
γi{0,1}|Ni|,i[h]\displaystyle\gamma_{i}\in\{0,1\}^{|N_{i}|},~{}i\in[h] (6c)
0ViMγi,i[h]\displaystyle 0\leq V_{i}\leq M\gamma_{i},~{}i\in[h] (6d)
WiVi1+BiVi(WiVi1+Bi)+M(1γi),\displaystyle W_{i}V_{i-1}+B_{i}\leq V_{i}\leq(W_{i}V_{i-1}+B_{i})+M(1-\gamma_{i}), (6e)
i[h]\displaystyle i\in[h]
Vh+1=Wh+1Vh+Bh+1\displaystyle V_{h+1}=W_{h+1}V_{h}+B_{h+1} (6f)
Wi0δWiWi0+δ,i[h+1]\displaystyle W_{i}^{0}-\delta\leq W_{i}\leq W_{i}^{0}+\delta,~{}i\in[h+1] (6g)
Bi0δBiBi0+δ,i[h+1]\displaystyle B_{i}^{0}-\delta\leq B_{i}\leq B_{i}^{0}+\delta,~{}i\in[h+1] (6h)

Due to the flexibility of such MILP programs, the framework accommodates continuous, ordinal, and categorical features Mohammadi et al. (2021). Specific requirements like feature immutability or associations between features can also be encoded Ustun et al. (2019). These MILP problems can be directly solved using off-the-shelf solvers such as Gurobi Gurobi Optimization, LLC (2023).

4.3 Soundness, Completeness and Convergence of Algorithm 1

We now discuss the soundness and completeness of our method by restricting the search space for the CE to the plausible region 𝒳plaus\mathcal{X}_{plaus}. From its definition, the vertices (except the input xx) of 𝒳plaus\mathcal{X}_{plaus} are Δ\Delta-robust, which thus satisfies the robustness requirement of our target problem (Definition 1). This means that there exist at least kk points in the search space satisfying constraint (2c) that also satisfy constraint (2b), making these points feasible solutions for the target problem. We may thus make the following remark:

Proposition 1.

Algorithm 1 is sound and complete if \exists x𝒟x^{\prime}\in\mathcal{D} such that xx^{\prime} is Δ\Delta-robust.

Next, we adapt the method in (Mutapcic and Boyd, 2009, Section 5.2) to provide an upper bound on the maximum number of iterations of Algorithm 1.

Proposition 2.

Given the requirements of Algorithm 1, assume the classifier Θ\mathcal{M}_{\Theta} is Lipschitz continuous in xx^{\prime}. Then, the maximum number of iterations before Algorithm 1 terminates is bounded.

Proof.

Firstly, we assume two small tolerance variables σ>t>0\sigma>t>0 and modify the robustness constraint (2b) of Definition 1 to: maxΘΔ[Θ(x)+σ]t\underset{\mathcal{M}_{\Theta^{\prime}}\in\Delta}{\max}[-\mathcal{M}_{\Theta^{\prime}}(x^{\prime})+\sigma]\leq t, such that the correctness of the robustness guarantee is not affected. The termination condition for Algorithm 1 therefore becomes Θ(x)+σt-\mathcal{M}_{\Theta^{\prime}}(x^{\prime})+\sigma\leq t.

Consider the plausible CE problem (Definition 2 with the validity constraint modified to Θ(x)+σt)-\mathcal{M}_{\Theta}(x^{\prime})+\sigma\leq t), which is the problem solved by the first execution (iteration 11) of the outer minimisation problem in Algorithm 1. We denote its feasible region as \mathcal{F}. Suppose Θ\mathcal{M}_{\Theta} is a ReLU NN without the final (softmax or sigmoid) activation layer, then Θ\mathcal{M}_{\Theta} is Lipschitz continuous. Let f(x,Θ):=Θ(x)+σf(x^{\prime},\mathcal{M}_{\Theta}):=-\mathcal{M}_{\Theta}(x^{\prime})+\sigma, then ff is Lipschitz continuous in xx^{\prime} over \mathcal{F} with some Lipschitz constant LL. For a distance metric dist:d×d+dist:\mathbb{R}^{d}\times\mathbb{R}^{d}\rightarrow\mathbb{R}^{+}, and any x1,x2x_{1},x_{2}\in\mathcal{F}, we have:

|f(x1,Θ)f(x2,Θ)|L×dist(x1,x2)|f(x_{1},\mathcal{M}_{\Theta})-f(x_{2},\mathcal{M}_{\Theta})|\leq L\times dist(x_{1},x_{2}) (7)

At iteration mm, we denote the CE found by the outer minimisation as x(m)x^{\prime(m)}, and the shifted model found by the inner maximisation as Θ(m)\mathcal{M}_{\Theta}^{(m)}. Then, f(x(m),Θ(m)):=Θ(m)(x(m))+σf(x^{\prime(m)},\mathcal{M}_{\Theta}^{(m)}):=-\mathcal{M}_{\Theta}^{(m)}(x^{\prime(m)})+\sigma. Assume at step mm the algorithm has not terminated, then

f(x(m),Θ(m))>tf(x^{\prime(m)},\mathcal{M}_{\Theta}^{(m)})>t (8)

For the iteration steps n>mn>m, x(n)x^{\prime(n)} is required to be valid on Θ(m)\mathcal{M}_{\Theta}^{(m)} as specified in the outer minimisation problem, we therefore have:

f(x(n),Θ(m))0f(x^{\prime(n)},\mathcal{M}_{\Theta}^{(m)})\leq 0 (9)

Combining (8) and (9) yields

f(x(m),Θ(m))f(x(n),Θ(m))>t\displaystyle f(x^{\prime(m)},\mathcal{M}_{\Theta}^{(m)})-f(x^{\prime(n)},\mathcal{M}_{\Theta}^{(m)})>t (10)

Further combining (10) with (7), for the iteration steps n>mn>m,

dist(x(m),x(n))>tLdist(x^{\prime(m)},x^{\prime(n)})>\frac{t}{L} (11)

Consider the balls Bi,i=1,,m,B_{i},i=1,\ldots,m, of diameter tL\frac{t}{L} centred at each intermediate result of the outer minimisation problem, x(i)x^{\prime(i)}. From (11), it can be concluded that for any two intermediate x(i)x^{\prime(i)}, x(j)x^{\prime(j)}, 1<i,j<m1<i,j<m, dist(x(i),x(j))>tLdist(x^{\prime(i)},x^{\prime(j)})>\frac{t}{L}, and x(i){x^{\prime(i)}} and x(j){x^{\prime(j)}} are the centres of the balls B(i)B^{(i)} and B(j)B^{(j)}. Therefore, any two circles will not intercept. The total volume of these balls is thus m×U×(tL)dm\times U\times(\frac{t}{L})^{d}, where UU is the unit volume in d\mathbb{R}^{d}.

Consider a ball that encompasses the feasible solution region \mathcal{F} and let RR be its radius. We know that xi,i=1,,m,x^{\prime}_{i},i=1,\ldots,m, are all within the feasible region \mathcal{F}, therefore, the ball BB that has a radius R+t2LR+\frac{t}{2L} will cover the spaces of the small balls Bi,i=1,,mB_{i},i=1,\ldots,m. Also, the volume of BB is U×(2R+tL)dU\times(2R+\frac{t}{L})^{d} and will be greater than the total volume of the small balls, which means:

U×(2R+tL)d>m×U×(tL)dm<(2RLt+1)d\displaystyle U\times\left(2R+\frac{t}{L}\right)^{d}>m\times U\times\left(\frac{t}{L}\right)^{d}\quad\implies\quad m<\left(\frac{2RL}{t}+1\right)^{d}

It can be concluded that the step number at which Algorithm 1 has not terminated is bounded above by the (2RLt+1)d(\frac{2RL}{t}+1)^{d}. ∎

5 Experiments

In this section, we demonstrate that our proposed method achieves state-of-the-art performances compared with existing robust CEs generation methods.

Datasets and Classifiers.

Our experiments use four benchmark datasets in financial and legal contexts: the Adult Income (ADULT), COMPAS, Give Me Some Credits (GMC), and HELOC datasets. We adopt the pre-processed versions available in the CARLA library Pawelczyk et al. (2021) where each dataset contains binarised categorical features and min-max scaled continuous features. Labels 0 and 1 are the unwanted and the desired class, respectively. We split each dataset into two halves. We use the first half for training NNs with which the robust CEs are generated, and the second half for model retraining and evaluating the robustness of the CEs.

For making predictions and generating CEs, the NNs contain two hidden layers with ReLU activation functions. They are trained using the Adam optimiser with a batch size of 32, and under the standard 80%,20%80\%,20\% train-test dataset split setting. The classifiers achieved 84%, 85%, 94%, and 76% accuracies on the test set of ADULT, COMPAS, GMC, and HELOC datasets, respectively.

The retrained models have the same hyperparameters and training procedures as the original classifiers. Following the experimental setup in previous works  Dutta et al. (2022); Ferrario and Loi (2022); Nguyen et al. (2022); Black et al. (2022); Upadhyay et al. (2021), for each dataset, we train 10 new models using both halves of the dataset to simulate the possible retrained models after new data are collected. We also train 10 new models using 99% of the first half of the dataset (different 1% data are discarded for each training), to simulate the leave-one-out retraining procedures. The random seed is perturbed for retraining. These 20 retrained models are used for evaluating the robustness of CEs.

Evaluation Metrics.

The CEs are evaluated by the following metrics for their proximity, plausibility, and robustness.

  • 1\ell_{1} measures the average L1L_{1} distance between a CE and its corresponding input.

  • loflof is the average 10-Local Outlier Factor Breunig et al. (2000) of the generated CEs, which indicates to what extent a data point is an outlier wrt its kk nearest neighbours in a specified dataset. loflof values close to 11 indicate inliers, larger values (especially if greater than 1.51.5) indicate outliers.

  • vrvr, the validity of CEs on the retrained models, is defined as the average percentage of CEs that remain valid (classified to class 1) under the retrained models.

  • vΔv\Delta is the percentage of CEs that are Δ\Delta-robust. The bound of model parameter changes δ\delta is specified to be the same as the value used in our algorithm.

Baselines.

We compare our method with six state-of-the-art methods for generating CEs, including five which target robustness. WCE Wachter et al. (2017) is the first method to generate CEs for NNs, which minimises the 1\ell_{1} distance between the CEs and the inputs. Robust Bayesian Recourse (RBR) Nguyen et al. (2022) addresses the proximity, robustness, and plausibility of CEs. RobXNN Dutta et al. (2022) is a nearest-neighbour-based method that focuses on a different notion of robustness to model changes. Robust Algorithmic Recourse (ROAR) Upadhyay et al. (2021) optimises for proximity and the same Δ\Delta notion of robustness. Proto-R and MILP-R are the methods proposed by Jiang et al. (2023) which embed the Δ\Delta-robustness test into the base methods of Van Looveren and Klaise (2021) and Mohammadi et al. (2021). For all methods including ours, we tune their hyperparameters to maximise the validity after retraining vrvr.

ADULT COMPAS GMC HELOC
​​vr\uparrow​​ ​​vΔ\Delta\uparrow​​ ​​1\ell_{1}\downarrow​​ ​​lof\downarrow​​ ​​vr\uparrow​​ ​​vΔ\Delta\uparrow​​ ​​​1\ell_{1}\downarrow​​​ ​​lof\downarrow​​ ​​vr\uparrow​​ ​​vΔ\Delta\uparrow​​ ​​​1\ell_{1}\downarrow​​​ ​​lof\downarrow​​ ​​vr\uparrow​​ ​​vΔ\Delta\uparrow​​ ​​​1\ell_{1}\downarrow​​​ ​​lof\downarrow​​
​​​​WCE​​​​ ​​89​​ ​​78​​ ​​.175​​ ​​1.59​​ ​​57​​ ​​0​​ ​​.170​​ ​​1.81​​ ​​84​​ ​​18​​ ​​.148​​ ​​2.80​​ ​​49​​ ​​0​​ ​​.045​​ ​​1.16​​
​​​​RBR​​​​ ​​100​​ ​​0​​ ​​.031​​ ​​1.28​​ ​​100​​ ​​62​​ ​​.043​​ ​​1.34​​ ​​90​​ ​​0​​ ​​.050​​ ​​1.31​​ ​​80​​ ​​0​​ ​​.038​​ ​​1.10​​
​​​​RobXNN​​​​ ​​100​​ ​​82​​ ​​.064​​ ​​1.28​​ ​​100​​ ​​82​​ ​​.050​​ ​​1.11​​ ​​100​​ ​​96​​ ​​.073​​ ​​1.35​​ ​​100​​ ​​30​​ ​​.073​​ ​​1.04​​
​​​​ROAR​​​​ ​​99​​ ​​98​​ ​​.279​​ ​​1.96​​ ​​98​​ ​​100​​ ​​.219​​ ​​2.84​​ ​​96​​ ​​88​​ ​​.188​​ ​​4.22​​ ​​98​​ ​​98​​ ​​.109​​ ​​1.57​​
​​​​PROTO-R​​​​ ​​98​​ ​​55​​ ​​.068​​ ​​1.60​​ ​​100​​ ​​100​​ ​​.084​​ ​​1.36​​ ​​100​​ ​​100​​ ​​.066​​ ​​1.49​​ ​​100​​ ​​100​​ ​​.057​​ ​​1.21​​
​​​​MILP-R​​​​ ​​100​​ ​​100​​ ​​.024​​ ​​1.69​​ ​​100​​ ​​100​​ ​​.040​​ ​​1.71​​ ​​100​​ ​​100​​ ​​.059​​ ​​2.08​​ ​​100​​ ​​100​​ ​​.044​​ ​​2.48​​
​​​​OURS​​​​ ​​100​​ ​​100​​ ​​.046​​ ​​1.22​​ ​​100​​ ​​100​​ ​​.039​​ ​​1.24​​ ​​100​​ ​​100​​ ​​.058​​ ​​1.24​​ ​​100​​ ​​100​​ ​​.057​​ ​​1.04​​
Table 1: Evaluations of PROPLACE (OURS) and baselines on NNs. The \uparrow (\downarrow) indicates that higher (lower) values are preferred for the evaluation metric.
Results.

We randomly select 50 test points from each dataset that are classified to be the unwanted class, then apply our method and each baseline to generate CEs for these test points. Results are shown in Table 1.

As a non-robust baseline, the WCE method is the least robust while producing high 1\ell_{1} costs and poor plausibility. Though RBR shows the lowest 1\ell_{1} results on three datasets, it has only moderate robustness against the naturally retrained models and is not Δ\Delta-robust on any dataset. The rest of the baselines all show strong robustness on at least three datasets, with our method having slightly better vrvr and vΔv\Delta results, evaluated at 100% in every experiment. This indicates that our method PROPLACE can not only guarantee robustness under bounded model parameter changes but also induce reliable robustness against unbounded model changes. In terms of plausibility, our method shows the best lof score in most experiments. Therefore, our method has addressed the limitation in the literature that no method optimises for guaranteed Δ\Delta-robustness and plausibility. Though the two properties have established trade-offs with proximity Pawelczyk et al. (2020b, 2022); Upadhyay et al. (2021), our method still shows 1\ell_{1} costs lower than all methods except RBR, which is significantly less robust, and MILP-R, which finds outliers. For the COMPAS dataset, our method has the best proximity result among all baselines.

Note that the PROTO-R baseline from the work which proposed certification for Δ\Delta-robustness failed to find Δ\Delta-robust CEs on the ADULT dataset, as was the case in their results (see Table 1, Jiang et al. (2023)). This is due to the fact that their method rely heavily on a base method to find CEs, and it is not straightforward to be always able to direct the hyperparameters search for optimising Δ\Delta-robustness. With improved soundness and completeness (Proposition 1), PROPLACE always finds provably robust results.

6 Conclusions

We proposed a robust optimisation framework PROPLACE to generate provably robust and plausible CEs for neural networks. The method addresses the limitation in the literature that existing methods lack formal robustness guarantees to bounded model parameter changes and do not generate plausible CEs. We proved the soundness, completeness, and convergence of PROPLACE. Through a comparative study, we show the efficacy of our method, demonstrating the best robustness and plausibility results with better proximity than the most robust baselines. Despite the specific form of robustness we target, PROPLACE is also empirically robust to model retraining with unbounded parameter changes. Future work could include investigating the properties of actionability and diversity, evaluations with user studies, and investigating connections between Δ\Delta-robustness and different notions of robustness measures.

Acknowledgement

Jiang, Rago and Toni were partially funded by J.P. Morgan and by the Royal Academy of Engineering under the Research Chairs and Senior Research Fellowships scheme. Jianglin Lan is supported by a Leverhulme Trust Early Career Fellowship under Award ECF-2021-517. Leofante is supported by an Imperial College Research Fellowship grant. Rago and Toni were partially funded by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No. 101020934). Any views or opinions expressed herein are solely those of the authors listed.

References

  • Alvarez-Melis and Jaakkola [2018] David Alvarez-Melis and Tommi S. Jaakkola. Towards robust interpretability with self-explaining neural networks. In Adv. in Neural Information Processing Systems 31, NeurIPS, 2018.
  • Bajaj et al. [2021] Mohit Bajaj, Lingyang Chu, Zi Yu Xue, Jian Pei, Lanjun Wang, Peter Cho-Ho Lam, and Yong Zhang. Robust counterfactual explanations on graph neural networks. In Adv. in Neural Information Processing Systems 34, NeurIPS, 2021.
  • Ben-Tal et al. [2009] Aharon Ben-Tal, Laurent El Ghaoui, and Arkadi Nemirovski. Robust optimization, volume 28. Princeton university press, 2009.
  • Black et al. [2022] Emily Black, Zifan Wang, and Matt Fredrikson. Consistent counterfactuals for deep models. In 10th Int. Conf. on Learning Representations, ICLR, 2022.
  • Breunig et al. [2000] Markus M. Breunig, Hans-Peter Kriegel, Raymond T. Ng, and Jörg Sander. LOF: Identifying density-based local outliers. In ACM SIGMOD Int. Conf. on Management of Data, 2000.
  • Bui et al. [2022] Ngoc Bui, Duy Nguyen, and Viet Anh Nguyen. Counterfactual plans under distributional ambiguity. In 10th Int. Conf. on Learning Representations, ICLR, 2022.
  • Celar and Byrne [2023] Lenart Celar and Ruth M. J. Byrne. How people reason with counterfactual and causal explanations for AI decisions in familiar and unfamiliar domains. Memory & Cognition, 2023.
  • Dhurandhar et al. [2018] Amit Dhurandhar, Pin-Yu Chen, Ronny Luss, Chun-Chen Tu, Paishun Ting, Karthikeyan Shanmugam, and Payel Das. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. In Adv. in Neural Information Processing Systems 31, NeurIPS, 2018.
  • Dominguez-Olmedo et al. [2022] Ricardo Dominguez-Olmedo, Amir-Hossein Karimi, and Bernhard Schölkopf. On the adversarial robustness of causal algorithmic recourse. In 39th Int. Conf. on Machine Learning, ICML, 2022.
  • Dutta et al. [2022] Sanghamitra Dutta, Jason Long, Saumitra Mishra, Cecilia Tilli, and Daniele Magazzeni. Robust counterfactual explanations for tree-based ensembles. In 39th Int. Conf. on Machine Learning, ICML, 2022.
  • Ferrario and Loi [2022] Andrea Ferrario and Michele Loi. The robustness of counterfactual explanations over time. IEEE Access, 10:82736–82750, 2022.
  • Guidotti [2022] Riccardo Guidotti. Counterfactual explanations and how to find them: literature review and benchmarking. Data Mining and Knowledge Discovery, pages 1–55, 2022.
  • Gurobi Optimization, LLC [2023] Gurobi Optimization, LLC. Gurobi Optimizer Reference Manual, 2023.
  • Hamman et al. [2023] Faisal Hamman, Erfaun Noorani, Saumitra Mishra, Daniele Magazzeni, and Sanghamitra Dutta. Robust counterfactual explanations for neural networks with probabilistic guarantees. In 40th Int. Conf. on Machine Learning, ICML, 2023.
  • Huai et al. [2022] Mengdi Huai, Jinduo Liu, Chenglin Miao, Liuyi Yao, and Aidong Zhang. Towards automating model explanations with certified robustness guarantees. In 36th AAAI Conf. on Artificial Intelligence, AAAI, 2022.
  • Jiang et al. [2023] Junqi Jiang, Francesco Leofante, Antonio Rago, and Francesca Toni. Formalising the robustness of counterfactual explanations for neural networks. In 37th AAAI Conf. on Artificial Intelligence, AAAI, 2023.
  • Leofante and Lomuscio [2023a] Francesco Leofante and Alessio Lomuscio. Robust explanations for human-neural multi-agent systems with formal verification. In 20th Eur. Conf. on Multi-Agent Systems EUMAS, 2023.
  • Leofante and Lomuscio [2023b] Francesco Leofante and Alessio Lomuscio. Towards robust contrastive explanations for human-neural multi-agent systems. In 22nd Int. Conf. on Autonomous Agents and Multiagent Systems, AAMAS, 2023.
  • Leofante et al. [2023] F Leofante, E Botoeva, and V Rajani. Counterfactual explanations and model multiplicity: a relational verification view. In The 20th Int. Conf. on Principles of Knowledge Representation and Reasoning, KR, 2023.
  • Maragno et al. [2023] Donato Maragno, Jannis Kurtz, Tabea E Röber, Rob Goedhart, Ş Ilker Birbil, and Dick den Hertog. Finding regions of counterfactual explanations via robust optimization. arXiv:2301.11113, 2023.
  • Marques-Silva and Ignatiev [2022] João Marques-Silva and Alexey Ignatiev. Delivering trustworthy AI through formal XAI. In 36th AAAI Conf. on Artificial Intelligence, AAAI, 2022.
  • Mohammadi et al. [2021] Kiarash Mohammadi, Amir-Hossein Karimi, Gilles Barthe, and Isabel Valera. Scaling guarantees for nearest counterfactual explanations. In Conf. on AI, Ethics, and Society, AIES, 2021.
  • Mothilal et al. [2020] Ramaravind Kommiya Mothilal, Amit Sharma, and Chenhao Tan. Explaining machine learning classifiers through diverse counterfactual expl. In Conf. on Fairness, Accountability, and Transparency, 2020.
  • Mutapcic and Boyd [2009] Almir Mutapcic and Stephen Boyd. Cutting-set methods for robust convex optimization with pessimizing oracles. Optimization Methods & Software, 24(3):381–406, 2009.
  • Nguyen et al. [2022] Tuan-Duy H Nguyen, Ngoc Bui, Duy Nguyen, Man-Chung Yue, and Viet Anh Nguyen. Robust bayesian recourse. In 38th Conf. on Uncertainty in AI, UAI, 2022.
  • Pawelczyk et al. [2020a] Martin Pawelczyk, Klaus Broelemann, and Gjergji Kasneci. Learning model-agnostic counterfactual explanations for tabular data. In The ACM Web Conference, WWW, 2020.
  • Pawelczyk et al. [2020b] Martin Pawelczyk, Klaus Broelemann, and Gjergji Kasneci. On counterfactual explanations under predictive multiplicity. In 36th Conf. on Uncertainty in AI, UAI, 2020.
  • Pawelczyk et al. [2021] Martin Pawelczyk, Sascha Bielawski, Johannes van den Heuvel, Tobias Richter, and Gjergji Kasneci. CARLA: A python library to benchmark algorithmic recourse and counterfactual explanation algorithms. Adv. in Neural Information Processing Systems 34, NeurIPS (Datasets and Benchmarks), 2021.
  • Pawelczyk et al. [2022] Martin Pawelczyk, Teresa Datta, Johannes van-den Heuvel, Gjergji Kasneci, and Himabindu Lakkaraju. Probabilistically robust recourse: Navigating the trade-offs between costs and robustness in algorithmic recourse. arXiv:2203.06768, 2022.
  • Poyiadzi et al. [2020] Rafael Poyiadzi, Kacper Sokol, Raúl Santos-Rodríguez, Tijl De Bie, and Peter A. Flach. FACE: feasible and actionable counterfactual explanations. In Conf. on AI, Ethics, and Society, AIES, 2020.
  • Prabhakar and Afzal [2019] Pavithra Prabhakar and Zahra Rahimi Afzal. Abstraction based output range analysis for neural networks. In Adv. in Neural Information Processing Systems 32, NeurIPS, 2019.
  • Rawal et al. [2020] Kaivalya Rawal, Ece Kamar, and Himabindu Lakkaraju. Algorithmic recourse in the wild: Understanding the impact of data and model shifts. arXiv:2012.11788, 2020.
  • Sharma et al. [2020] Shubham Sharma, Jette Henderson, and Joydeep Ghosh. CERTIFAI: A common framework to provide explanations and analyse the fairness and robustness of black-box models. In Conf. on AI, Ethics, and Society, AIES, 2020.
  • Tolomei et al. [2017] Gabriele Tolomei, Fabrizio Silvestri, Andrew Haines, and Mounia Lalmas. Interpretable predictions of tree-based ensembles via actionable feature tweaking. In 23rd ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, 2017.
  • Upadhyay et al. [2021] Sohini Upadhyay, Shalmali Joshi, and Himabindu Lakkaraju. Towards robust and reliable algorithmic recourse. Adv. in Neural Information Processing Systems 34, NeurIPS, 2021.
  • Ustun et al. [2019] Berk Ustun, Alexander Spangher, and Yang Liu. Actionable recourse in linear classification. In Conf. on Fairness, Accountability, and Transparency, 2019.
  • Van Looveren and Klaise [2021] Arnaud Van Looveren and Janis Klaise. Interpretable counterfactual explanations guided by prototypes. In Eur. Conf. on Machine Learning and Knowledge Discovery in Databases, ECML PKDD, 2021.
  • Verma et al. [2022] Sahil Verma, Keegan Hines, and John P Dickerson. Amortized generation of sequential algorithmic recourses for black-box models. In 36th AAAI Conf. on Artificial Intelligence, AAAI, 2022.
  • Virgolin and Fracaros [2023] Marco Virgolin and Saverio Fracaros. On the robustness of sparse counterfactual expl. to adverse perturbations. Artificial Intelligence, 316:103840, 2023.
  • Vo et al. [2023] Vy Vo, Trung Le, Van Nguyen, He Zhao, Edwin V. Bonilla, Gholamreza Haffari, and Dinh Phung. Feature-based learning for diverse and privacy-preserving counterfactual explanations. In ACM SIGKDD Conf. on Knowledge Discovery and Data Mining, 2023.
  • Wachter et al. [2017] Sandra Wachter, Brent D. Mittelstadt, and Chris Russell. Counterfactual explanations without opening the black box: Automated decisions and the GDPR. Harv. JL & Tech., 31:841, 2017.
  • Zhang et al. [2023] Songming Zhang, Xiaofeng Chen, Shiping Wen, and Zhongshan Li. Density-based reliable and robust explainer for counterfactual explanation. Expert Systems with Applications, pages 120–214, 2023.