1 Introduction
Consider the following setting: a 28-year-old female professional working as a software engineer seeks a mortgage to purchase a home. Consider further that the loan-granting institution (e.g., bank) uses a binary classifier and denies the individual the loan based on her attributes. Naturally, in this context, answering the following questions become relevant to the individual:
Q1:
Why was I rejected the loan?
Q2:
What can I do to get the loan in the future?
In the setting of the preceding example, unless the policy of the bank is relaxed, the individual must expend effort to change their situation to be favorably treated by the decision-making system. Examples such as this are prevalent not only in finance [
20,
168] but also in justice (e.g., pretrial bail) [
10], healthcare (e.g., prescription of life-altering medication) [
24,
28,
91], and other settings (e.g., hiring) [
50,
149,
169,
207] broadly classified as
consequential decision-making settings [
20,
35,
52,
113]. Given the rapid adoption of automated decision-making systems in these settings, designing models that not only have high objective accuracy but also afford the individual with
explanations and
recommendations to favorably change their situation is of paramount importance, and even argued to be a legal necessity (GDPR [
231]). This is the concern of
algorithmic recourse.
Contributions. Our review brings together the plethora of recent works on algorithmic recourse. In reviewing the literature, we identify opportunities to consolidate definitions, construct technical baselines for future comparison, and situate recourse in the broader ethical
machine learning (ML) literature. The primary contribution is thus a unified overview of the definitions (Section
2), formulations (Section
3), and technical solutions (Section
4) for computing
(nearest) contrastive explanations and
(minimal) consequential recommendations, across a broad range of setups. A high-level summary is curated and presented in Table
1. Clearly distinguishing between the preceding two recourse offerings and presenting their often overlooked different technical treatment is a common theme throughout this work. Additionally, we identify a number of prospective research directions which challenge the assumptions of existing setups, and present extensions to better situate recourse in the broader ethical ML literature (Section
5).
Who is this article for? This article aims to engage different audiences: practitioners (
P), researchers (
R), and legal scholars (
L). Section
2 builds on the rich literature in the philosophy of science and presents a clear distinction between contrastive explanations and consequential recommendations based on the causal history of events (
R,
L). Table
1 presents an overview of 50+ technical papers that offer recourse in a broad range of setups (
P,
R). Section
3 formulates the recourse problem as constrained optimization and presents the variety of constraints needed to model real-world settings (
R). Section
4 presents the difficulty in solving for recourse in general settings and summarizes existing approximate solutions that trade off various desirable properties (
P,
R,
L). Finally, Section
5 argues that recourse should be considered in relation to other ethical ML considerations (
L) and outlines future research directions (
R).
Scope. To adequately present the material, we provide brief introductions to various related topics, such as explainable ML (or XAI), causal and counterfactual inference, and optimization, among others. Our work is not meant to be a comprehensive review of these topics and merely aims to situate algorithmic recourse in the context of these related fields. For the interested reader, an in-depth discussion on the social, philosophical, cognitive, and psychological aspects of explanations can be found elsewhere [
21,
36,
157,
226]. Other works of interest are available on XAI [
1,
2,
32,
68,
79,
86,
94,
139,
150,
163], causal and counterfactual inference, [
82,
95,
166,
185,
186,
206,
245], and optimization [
33,
174,
214].
3 Formulation
Given a fixed predictive model, commonly assumed to be a binary classifier,
\(h: \mathcal {X} \rightarrow \lbrace 0, 1\rbrace\), with
\(\mathcal {X} = \mathcal {X}_1 \times \cdots \times \mathcal {X}_D\), we can define the set of
contrastive explanations for a (factual) input
\(\boldsymbol {x}^\texttt {F}\in \mathcal {X}\) as
\(\mathcal {E} := \lbrace \boldsymbol {x}^\texttt {CF}\in \mathcal {P}(\mathcal {X}) ~|~ h(\boldsymbol {x}^\texttt {CF}) \not= h(\boldsymbol {x}^\texttt {F}) \rbrace\). Here,
\(\mathcal {P}(\mathcal {X}) \subseteq \mathcal {X}\) is a
plausible subspace of
\(\mathcal {X}\), according to the distribution of training data (see Section
3.3.1). Descriptively, contrastive explanations identify alternative feature combinations (in nearby worlds [
136]) that result in a favorable prediction from the fixed model. Assuming a notion of dissimilarity between instances, represented as
\(\texttt {dist}(\cdot , \cdot) : \mathcal {X} \times \mathcal {X} \rightarrow \mathbb {R}_+\), one can identify
nearest contrastive explanations (a.k.a. counterfactual explanations) as follows:
where
\(\boldsymbol {\delta }\) is the perturbation applied independently to the feature vector
\(\boldsymbol {x}^\texttt {F}\) to obtain the counterfactual instance
\(\boldsymbol {x}^\texttt {CFE}\) [
234]. As discussed in Section
2, although contrastive explanations identify the
feature vectors that would achieve recourse, in general, the
set of actions that would need to be performed to realize these features are not directly implied from the explanation [
114]. Thus, a
consequential recommendation for (factual) input
\(\boldsymbol {x}^\texttt {F}\in \mathcal {X}\) is defined as
\(\mathcal {R} := \lbrace a \in \mathcal {A}(\boldsymbol {x}^\texttt {F}) ~|~ h(\boldsymbol {x}^\texttt {SCF}(\boldsymbol {a}; \boldsymbol {x}^\texttt {F})) \not= h(\boldsymbol {x}^\texttt {F}) \rbrace\). Here,
\(\mathcal {A}(\boldsymbol {x}^\texttt {F})\) is the set of
feasible actions that can be performed by the individual seeking recourse (see Section
3.3.2). Approaching the recourse problem from a causal perspective within the SCM framework [
114], actions are considered as interventions of the form
\(\boldsymbol {a}= \mathrm{do}(\lbrace \mathrm{X}_i := x^\texttt {F}_i + \theta _i \rbrace _{i\in \mathcal {I}}) \in \mathcal {A}(\boldsymbol {x}^\texttt {F})\), and
\(\boldsymbol {x}^\texttt {SCF}(\boldsymbol {a}; \boldsymbol {x}^\texttt {F})\) denotes the structural counterfactual of
\(\boldsymbol {x}^\texttt {F}\) had action
\(\boldsymbol {a}\) been performed, all else being equal [
184]. Finally, given a notion of cost of actions, capturing the effort expended by the individual as
\(\texttt {cost}(\cdot ; \cdot) : \mathcal {A} \times \mathcal {X} \rightarrow \mathbb {R}_+\), one can identify
minimal consequential recommendations as follows:
Importantly, the solution of (
1) yields a nearest contrastive explanation (i.e.,
\(\boldsymbol {x}^\texttt {CF}= \boldsymbol {x}^\texttt {F}+ \boldsymbol {\delta }^*\)), with no direct mapping to a set of (minimal) consequential recommendations [
114]. Conversely, solving (
2) yields both a minimal consequential recommendation (i.e.,
\(\boldsymbol {a}^*\)) and a contrastive explanation (i.e., by construction
\(\boldsymbol {x}^\texttt {CF}= \boldsymbol {x}^\texttt {SCF}(\boldsymbol {a}^*; \boldsymbol {x}^\texttt {F})\)).
3Our position is that, with the aim of providing recourse, the primary goal should be to provide minimal consequential recommendations that result in a (not necessarily nearest) contrastive explanation when acted upon. Offering nearest contrastive explanations that are not necessarily attainable through
minimal effort is of secondary importance to the individual. In practice, however, due to the additional assumptions needed to solve (
2) (specifically for computing
\(\boldsymbol {x}^\texttt {SCF}\)), the literature often resorts to solving (
1). Despite the prevalance of methods assuming independent features, we quote the negative result of Karimi et al. [
115, prop. 2], whereby the authors show that without full specification of the causal relations among variables, recourse cannot be guaranteed. As pointed out by Chou et al. [
47], explanations derived from popular algorithms based on non-causal spurious correlations “may yield sub-optimal, erroneous, or even biased explanations.” Therefore, for the rest of this section, we operate on the assumption of knowledge of an SCM, which allows for a deterministic formulation of recourse yielding both minimal consequential recommendation and resulting contrastive explanations. Later in Sections
5 and
6, we relax these assumptions and discuss the relatively easier settings in which causal assumptions are scarce or absent entire. This is not to say that counterfactual explanations not accompanied by recommendation actions are not useful. On the contrary, as noted by Karimi et al. [
114], counterfactual explanations hold promise for “guided audit of the data” [
234] and evaluating various desirable model properties, such as robustness [
98,
209] or fairness [
96,
113,
209,
223] (see also Section
5). Besides this, it has been shown that designers of interpretable ML systems use counterfactual explanations for predicting model behavior [
125], uncovering inaccuracies in the data profile of individuals [
226], and also for medical outcome prediction under hypothesized treatments [
153] or realistic medical image generation [
238]. Beyond the consequential domains mentioned earlier and in Section
1, counterfactual explanations have been used for changing behavior of source code [
48], overcoming class imbalance issues by amending underrepresented classes with plausible counterfactuals [
217], and aiding the interpretable design of molecules [
239].
In the rest of this section, we provide an overview of the objective function and constraints used in (
1) and (
2), followed by a description of the datatypes commonly used in recourse settings. Finally, we conclude with related formulations. Then in Section
4, we review the tools used to solve the formulations defined here.
3.1 Optimization Objective
Generally, it is difficult to define dissimilarity (
dist) between individuals or
cost functions for effort expended by individuals. Notably, this challenge was first discussed in the algorithmic fairness literature [
73,
107] and later echoed throughout the algorithmic recourse community [
21,
226]. In fact, “the law provides no formal guidance as to the proper metric for determining what reasons are most salient” [
208]. Despite this, existing works have presented various ad hoc formulations with sensible intuitive justifications or practical allowance, which we review in the following.
3.1.1 On dist.
Wachter et al. [
234] define
dist as the Manhattan distance, weighted by the inverse
median absolute deviation (MAD):This distance has several desirable properties, including accounting and correcting for the different ranges across features through the MAD heuristic, robustness to outliers with the use of the median absolute difference, and, finally, favoring sparse solutions through the use of \(\ell _1\) Manhattan distance.
Karimi et al. [
113] propose a weighted combination of
\(\ell _p\) norms as a flexible measure across a variety of situations. The weights,
\(\alpha , \beta , \gamma , \zeta\) as shown in the following, allow practitioners to balance between sparsity of changes (i.e., through the
\(\ell _0\) norm), an elastic measure of distance (i.e., through the
\(\ell _1, \ell _2\) norms) [
63], and a maximum normalized change across all features (i.e., through the
\(\ell _\infty\) norm):
where
\(\boldsymbol {\delta }= [\delta _1, \ldots , \delta _D]^T\) and
\(\delta _k: \mathcal {X}_k \times \mathcal {X}_k \rightarrow [0,1] ~ \forall ~ k \in [D]\). This measure accounts for the variability across heterogeneous features (see Section
3.5) by independently normalizing the change in each dimension according to its spread. Additional weights may also be used to relative emphasize changes in specific variables. Finally, other works aim to minimize dissimilarity on a graph manifold [
189], in terms of Euclidean distance in a learned feature space [
109,
183], or using a Riemannian metric in a latent space [
14,
15].
3.1.2 On cost.
Similar to Karimi et al. [
113], various works explore
\(\ell _p\) norms to measure cost of actions. Ramakrishnan et al. [
190] explore
\(\ell _1, \ell _2\) norm as well as constant cost if specific actions are undertaken; Karimi et al. [
114],
115] minimize the
\(\ell _2\) norm between
\(\boldsymbol {x}^\texttt {F}\) and the action
\(\boldsymbol {a}\) assignment (i.e.,
\(||\boldsymbol {\theta }||_2\)); and Cui et al. [
55] explore combinations of
\(\ell _0, \ell _1, \ell _2\) norms over a user-specified cost matrix. Encoding individual-dependent restrictions is critical—for example, obtaining
credit is more difficult for an foreign student than for a local resident.
Beyond
\(\ell _p\) norms, the work of Ustun et al. [
223] propose the total- and maximum-log percentile shift measures, to automatically account for the distribution of points in the dataset—for example,
where
\(Q_k(\cdot)\) is the CDF of
\(x_k\) in the target population. This type of metric naturally accounts for the relative difficulty of moving to unlikely (high or low percentile) regions of the data distribution. For instance, going from a 50th to a 55th percentile in
school grades is simpler than going from the 90th to the 95th percentile.
3.1.3 On the Relation Between dist and cost.
In a world in which changing one variable does not affect others, one can see a parallel between the counterfactual instance of (
1) (i.e.,
\(\boldsymbol {x}^\texttt {CF}= \boldsymbol {x}^\texttt {F}+ \boldsymbol {\delta }\)) and that of (
2) (i.e.,
\(\boldsymbol {x}^\texttt {CF}= \boldsymbol {x}^\texttt {SCF}(\boldsymbol {a}; \boldsymbol {x}^\texttt {F}) = \boldsymbol {x}^\texttt {F}+ \boldsymbol {\theta }\)). This mirroring form suggests that definitions of dissimilarity between individuals (i.e.,
dist) and effort expended by an individual (i.e.,
cost) may be used interchangeably. Following Karimi et al. [
114], however, we do not consider a general 1-1 mapping between
dist and
cost. For instance, in a two-variable system with
medication as the parent of
headache, an individual who consumes more than the recommended amount of medication may not recover from the headache (i.e., higher cost but smaller
symptomatic distance relative to another individual who consumed the correct amount). Furthermore, although dissimilarity is often considered to be symmetric (i.e.,
\(\texttt {dist}(\boldsymbol {x}^\texttt {F}_A, \boldsymbol {x}^\texttt {F}_B) = \texttt {dist}(\boldsymbol {x}^\texttt {F}_B, \boldsymbol {x}^\texttt {F}_A)\)), the effort needed to go from one profile to another need not satisfy symmetry—for example, spending money is easier than saving money (i.e.,
\(\texttt {cost}(\boldsymbol {a}= \mathrm{do}(\mathrm{X}_{\$} := \boldsymbol {x}^\texttt {F}_{A,\$} - \$500); \boldsymbol {x}^\texttt {F}_A) \le \texttt {cost}(\boldsymbol {a}= \mathrm{do}(\mathrm{X}_{\$} := \boldsymbol {x}^\texttt {F}_{B,\$} + \$500); \boldsymbol {x}^\texttt {F}_B)\). These examples illustrate that the interdisciplinary community must continue to engage to define the distinct notions of
dist and
cost, and such definitions cannot arise from a technical perspective alone.
3.2 Model and Counterfactual Constraints
3.2.1 Model.
A variety of fixed models have been explored in the literature for which recourse is to be generated. As summarized in Table
1, we broadly divide them in four categories: tree-based (TB), kernel-based (KB), differentiable (DF), and other (OT) types (e.g., generalized linear models, naive Bayes,
k-nearest neighbors). Although the literature on recourse has primarily focused on binary classification settings, most formulations can easily be extended to multi-class classification or regression settings. Extensions to such settings are straightforward, where the model constraint is replaced with
\(h(\boldsymbol {x}^\texttt {CF}) = k\) for a target class or
\(h(\boldsymbol {x}^\texttt {CF}) \in [a,b]\) for a desired regression interval, respectively. Alternatively, soft predictions may be used in place of hard predictions, where the goal may be, for example, to increase the prediction gap between the highest-predicted and second-highest-predicted class—that is,
\(\text{Pred}(\boldsymbol {x}^\texttt {CF})_i - \text{Pred}(\boldsymbol {x}^\texttt {CF})_j,\) where
\(i = \mathop {argmax}_{k \in K} \text{Pred}(\boldsymbol {x}^\texttt {CF})_k\),
\(j = \mathop {argmax}_{k \in K \setminus i} \text{Pred}(\boldsymbol {x}^\texttt {CF})_k\). In the end, the model constraint representing the change in prediction may be arbitrarily non-linear, non-differentiable, and non-monotone [
21], which may limit the applicability of solutions (c.f. Section
4).
3.2.2 Counterfactual.
The counterfactual constraint depends on the type of recourse offered. Whereas
\(\boldsymbol {x}^\texttt {CF}= \boldsymbol {x}^\texttt {F}+ \boldsymbol {\delta }\) in (
1) is a linear constraint, computing
\(\boldsymbol {x}^\texttt {CF}= \boldsymbol {x}^\texttt {SCF}(\boldsymbol {a}; \boldsymbol {x}^\texttt {F})\) in (
2) involves performing the three step abduction-action-prediction process of Pearl et al. [
186] and may thus be non-parametric and arbitrary involved. A closed-form expression for deterministically computing the counterfactual in additive noise models is presented in the work of Karimi et al. [
114], and probabilistic derivations for more general SCMs are presented in another work by Karimi et al. [
115].
3.3 Actionability and Plausibility Constraints
3.3.1 Plausibility.
Existing literature has formalized plausibility constraint as one of three categories:
domain-consistency,
density-consistency, and
prototypical-consistency. Whereas domain-consistency restricts the counterfactual instance to the range of admissible values for the domain of features [
113], density-consistency focuses on likely states in the (empirical) distribution of features [
63,
64,
109,
111,
131,
183] identifying instances close to the data manifold. A third class of plausibility constraints selects counterfactual instances that are either directly present in the dataset [
189,
241] or close to a prototypical example of the dataset [
11,
12,
124,
131,
225].
3.3.2 Actionability (Feasibility).
The set of feasible actions,
\(\mathcal {A}(\boldsymbol {x}^\texttt {F})\), is the set of interventions
\(\mathrm{do}(\lbrace \mathrm{X}_i := x^\texttt {F}_i + \theta _i\rbrace _{i\in \mathcal {I}})\) that the individual,
\(\boldsymbol {x}^\texttt {F}\), is able to perform. To determine
\(\mathcal {A}(\boldsymbol {x}^\texttt {F})\), we must identify the set of variables upon which interventions are possible, as well as the pre-/post-conditions that the intervention must satisfy. The actionability of each variable falls into three categories [
114,
130]:
(1)
Actionable (and mutable) (e.g., bank balance);
(2)
Mutable but non-actionable (e.g., credit score);
(3)
Immutable (and non-actionable) (e.g., birthplace).
Intuitively, mutable but non-actionable variables are not directly actionable by the individual but may change as a consequence of a change to its causal ancestors (e.g., regular debt payment).
Having identified the set of actionable variables, an intervention can change the value of a variable
unconditionally (e.g.,
bank balance can increase or decrease), or
conditionally to a specific value [
114] or in a specific direction [
223]. Karimi et al. [
114] present the following examples to show that the actionable feasibility of an intervention on
\(\mathrm{X}_i\) may be contingent on any number of conditions:
(1)
Pre-intervention value of the intervened variable (i.e., \(x^\texttt {F}_i\))—for example, an individual’s age can only increase (i.e., \([x^\texttt {SCF}_{\rm\small AGE} \ge x^\texttt {F}_{\rm\small AGE}])\);
(2)
Pre-intervention value of other variables (i.e., \(\lbrace x^\texttt {F}_j \rbrace _{j \subset [d] \setminus i}\))—for example, an individual cannot apply for credit on a temporary visa (i.e., \([x^\texttt {F}_{\rm\small VISA} = \texttt {PERMANENT}] \ge [x^\texttt {SCF}_{\rm\small CREDIT} = \texttt {TRUE}])\);
(3)
Post-intervention value of the intervened variable (i.e., \(x^\texttt {SCF}_i\))—for example, an individual may undergo heart surgery (an additive intervention) only if they will not remiss due to sustained smoking habits (i.e., \([x^\texttt {SCF}_{\rm\small HEART} \not= \texttt {REMISSION}]);\)
(4)
Post-intervention value of other variables (i.e., \(\lbrace x^\texttt {SCF}_j \rbrace _{j \subset [d] \setminus i}\))—for example, an individual may undergo heart surgery only after their blood pressure (bp) is regularized due to medicinal intervention (i.e., \([x^\texttt {SCF}_{\rm\small BP} = \texttt {O.K.}] \ge [x^\texttt {SCF}_{\rm\small HEART} = \texttt {SURGERY}]).\)
All such feasibility conditions can easily be encoded as Boolean/logical constraint into
\(\mathcal {A}(\boldsymbol {x}^\texttt {F})\) and jointly solved for in the constrained optimization formulations (
1), (
2). An important side note to consider is that
\(\mathcal {A}(\boldsymbol {x}^\texttt {F})\) is
not restricted by the SCM assumptions but instead by individual-/context-dependent consideration that determine the
form,
feasibility, and
scope of actions [
114].
3.3.3 On the Relation Between Actionability and Plausibility.
While seemingly overlapping, actionability (i.e., \(\mathcal {A}(\boldsymbol {x}^\texttt {F})\)) and plausibility (i.e., \(\mathcal {P}(\mathcal {X})\)) are two distinct concepts: whereas the former restrict actions to those that are possible to do, the latter require that the resulting counterfactual instance be possibly true,believable,or realistic. Consider a Middle Eastern Ph.D. student who is denied a U.S. visa to attend a conference. Although it is quite likely for there to be favorably treated foreign students from other countries with similar characteristics (plausible gender, field of study, academic record, etc.), it is impossible for our student to act on their birthplace for recourse (i.e., a plausible explanation but an infeasible recommendation). Conversely, an individual may perform a set of feasible actions that would put them in an implausible state (e.g., small \(p(\boldsymbol {x}^\texttt {CF})\); not in the dataset) where the model fails to classify with high confidence. Thus, actionability and plausibility constraints may be used in conjunction depending on the recourse setting they describe.
3.4 Diversity and Sparsity Constraints
3.4.1 Diversity.
Diverse recourse is often sought in the presence of uncertainty (e.g., unknown user preferences when defining dist and cost). Approaches seeking to generate diverse recourse generally fall in two categories: diversity through multiple runs of the same formulation or diversity via explicitly appending diversity constraints to prior formulations.
In the first camp, Wachter et al. [
234] show that different runs of their gradient-based optimizer over a non-convex model (e.g., multilayer perceptron) results in different solutions as a result of different random seeds. Sharma et al. [
209] show that multiple evolved instances of the genetic-based optimization approach can be used as diverse explanations, hence benefiting from not requiring multiple re-runs of the optimizer. Downs et al. [
69], Mahajan et al. [
146], and Pawelczyk et al. [
183] generate diverse counterfactuals by passing multiple samples from a latent space that is shared between factual and counterfactual instances through a decoder, and filtering those instances that correctly flip the prediction.
In the second camp, Russell [
205] pursues a strategy whereby subsequent runs of the optimizer would prevent changing features in the same manner as prior explanations/recommendations. Karimi et al. [
113] continue in this direction and suggest that subsequent recourse should not fall within an
\(\ell _p\)-ball surrounding any of the earlier explanations/recommendations. Cheng et al. [
46] and Mothilal et al. [
167] present a differentiable constraint that maximizes diversity among generated explanations by maximizing the determinant of a (kernel) matrix of the generated counterfactuals.
3.4.2 Sparsity.
It is often argued that sparser solutions are desirable as they emphasize fewer changes (in explanations) or fewer variables to act upon (in recommendations) and are thus more interpretable for the individual [
154]. Although this is not generally accepted [
182,
225], one can formulate this requirement as an additional constraint, whereby, for example,
\(||\boldsymbol {\delta }||_0 \le s\) or
\(||\boldsymbol {\theta }||_0 \le s\). Formulating sparsity as an additional (hard) constraint rather than optimizing for it in the objective grants the flexibility to optimize for a different object while ensuring that a solution would be sparse.
3.5 Datatypes and Encoding
A common theme in consequential decision-making settings is the use of datatypes that refer to real-world attributes of individuals. As a result, datasets are often tabular
, comprising heterogeneous features, with a mix of numeric (real, integer), binary, categorical, or ordinal variables. Most commonly, Adult [
3], Australian Credit [
71], German Credit [
17], GiveMeCredit [
248], COMPAS [
127], and HELOC [
103], among others, are used, which are highly heterogeneous.
Different feature types obey different statistical properties—for example, the integer-based
heart rate, real-valued
BMI, categorical
blood type, and ordinal
age group differ drastically in their range. Thus, heterogeneous data requires special handling to preserve their semantics. A common approach is to encode each variable according to a predetermined strategy, which preprocesses the data before model training and consequently during recourse generation (for either recourse type, i.e., consequential recommendations or contrastive explanations). For instance, categorical and ordinal features may be encoded using one-hot encoding and thermometer encoding, respectively. To preserve the semantics of each variable during recourse generation, we must also ensure that the generated explanations/recommendations result in counterfactual instances that also satisfy the encoding constraints. For instance, Boolean and linear constraints of the form
\(\sum _j \boldsymbol {x}_{i,j} = 1 ~ \forall ~ \boldsymbol {x}_{i,j} \in \lbrace 0,1\rbrace\) are used to ensure that multiple categories are be simultaneously active, and thermometer-encoded ordinal variables are required to satisfy
\(\boldsymbol {x}_{i,j} \ge \boldsymbol {x}_{i,j+1} ~ \forall ~ \boldsymbol {x}_{i,j} \in \lbrace 0,1\rbrace\). For a detailed overview, refer to the work of Nazabal et al. [
171].
In addition to
tabular data, one may require contrastive explanations for
image-based or
text-based datasets, as summarized in Table
1. For image-based datasets, the algorithm may optionally operate on the raw data, or on super-pixel or other forms of extracted features (e.g., a hidden representation in a neural network). Text-based datasets are also commonly encoded as vectors representing GloVe [
187] or bag-of-words embeddings [
198]. For a recent work that explores the use of counterfactual explanations for functional data, refer to Carrizosa et al. [
42].
3.6 Related Formulations
The problem formulation for recourse generation, and specifically that of contrastive explanations, (
1), is broadly related to several other problems in data mining and ML. For instance, the
cost-minimizing inverse classification problem [
4,
128,
129,
130,
132,
147] aims to identify the “minimum required change to a case in order to reclassify it as a member of a different preferred class” [
147].
Actionable knowledge extraction is employed in data mining to suggest “behaviors which render a state of an instance into a preferred state” [
70] according to a classifier [
37,
38,
39,
112,
247]. Finally,
adversarial perturbations are small imperceptible changes to the input of a classifier that would alter the output prediction to a false and highly confident region [
40,
88,
164,
165,
172,
177,
178,
216]. An additional parallel shared by the preceding methods is in their assumption of a fixed underlying model. Extensions of the preceding, in which model designers anticipate and aim to prevent malicious behavior, exist in the
strategic classification [
67,
99,
104,
120,
135,
140,
155,
158] and
adversarial robustness [
41,
49,
76,
246] literature. For recent expositions on the similarities and contrasts between these methods, refer to works elsewhere [
75,
78,
180].
Whereas there exists strong parallels in their formulations, the differences arise in their intended use cases and guarantees for the stakeholders involved. For example, as opposed to recourse that aims to build trust with affected individuals, the primary use case cited in the actionable knowledge extraction literature is to deliver cost-effective actions to maximize profit or other business objectives. Furthermore, whereas a contrastive explanation aims to inform an individual about ways in which their situation would have led to a desirable outcome, an adversarial perturbation aims to fool the human by being imperceptable (e.g., by leaving the data distribution). In a sense, imperceptability is the antithesis of explainability and trust. Finally, building on the presentation in Section
2, offering consequential recommendations relies on a causal modeling of the world, which is largely ignored by other approaches.
4 Solution
By definition, recourse is offered when an individual is presented with contrastive explanations and consequential recommendations, which can be obtained by solving (
1) and (
2), respectively. Notably, the objective that is to be minimized (i.e.,
dist or
cost) may be non-linear, non-convex, or non-differentiable. Furthermore, without restricting the classifier family, the model constraint also need not be linear, monotonic, or convex. Finally, based on individual-/context-specific restrictions, the problem setting may require optimizing over a constrained set of plausible instances,
\(\mathcal {P}(\mathcal {X})\), or feasible actions,
\(\mathcal {A}(\boldsymbol {x}^\texttt {F})\).
4 Thus, a distance/cost-agnostic and model-agnostic solution with support for plausibility, feasibility, sparsity, and diversity constraints over heterogeneous datasets will in general require complex approaches, trading-off various desirable properties in the process. In the following, we discuss the importance of these properties and provide an overview of utilized solutions.
4.1 Properties
We remark that the optimizer and the resulting solutions should ideally satisfy some desirable properties, as detailed next. In practice, methods typically trade off optimal guarantee \(\delta ^*\), perfect coverage \(\Omega ^*\), or efficient runtime \(\tau ^*\), and may otherwise require prohibitive access to the underlying data or predictive model.
4.1.1 Optimality.
Identified counterfactual instances should ideally be
proximal to the factual instance, corresponding to a small change to the individual’s situation. When optimizing for minimal
dist and
cost in (
1) and (
2), the objective functions and constraints determine the
existence and
multiplicity of recourse. For factual instance
\(\boldsymbol {x}^\texttt {F}\), there may exist zero, one, or multiple
5 optimal solutions, and an ideal optimizer should thus identify (at least) one solution (explanation or recommendation, respectively) if one existed, or terminate and return
N/A otherwise.
4.1.2 Perfect Coverage.
Coverage is defined as the number of individuals for which the algorithm can identify a plausible counterfactual instance (through either recourse type), if at least one solution existed [
113]. Communicating the domain of applicability to users is critical for building trust [
161,
234], and ideally the algorithm offers recourse to all individuals (i.e., perfect coverage).
4.1.3 Efficient Runtime.
As explanations/recommendations are likely to be offered in conversational settings [
101,
105,
157,
212,
240], it is desirable to generate recourse in near-real time. Despite the general underreporting of runtime complexities in the literature, in Table
1 we highlight those algorithms that report efficient wall-clock runtimes that may enable interaction between the algorithm and individual.
4.1.4 Access.
Different optimization approaches may require various levels of access to the underlying dataset or model. Access to the model may involve query access (where only the label is returned), gradient access (where the gradient of the output with respect to the input is requested), class probabilities access (from which one can infer the confidence of the prediction), or complete white-box access (where all model parameters are known).
Naturally, there are practical implications to how much access is permissible in each setting, which further restricts the choice of tools. Consider an organization that seeks to generate recourse for their clients. Unless these algorithms are run in-house by the said organization, it is unlikely that the organization would hand over training data, model parameters, or even non-rate-limited API of their models to a third-party to generate recourse.
4.2 Tools
We consider the richly explored field of optimization [
33,
174,
214] outside the scope of this work and suffice to briefly review the tools used specifically for recourse generation, highlighting their domain of applicability, and relegating technical details to appropriate references. Not only is solving (
1) and (
2) difficult in general settings [
130], it has even been shown to be NP-complete or NP-hard in restricted settings, such as solving for integer-based variables [
12], solving for additive tree models [
16,
55,
219] or neural networks [
116], and solving for quadratic objectives and constraints [
12,
33,
179]. Thus, except for exhaustive search over a potentially uncountable set of solutions, most works pursue
approximate solutions in restricted settings, trading off the preceding desirable properties (see Table
1). Solutions can be broadly categorized as
gradient-based optimization,
model-based,
search-based,
verification-based, and
heuristics-based.
6Under differentiability of the objective and constraints, gradient-based optimization solutions such as FISTA [
26] are employed [
63,
64,
225] to find globally optimal solutions under convex Lagrangian, and first-order methods such as (L-)BFGS or projected gradient descent may be used to identify local optima otherwise. Relatedly, rather than solving recourse for each individual independently, some works pursue a model-based approach, whereby a mapping from factual to counterfactual instances is learned through gradient optimization [
146,
183]. These methods enjoy efficient runtimes at the cost of coverage loss and poor handling of heterogeneous data.
For non-differentiable settings, branch-and-bound-based [
134] approaches split the
search domain into smaller regions within which a solution may be easier to find. Under linearity of the objectives and constraints,
integer linear programming (ILP) algorithms may be used when datatypes are discrete [
55,
223], and
mixed-integer linear programming (MILP) extensions are utilized when some variables are not discrete [
110,
205]. (M)ILP formulations are solved using powerful off-the-shelf solvers such as CPLEX [
54] and Gurobi [
176]. One may also use a combination of iterative binary
search and
verification tools to obtain solutions to (
1) and (
2). Here, the problem is reformulated as a constrained satisfaction problem, where the constraint corresponding to the objective (
dist or
cost) is updated in each iteration to reflect the bounds in which a solution is obtainable [
113,
114,
162]. As with (M)ILP, this approach benefits from the existence of off-the-shelf solvers such as Z3 [
57], CVC4 [
23], and pySMT [
81]. The problem may also be cast and solved as program synthesis [
59,
190] or answer-set programming [
29]. The preceding methods typically offer optimality and perfect coverage while relying on white-box access to the fixed model parameters.
A number of heuristics-based approaches are also explored, such as finding the shortest path (Dijkstra’s algorithm [
53]) between
\(\boldsymbol {x}^\texttt {F}\) and potential
\(\boldsymbol {x}^\texttt {CF}\)s on an empirical graph where edges are placed between similar instances (according to, e.g., Gaussian kernel) [
189]. Finally, genetic-based approaches [
243,
252] find solutions over different evolutions of candidate solutions according to various heuristics [
22,
56,
93,
124,
209] and benefit from being model-/datatype-/norm-agnostic via only requiring query access to the model.
5 Prospects
In the previous sections, we covered the definitions, formulations, and solutions of existing works aiming to offer algorithmic recourse. We showed that generating recourse explanations and recommendations required counterfactual reasoning based on different levels of causal knowledge. Counterfactual reasoning has roots not only in the philosophy of science [
101,
102,
136,
137,
138,
244] but also in the psychology of human agents [
36,
156,
157] and benefits from strong technical foundations [
19,
97]. User studies have demonstrated that causal relationships are assessed by evaluating counterfactuals [
151], and counterfactual simulation is used to predict future events [
83]. Specifically in the context of XAI, it has been shown that counterfactuals can “make the decisions of inscrutable systems intelligible to developers and users” [
36], and that people perform better at predicting model behavior when presented with counterfactual instances [
125]. Organizations seek to deploy counterfactual-based explanations citing their easy-to-understand nature [
30,
31] and GDPR compliance [
234]. Finally, from a practitioner’s standpoint, not only does algorithmic recourse benefit from the widely exercised practice of sharing open source implementations (see Table
1), but various graphical interfaces have also been developed to assist the onboarding of non-technical stakeholders [
46,
87,
241].
There are, however, several implicit assumptions made in existing setups—for example, that the world dynamics are known and do not change, that the predictive (supervised) model is fixed, and that changes only arise due to the actions of the individual seeking recourse. Moreover, in the multi-agent settings considered (with, e.g., banks and loan seekers), agents are assumed to act truthfully with no gaming or false reporting of features, and agents are aligned in the aim to minimize an agreed-upon objective function. In the following, we explore settings in which these assumptions do not hold and offer potential solutions for extending to more realistic recourse settings.
5.1 Beyond Deterministic Recourse
In (
2), we saw that minimal consequential recommendations are generated subject to the constraint that the counterfactual instance,
\(\boldsymbol {x}^\texttt {CF}\), is assigned to be the structural counterfactual of
\(\boldsymbol {x}^\texttt {F}\) under hypothetical actions
\(\boldsymbol {a}\)—that is,
\(\boldsymbol {x}^\texttt {CF}= \boldsymbol {x}^\texttt {SCF}(\boldsymbol {a}; \boldsymbol {x}^\texttt {F})\) [
114]. Computing the structural counterfactual exactly, however, relies on strong assumptions (i.e., the true SCM is an additive noise model and is known). Karimi et al. [
115] show that without complete knowledge of the true SCM, counterfactual analysis cannot be done exactly and thus recourse cannot be offered deterministically. Although they then present methods that offer recourse with high probability, they do so under specification of a causally sufficient graph. Future research in this direction may explore less strict settings, perhaps accounting for hidden confounders [
232], or partially observed graphs [
9,
51,
218], further adding to the uncertainty of recourse recommendations [
27,
122]. Alternatively, sources of stochasticity may enter the recourse process via a non-deterministic decision-making system. For example, it has been demonstrated that for models trained on
selective labels, fair and optimal decisions should be made stochastically [
25,
119,
221].
5.2 Beyond Supervised Recourse
In Section
3.2 we discussed how the standard binary classification setting could be extended to support multi-class classification and regression. Beyond these classic supervised learning settings, an individual may be subject to an automated decision maker that determines a matching of applicants to resources across a population (e.g., kindergarten assignment for children or housing for low-income families). Alternatively, one can expect to generate explanations in more interactive settings, such as for the actions and policies of a reinforcement learning agent [
144,
145,
201,
224] or for recommender systems [
60,
84]. Finally, explanations may also be generated for time-series data [
5,
16,
142], which can be extended to support online data streams and models that change over time [
21,
61,
182,
200,
226].
5.3 Beyond Individualized Recourse
So far, the presented formulations aimed to offer recourse explanations pertaining to a single individual and assumed that recourse recommendations would be undertaken by that individual. However, it is natural to extend the notion of recourse beyond the data subject in question, or beyond a single individual in the population.
An example of the former setting is when the family member of a patient decides on a treatment on their behalf when the patient cannot directly exercise their agency due to incapacitation [
226]. One may also consider common cases in judicial processes where a legal counsel represents and seeks recourse for their client that may then be exercised by another fiduciary. In such settings, the formulation of cost and feasibility of actions may need to be adjusted to account for restrictions on both the subject and the actor.
Alternatively, recourse may be achieved through the collective action of a group of people rather than that of a single individual [
114]. For instance, the efforts of social and political activists may culminate in a
law change that offers better conditions for a group of individuals. In such settings, a (background) variable that is non-actionable (or incurs high cost) on an individual level may be rendered as actionable on a group level, which may in turn bring down the cost for all members of the group. This example also suggests that background variables may capture contextual information (e.g.,
economy) that are not characteristics of, but nonetheless affect, the individual. Furthermore, the individual may not have control over these macro variables that change over time and violate the stationarity assumption of the world. Additionally, explicit consideration for multiple agents [
175], building theories of minds for each agent so as to best persuade [
123], and viewing recourse as a sequential game under non-instantaneous effects of actions taken place [
170,
228] is an underexplored area for future research. Finally, the need to analyze recourse on a subpopulation level may arise due to uncertainty in assumptions [
115] or as an intentional study of other properties of the system, such as fairness [
46,
96,
113,
195,
223], which we explore further in the following.
5.4 On the Interplay of Recourse and Ethical ML
The preceding research questions have primarily focused on one stakeholder: the affected individual. However, giving the right of recourse to individuals should not be considered in a vacuum and independently of the effect that providing explanations/recommendations may have on other stakeholders (e.g., model deployer and regulators), or in relation to other desirable properties (e.g., fairness, security, privacy, robustness), broadly referred to as ethical ML. We explore this interplay next.
5.4.1 Recourse and Fairness.
Fairness in ML is a primary area of study for researchers concerned with uncovering and correcting potentially discriminatory behavior of ML models. In this regard, prior work has informally used the concept of
fairness of recourse as a means to evaluate the
fairness of predictions. For instance, Ustun et al. [
223] look at comparable male/female individuals that were denied a loan and show that a disparity can be detected if the suggested recourse actions (namely,
flipsets) require relatively more effort for individuals of a particular subgroup. Along these lines, Sharma et al. [
209] evaluate group fairness via aggregating and comparing the cost of recourse (namely,
burden) over individuals of different subpopulations. Karimi et al. [
113] show that the addition of feasibility constraints (e.g., non-decreasing
age) that results in an increase in the cost of recourse indicates a reliance of the fixed model on the sensitive attribute
age, which is often considered as legally and socially unfair. Here we clarify that these notions are distinct and would benefit from a proper mathematical study of the relation between them.
The preceding examples suggest that evidence of discriminatory recourse (e.g., reliance on race) may be used to uncover unfair classification. We show, however, that the contrapositive statement does not hold: consider, for example, a 2D dataset comprising of two subgroups (i.e., \(s \in \lbrace 0,1\rbrace\)), where \(p(x|s) = \mathcal {N}(0, 10^s)\). Consider a binary classifier, \(h: \mathbb {R} \times \mathbb {S} \rightarrow \lbrace 0,1\rbrace\), where \(h(x, s) = \text{sign}(x)\). Although the distribution of predictions satisfies demographic parity, the (average) recourse actions required of negatively predicted individuals in \(s=1\) is larger than those in \(s=0\). Thus, we observe unfair recourse even when the predictions are demographically fair.
This contradiction (the conditional and contrapositive not holding simultaneously) can be resolved by considering a new and distinct notion of fairness (i.e.,
fairness of recourse) that does not imply or is not implied by the
fairness of prediction. In this regard,
equalizing recourse was recently presented by Gupta et al. [
96], who offered the first recourse-based (and prediction-independent) notion of fairness. The authors demonstrate that one can directly calibrate for the average distance to the decision boundary to be equalized across different subgroups during the training of both linear and non-linear classifiers. A natural extension would involve considering the
cost of recourse actions, as opposed to the
distance to the decision boundary, in flipping the prediction across subgroups. In summary, recourse may trigger new definitions of fairness to ensure non-discriminatory behavior of ML models, as in the work of Von Kügelgen et al. [
233].
5.4.2 Recourse and Robustness.
Robustness often refers to our expectation that model outputs should not change as a result of (small) changes to the input. In the context of recourse, we expect that similar individuals should receive similar explanations/recommendations, or that recourse suggestions for an individual should be to some extent invariant to the underlying decision-making system trained on the same dataset [
204]. In practice, however, the stability of both gradient-based [
8,
65,
85,
152] and counterfactual-based [
131,
133] explanation systems has been called into question. Interestingly, it has been argued that it is possible for a model to have
robust predictions but
non-robust explanations [
98], and vice versa [
131] (similar to relation between fair predictions and fair recourse). Parallel studies argue that the sparsity of counterfactuals may contribute to non-robust recourse when evaluating explanations generated under different fixed models [
182]. Finally, in the case of consequential recommendations, robustness will be affected by assumptions of the causal generative process (see Figure
1). Carefully reviewing assumptions and exploring such robustness issues in more detail is necessary to build trust in the recourse system and in turn in the algorithmic decision-making system. For recent works in this direction, which study the sensitivity of recourse methods to uncertainty in the features, model, or causal assumptions, refer to works found elsewhere [
62,
66,
160,
173,
211,
222,
230]. Another interesting research question pertains to the generation of recourse when actionability of the recommended actions changes over time or due to external sources (e.g., inflation). Studying and accounting for various failure modes of recourse is important to offer robust and ultimately useful recourse.
5.4.3 Recourse, Security, and Privacy.
Model extraction concerns have been raised in various settings for ML APIs [
141,
197,
220,
236]. In such settings, an adversary aims to obtain a surrogate model,
\(\hat{f}\), that is similar (e.g., in fidelity) to the target model,
f:
Here, an adversary may have access to various model outputs (classification label [
141], class probabilities [
220], etc.) under different query budgets (unlimited, rate- limited, etc. [
44,
108]). Model extraction may be accelerated in the presence of additional information, such as gradients of outputs with respect to inputs
7 [
159], or contrastive explanations [
7]. Related to recourse, and of practical concern [
21,
208,
213,
223], is a study of the ability of an adversary with access to a recourse API in extracting a model. Specifically, we consider a setting in which the adversary has access to a
prediction API and a
recourse API, which given a factual instance,
\(\boldsymbol {x}^\texttt {F}\), returns a nearest contrastive explanation,
\(\boldsymbol {x}^\texttt {*CF}\), using a known distance function,
\(\Delta : \mathcal {X} \times \mathcal {X} \rightarrow \mathbb {R}_+\).
8 How many queries should be made to these APIs to perform a functionally equivalent extraction of
\(f(\cdot)\)?
In a first attack strategy, one could learn a surrogate model on a dataset where factual instances and labels (form the training set or randomly sampled) are augmented with counterfactual instances and counterfactual labels. This idea was explored by Aïvodji et al. [
7], where they demonstrated that a high-fidelity surrogate model can be extracted even under low query budgets. Although easy to understand and implement, this attack strategy implicitly assumes that constructed dataset has independent and identically distributed data, and thus does not make use of the relations between factual and counterfactual pairs.
An alternative attack strategy considers that the model
f can be fully represented by its decision boundaries, or the complementary
decision regions \(\lbrace \mathcal {R}_1, \ldots , \mathcal {R}_l\rbrace\). Every contrastive explanation returned from the recourse API informs us that all instances surrounding the factual instance, up to a distance of
\(\Delta (\boldsymbol {x}^\texttt {F}, \boldsymbol {x}^\texttt {*CF})\), share the same class label as
\(\boldsymbol {x}^\texttt {F}\) according to
f (otherwise, that instance would be the nearest contrastive explanation). More formally,
\(f(\boldsymbol {x}^\texttt {F}) = f(\boldsymbol {x}) ~ \forall ~ \boldsymbol {x}\in \mathcal {B}^\Delta _{\boldsymbol {x}^\texttt {F}}\), where
\(\mathcal {B}^\Delta _{\boldsymbol {x}^\texttt {F}}\) is referred to as the
\(\Delta\)-
factual ball, centered at
\(\boldsymbol {x}^\texttt {F}\) and with radius
\(\Delta (\boldsymbol {x}^\texttt {F}, \boldsymbol {x}^\texttt {*CF})\). The model extraction problem can thus be formulated as the number of factual balls needed to maximally pack all decision regions (Figure
2):
As in other extraction settings, \(\hat{f}\) can then used to infer private information on individuals in the training set, to uncover exploitable system vulnerabilities, or for free internal use. Understanding attack strategies may guide recourse policy and the design of defensive mechanisms to hinder the exploitation of such vulnerabilities.
Surprisingly, a model need not be extracted in the preceding sense to be revealing of sensitive information. Building on the preceding intuition, we note that a single contrastive explanation informs the data subject that there are no instances in a certain vicinity (i.e., within
\(\mathcal {B}^\Delta _{\boldsymbol {x}^\texttt {F}}\)) such that their prediction is different. This information informs the data subject about, for example, whether their similar friend was also denied a loan, violating their predictive privacy. Even under partial knowledge of the friend’s attributes, an adversary may use the information about the shared predictions in
\(\mathcal {B}^\Delta _{\boldsymbol {x}^\texttt {F}}\) to perform membership inference attacks [
210] or infer missing attributes [
72]. This problem is worsened when multiple diverse explanations are generated and is an open problem.
5.4.4 Recourse and Manipulation.
Although a central goal of recourse is to foster trust between an individual and an automated system, it would be simplistic to assume that all parties will act truthfully in this process. For instance, having learned something about the decision-making process (perhaps through recommendations given to similar individuals), an individual may exaggerate some of their attributes for a better chance of favorable treatment [
226]. Trust can also be violated by the recourse-offering party. As discussed earlier, the multiplicity of recourse explanations/recommendations (see Section
4.1.1) may allow for an organization to cherry-pick “the most socially acceptable explanation out of many equally plausible ones” [
21,
98,
126] (also see
fairwashing [
6]). In such cases of misaligned incentives, the oversight of a regulatory body, perhaps with random audits of either party, seems necessary. Another solution may involve mandating a minimum number of diverse recourse offerings, which would conflict with security considerations.
5.5 Toward Unifying Benchmarks
Table
1 presents an overview of the diverse settings in which recourse is sought. Despite the abundance of open source implementations built on robust tools and working well in their respective settings, a comparative benchmark for recourse is lacking. This problem is exacerbated for consequential recommendations that further rely on assumptions about the causal generative process. To make objective progress, however, new contributions should be evaluated against existing methods. Thus, a next step for the community is the curation of an online challenge (e.g., using Kaggle) to benchmark the performance of existing and new methods. To broadly cover practical settings, we envision multiple tracks where authors can submit their generated explanations/recommendations given a fixed classifier, test dataset, and pre-determined
dist/
cost definition, and be evaluated using the metrics defined in Section
4.1. Authors may also submit results that satisfy additional actionability, plausibility, and diversity constraints, and be compared as such.
9 Finally, besides striving for all-encompassing evaluation benchmarks for recourse, a push toward understanding the relations between recourse methods and other explainability methods is quite welcome. In particular, refer to other works [
74,
77,
80,
121] for recent efforts that show how recourse can lead to, or result from, other such methods as attribution methods. User testing of recourse, given the consequential nature of the domain, is more difficult (if at all ethically possible) than explainability methods targeting other stakeholders in non-consequential domains. Nevertheless, with the ultimate objective of building not just explanations but reliable and robust explanations, future research should investigate ways that such uncertainties can be modeled (see Section
5.4.2). This is important not just for the recipient of the recourse recommendation but also for the institution offering such recommendation, as, in a sense, it is a binding contract between the parties to offer some resource when certain conditions are met.
6 Conclusion
Our work started with a case study of a 28-year-old female professional who was denied a loan by an automated decision-making system. We aimed to assist this individual in overcoming their difficult situation—that is, to achieve algorithmic recourse, which was contingent on offering answers to two questions: why and how? We studied the relation between these questions and arrived at distinct responses, namely contrastive explanations and consequential recommendations. Mindful of the goal of recourse, we emphasized minimal consequential recommendations over nearest contrastive explanations, as the former directly optimizes for the least effort from the individual. Furthermore, we noted that offering recourse recommendations automatically implied recourse explanations (through simulation of the causal effect of undertaken actions), whereas the converse would not. In reviewing the literature, however, we observed an underexploration of consequential recommendations, which we attribute to its reliance on additional assumptions at the level of the causal generative process of the world in which actions take place.
We remark that the primary emphasis of recourse is on providing actionable recommendations for overcoming adverse predictions, unlike counterfactual explanations that highlight similar instances with different predictions. This subtlety is also apparent in parallel literature on adversarial ML (see Section
3.6): counterfactual explanations identify the adversarial instances, whereas recourse informs on how to create them. Therefore, to provide actionable recommendations under real-world constraints, we must consider that actions on one subset of variables may have consequences on other variables. Such constraints can be accounted for using the SCM framework, and therefore we primarily take this approach in our work.
In addition to unifying and precisely defining recourse, we present an overview of the many constraints (e.g., actionability, plausibility, diversity, sparsity) that are needed to model realistic recourse settings. With accompanying illustrative examples, we distinguish between the notions of
dist versus
cost, and plausibility versus actionability (feasibility), whose distinctions are largely ignored in the literature. Throughout, we reiterate that these notions are individual-/context dependent, and that formulations cannot arise from a technical perspective alone. We summarize the technical literature in Table
1 as a guide for practitioners looking for methods that satisfy certain properties, and researchers who want to identify open problems and methods to further develop.
Finally, we identify several prospective research directions that challenge the assumptions of existing setups and present extensions to better situate recourse in the broader ethical ML literature. The presented examples and discussion serve to illustrate the diversity of stakeholder needs and a tension between the desirable system properties (fairness, security, privacy, robustness) that we seek to offer alongside recourse. Satisfyingly addressing these needs and navigating the entailed trade-offs may require new definitions and techniques, and relies on the cross-disciplinary expertise of a panel of technical and social scientists. It is our hope that this document may guide further discussion and progress in this direction.