Newer
Older
\begin{figure}
\begin{tikzpicture}[->,>=stealth',node distance=1.5cm, semithick]
\tikzstyle{every state}=[fill=none,draw=black,text=black]
\node[state] (R) {$\leniency$};
\node[state] (X) [right of=R] {$\obsFeatures$};
\node[state] (T) [below of=X] {$\decision$};
\node[state] (Z) [rectangle, right of=X] {$\unobservable$};
\node[state] (Y) [below of=Z] {$\outcome$};
\path (R) edge (T)
(X) edge (T)
edge (Y)
(Z) edge (T)
edge (Y)
(T) edge (Y);
\end{tikzpicture}
\caption{The causal model of the decision makers \human and \machine.
$\leniency$ is the leniency of the decision maker, $\decision$ is a binary decision, $\outcome$ is the outcome that is selectively labeled. Background features $\obsFeatures$ for a subject affect the decision and the outcome. Additional background features $\unobservable$ are visible only to decision maker \human. }\label{fig:causalmodel}
The setting we consider is described in terms of {\it two decision processes}.
In the first one, a decision maker \human considers a case described by a set of features $F$ and makes a binary decision $\decision = \decision_{_\human} \in\{0, 1\}$, nominally referred to as {\it positive} ($\decision = 1$) or non-positive ($\decision = 0$).
Intuitively, in our bail-or-jail example of Section~\ref{sec:introduction}, \human corresponds to the human judge deciding whether to grant bail ($\decision = 1$) or not ($\decision = 0$).
The decision is followed with a binary outcome $\outcome = \outcome_{_\human}$, which is nominally referred to as {\it successful} ($\outcome = 1$) or {\it unsuccessful} ($\outcome = 0$).
An outcome can be {\it unsuccessful} ($\outcome = 0$) only if the decision that preceded it was positive ($\decision = 1$).
If the decision was not positive ($\decision = 0$), then the outcome is considered by default successful ($\outcome = 1$).
Back in our example, the decision of the judge is unsuccessful only if the judge grants bail ($\decision = 1$) but the defendant violates its terms ($\outcome = 0$).
Otherwise, if the decision of the judge was to keep the defendant in jail ($\decision = 0$), the outcome is by default successful ($\outcome = 1$) since there can be no bail violation.
Moreover, we assume that decision maker \human is associated with a leniency level $\leniency$, which determines the fraction of cases for which they produce a positive decision, in expectation.
Formally, for leniency level $\leniency = r\in [0, 1]$, we have
P(\decision = 1 | \leniency = \leniencyValue) = \sum_{\allFeatures} P(\decision = 1, \allFeatures | \leniency = \leniencyValue) = \leniencyValue .
The product of this process is a record $(\human, \obsFeatures, \decision, \outcome)$ that contains only a subset $\obsFeatures\subseteq \allFeatures$ of the features of the case, the decision $\decision$ of the judge and the outcome $\outcome$ -- but leaves no trace for a subset $\unobservable = \allFeatures - \obsFeatures$ of the features.
Intuitively, in our example, $\obsFeatures$ corresponds to publicly recorded information about the bail-or-jail case decided by the judge (e.g., the gender and age of the defendant) and $\unobservable$ corresponds to features that are observed by the judge but do not appear on record (e.g., whether the defendant appeared anxious in court).
The set of records $\dataset = \{(\human, \obsFeatures, \decision, \outcome)\}$ produced by decision maker \human becomes part of what we refer to as the {\bf dataset} -- and the dataset may include records from more than one decision makers.
Figure~\ref{fig:causalmodel} shows the causal diagram that describes the operation of a single decision-maker \human.
In the second decision process, a decision maker \machine considers a case from the dataset -- and makes its own binary decision $\decision = \decision_{_\machine}$ based on recorded features $\obsFeatures$, followed by a binary outcome $\outcome = \outcome_{_\machine}$.
In our example, \machine corresponds to a machine-based automated-decision system that is considered for replacing the human judge in bail-or-jail decisions.
% Notice that we assume \machine has access only to some of the features that were available to \human, to model cases where the system would use only the recorded features and not other ones that would be available to a human judge.
The definitions and semantics of decision $\decision$ and outcome $\outcome$ follow those of the first process.
Moreover, decision maker \machine is also associated with a leniency level $\leniency$, defined as before for \human.
The causal diagram for decision-maker \machine is the same as that for \human (Figure~\ref{fig:causalmodel}), except that \machine does not observe variables $\unobservable$.
\subsection{Model definition} \label{sec:model_definition}
The causal diagram of Figure~\ref{fig:causalmodel} provides the structure of causal relationships for quantities of interest.
%
In addition, we consider $J$ instances $\{\human_j, j = 1..J\}$ of decision makers \human --
and, for the purposes of applied data analysis, we explicate probabilistically our assumptions about them below.
Note that index $j$ refers to decision maker $\human_j$ and \invlogit is the logistic function.
\derivative{\prob{\obsFeatures = \obsFeaturesValue}}{\obsFeaturesValue} = \label{eq:priorx}\\ %probability density for X
\derivative{\prob{\unobservable = \unobservableValue}}{\unobservableValue} = \\ %probability density for Z
\prob{\decision = 1~|~\leniency_j, \obsFeatures, \unobservable}
& = \invlogit(\alpha_{j} + \gamma_{_\obsFeatures} \obsFeatures + \gamma_{_\unobservable}\unobservable), \label{eq:judgemodel}\\
\text{with}\ \alpha_{j} & = \logit(\leniency_j) \label{eq:leniencymodel}\\
\prob{\outcome=1~|~\decision, \obsFeatures, \unobservable}
& =
\begin{cases}
1,~\text{if}~\decision = 0\\
\invlogit(\alpha+\beta_{_\obsFeatures}\obsFeatures + \beta_{_\unobservable}\unobservable),~\text{o/w}
\derivative{\prob{\features \leq \featuresValue}}{\featuresValue} & = \dfrac{1}{\sqrt{2\pi\sigma^2}}\exp\left(-\dfrac{(\featuresValue-\mu)^2}{2\sigma^2}\right) \label{eq:priorx}\\ % Probability density for X. Standard gaussian.
\derivative{\prob{\unobservable \leq \unobservableValue}}{\unobservableValue} & = \dfrac{1}{\sqrt{2\pi\sigma^2}}\exp\left(-\dfrac{(\unobservableValue-\mu)^2}{2\sigma^2}\right) \\ % Probability density for Z. Standard gaussian.
& \ldots . \nonumber \\
\prob{\decision = 1~|~\leniency_j = \leniencyValue, \features = \featuresValue, \unobservable = \unobservableValue}
& = \invlogit(\alpha_{j} + \gamma_{x} x + \gamma_{z}z +\epsilon_\decisionValue), \label{eq:judgemodel}\\
\text{with}\ \alpha_{j} & = \logit(\leniencyValue) \label{eq:leniencymodel}\\
\prob{\outcome=1~|~\decision, \features=\featuresValue, \unobservable=\unobservableValue}
& =
\begin{cases}
1,~\text{if}~\decision = 0\\
\invlogit(\alpha_\outcomeValue+\beta_{x}x+\beta_{z}z +\epsilon_\outcomeValue),~\text{o/w}
As stated in the equations above, we consider normalized features \obsFeatures and \unobservable.
%
Moreover, the probability that the decision maker makes a positive decision takes the form of a logistic function (Equation~\ref{eq:judgemodel}).
%
Note that we are making the simplifying assumption that coefficients $\gamma$ are the same for all defendants, but decision makers are allowed to differ in coefficient $\alpha_j$ so as to model varying leniency levels among them (Equation~\ref{eq:leniencymodel}).
The probability that the outcome is successful conditional on a positive decision (Equation~\ref{eq:defendantmodel}) is also provided by a logistic function, applied on the same features as the logistic formula of (Eq.~\ref{eq:judgemodel}).
In general, these two logistic functions may differ in their coefficients.
However, in many settings, a decision maker would be considered good if the two functions were the same -- i.e., if the probability to make a positive decision was the same as the probability to obtain a successful outcome after a positive decision.
\todo{Michael to Riku}{Define the full model above.}
\rcomment{Defined. Both equations~\ref{eq:judgemodel} and~\ref{eq:defendantmodel} now contain error terms $\epsilon$ which are normally distributed with variance parameters $0.1$ and $0.2^2$ respectively. Question: should the intercept be modelled explicitly as in Eq. \ref{eq:leniencymodel}?}
The quality of a decision maker is measured in terms of its {\bf failure rate} \failurerate -- i.e., the fraction of undesired outcomes ($\outcome=0$) out of all the cases for which a decision is made.
A good decision maker achieves as low failure rate \failurerate as possible.
Note, however, that a decision maker that always makes a negative decision $\decision=0$, has failure rate $\failurerate = 0$, by definition.
For comparisons to be meaningful, we compare decision makers at the same leniency level $\leniency$.
The main challenge in estimating \failurerate is that in general the dataset does not directly provide a way to evaluate \failurerate.
In particular, let us consider the case where we wish to evaluate decision maker \machine\ -- and suppose that \machine is making a decision $\decision_{_\machine}$ for the case corresponding to record $(\human, \obsFeatures, \decision_{_\human}, \outcome_{_\human})$, based on the recorded features \obsFeatures.
Suppose also that the decision by \human was $\decision_{_\human} = 0$, in which case the outcome is always positive, $\outcome_{_\human} = 1$.
If the decision by \machine is $\decision_{_\machine} = 1$, then it is not possible to tell directly from the dataset what its outcome $\outcome_{_\machine}$ would be.
The approach we take to deal with this challenge is to use counterfactual reasoning to infer $\outcome_{_\machine}$(see Section~\ref{sec:imputation} below).
Ultimately, our goal is to obtain an estimate of the failure rate \failurerate for a decision maker \machine.
Given a dataset $\{(\human, \obsFeatures, \decision, \outcome)\}$, and a decision maker \machine, provide an estimate of the failure rate \failurerate.
Sometimes, we may have control over the leniency level of the decision maker we evaluate.
%
In such cases, we would like to evaluate decision maker $\machine = \machine(\leniency = \leniencyValue)$ at various leniency levels $\leniency$.
Ideally, the estimate returned by the evaluation should also be accurate for all levels of leniency.