Skip to content
Snippets Groups Projects
Commit fde91e39 authored by Antti Hyttinen's avatar Antti Hyttinen
Browse files

...

parent 50ebfdc2
No related branches found
No related tags found
No related merge requests found
......@@ -55,13 +55,14 @@ Otherwise, if the decision of the judge was to keep the defendant in jail ($\dec
%\end{equation}
%Antti I think this formula is mostly misleading
In the likely case that a dataset includes decision of many decision makers we mark $\human=\{\human_\judgeValue\}$ and use
$\judge$ as an index to the decision maker the subject is assigned.
$\judge$ as an index to the decision maker the subject is assigned. This allows to in particular us to model decision makers differing in their leniency, the portion of subjects they make a positive decision for.
The product of this process is a record $(\judgeValue, \obsFeaturesValue, \decisionValue, \outcomeValue)$, that contains only observations on a subset $\obsFeatures\subseteq \allFeatures$ of the features of the case, the decision $\decision$ of the judge and the outcome $\outcome$ -- but leaves no trace for a subset $\unobservable = \allFeatures \setminus \obsFeatures$ of the features.
%
Intuitively, in our example, $\obsFeatures$ corresponds to publicly recorded information about the bail-or-jail case decided by the judge (e.g., the gender and age of the defendant) and $\unobservable$ corresponds to features that are observed by the judge but do not appear on record (e.g., whether the defendant appeared anxious in court).
%
The set of records $\dataset = \{(\judgeValue, \obsFeaturesValue, \decisionValue, \outcomeValue)\}$ produced by decision maker \human becomes part of what we refer to as the {\bf dataset}.
The set of records $\dataset = \{(\judgeValue, \obsFeaturesValue, \decisionValue, \outcomeValue)\}$ %produced by decision maker \human
becomes part of what we refer to as the {\bf dataset}.
% -- and the dataset generally includes records from \emph{more than one} decision makers, indexed by $\judgeValue$.
%
Figure~\ref{fig:causalmodel} shows the causal diagram of this decision making process.
......@@ -82,7 +83,7 @@ In our example, \machine corresponds to a machine-based automated-decision syste
%
The definitions and semantics of decision $\decision$ and outcome $\outcome$ follow those of the first process.
%
Moreover, decision maker \machine is also associated with a leniency level $\leniency$, defined as before for \human.
%Moreover, decision maker \machine is also associated with a leniency level $\leniency$, defined as before for \human.
%
%The causal diagram for decision-maker \machine is the same as that for \human (Figure~\ref{fig:causalmodel}), except that \machine does not observe variables $\unobservable$.
......@@ -96,6 +97,9 @@ A good decision maker achieves as low failure rate \failurerate as possible.
%
Note, however, that a decision maker that always makes a negative decision $\decision=0$, has failure rate $\failurerate = 0$, by definition.
%
%To produce sensible evaluation of decision maker at varying leniency levels.
%Moreover, decision maker \machine is also associated with a leniency level $\leniency$, defined as before for \human.
%
Thus the evaluation to be meaningful, we evaluate decision makers at the different leniency levels $\leniency$.
%Ultimately, our goal is to obtain an estimate of the failure rate \failurerate for a decision maker \machine.
......@@ -111,11 +115,11 @@ In particular, let us consider the case where we wish to evaluate decision maker
%$(\judgeValue, \obsFeaturesValue, \decisionValue, \outcomeValue)$,
based on the recorded features \obsFeaturesValue.
%
Suppose also that the decision in the data was $\decision = 0$, in which case the outcome is always positive, $\outcome = 1$.
Suppose also that the decision in the data was negative, $\decision = 0$, in which case the outcome is always positive, $\outcome = 1$.
%
If the decision by \machine is $\decision = 1$, then it is not possible to tell directly from the dataset what its outcome $\outcome$ would be.
%
The approach we take to deal with this challenge is to use counterfactual reasoning to infer $\outcome$(see Section~\ref{sec:imputation} below).
The approach we take to deal with this challenge is to use counterfactual reasoning to infer $\outcome$ if we had $\decision = 1$ (see Section~\ref{sec:imputation} below).
%Sometimes, we may have control over the leniency level of the decision maker we evaluate.
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment