@@ -110,37 +110,49 @@ Taking into account that we need to learn parameters from the data we integrate
To make inference we obviously have to learn the parametric model from the data instead of fixed functions of the previous section. We can define the model as a probabilistic due to the simplification of the counterfactual expression in the previous section.
We assume feature vectors $\obsFeaturesValue$ and $\unobservableValue$ representing risk can be consensed to unidimension risk values, for example by ... . Furthermore we assume their distribution as Gaussian. Since $Z$ is unobserved we can assume its variance to be 1.
We assume feature vectors $\obsFeaturesValue$ and $\unobservableValue$ representing risk can be consensed to unidimension risk factors, for example by propensity scores. Furthermore we assume their distribution as Gaussian. Since $Z$ is unobserved we can assume its variance to be 1 without loss of generalization.
According to the selective labels setting we have $Y=0$ whenever $T=0$. When $T=1$, subject behaviour is modeled as logistic regression over the risk factors and a noise terms.
$\parameters=\{\alpha_\outcomeValue, \alpha_j, \beta_\obsFeaturesValue, \beta_\unobservableValue, \gamma_\obsFeaturesValue, \gamma_\unobservableValue\}$. \acomment{Where are the variance parameters?} Our estimate is simply integrating over the posterior of these variables.
Here \invlogit is the standard \acomment{inverse?} logistic function.
Since the decision are ultimately based on the risk factors for behaviour, we model the decisions similarly according to a logistic regression over the risk factors, leniency for the decision maker and a noise term:
Note that index $j$ refers to decision maker $\human_j$. Parameter $\alpha_{j}$ provides for the leniency of a decision maker by $\logit(\leniencyValue_j)$. The decision makers in the data differ from each other only by leniency.