Causal Inference and Digital Causality Lab
University of Hamburg
Throughout this course, we will occasionally refer to a case study from the real-world (Source: zeit.de)
Question: How would you define a causal effect in this setting?
We define a causal effect for an individual as: Treatment \(A\) has a causal effect on \(Y\) if \[Y^{a=1} \neq Y^{a=0}.\]
Note that the definition is based on counterfactual outcomes \(Y^{a=0}, Y^{a=1}\).
Only one is factual, i.e., observable in reality.
Thus, individual effects cannot be identified, in general. This is sometimes called the fundamental problem of causal inference.
i | Y^{a=0} | Y^{a=1} |
---|---|---|
1 | 0 | 1 |
2 | 1 | 1 |
3 | 0 | 1 |
4 | 1 | 0 |
5 | 0 | 1 |
i | Y^{a=0} | Y^{a=1} | a |
---|---|---|---|
1 | 0 | 0 | |
2 | 1 | 0 | |
3 | 1 | 1 | |
4 | 0 | 1 | |
5 | 0 | 0 |
An average causal effect is present if \[P(Y^{a=1} = 1) \neq P(Y^{a=0} = 1),\]
or, more generally for nondichotomous outcomes \[E(Y^{a=1}) \neq E(Y^{a=0}).\]
Present if outcome depends on other individuals’ treatment status \(\Rightarrow\) \(Y_i^{a}\) is not well defined.
SUTVA-Assumption maintained throughout this course: “Stable-Unit-Treatment-Value Assumption” (Rubin 1980)
There could be different versions of a “treatment” \(\Rightarrow\) \(Y_i^{a}\) is not well defined.
Assumption of “treatment variation irrelevance” throughout this course.
\[\begin{align*} & P(Y^{a=1} = 1) - P(Y^{a=0} = 1) = 0 \quad \quad \text{ (Causal risk difference)} \\ \\ & \frac{P(Y^{a=1} = 1)}{P(Y^{a=0} = 1)} = 1 \quad \quad \quad \quad \quad \quad \quad \quad \quad \text{ (Causal risk ratio)} \\ \\ & \frac{ P(Y^{a=1} = 1) / P(Y^{a=1} = 0)}{ P(Y^{a=0} = 1) / P(Y^{a=0} = 0)} = 1 \quad \quad \quad \text{ (Causal odds ratio)} \end{align*}\]
Question: Can you interpret the different quantifications of the causal effect in the case study?
i | Y^{a=0} | Y^{a=1} |
---|---|---|
1 | 0 | 1 |
2 | 1 | 1 |
3 | 0 | 1 |
4 | 1 | 0 |
5 | 0 | 1 |
Hence, even if we could observe counterfactual outcomes, we cannot simply read them off. We have to use statistical estimation.
Also, simple calculations of causal effects are not meaningful if we work with samples. We have to perform proper statistical tests to reject causal hypotheses.
Further problems inherent to causal inference:
\(\Rightarrow\) These problems are the reasons why we need a profound statistical knowledge to perform causal research in practice.
\(\Rightarrow\) We will talk about the statistical methods later. First, we need to get familiar with the conceptual basics.
Suppose, we observe the following data in our case study.
i | Y | A |
---|---|---|
1 | 0 | 0 |
2 | 1 | 0 |
3 | 1 | 1 |
4 | 0 | 1 |
5 | 0 | 0 |
The probability to die if treated is
\(P(Y=1|A=1)=\) 0.5.
When \(P(Y=1|A=1)=P(Y=1|A=0)\), we say that \(Y\) and \(A\) are independent, i.e. \(Y \perp \!\!\! \perp A\), or equivalently, \(A \perp \!\!\! \perp Y\).
We say, that \(A\) and \(Y\) are dependent or associated when \(P(Y=1|A=1) \neq P(Y=1|A=0)\).
In our example, \(P(Y=1|A=0)=\) 0.3333 \(\Rightarrow\) \(A\) and \(Y\) are associated.
The associational risk difference (ARD), risk ratio (ARR), and odds ratio (AOR) quantify the strength of the association. These measures are subject to random variability!
\[\begin{align*} & P(Y=1| A=1) - P(Y=1| A = 0) \text{ (Associational risk difference)} \\ \\ & \frac{P(Y=1| A=1)}{P(Y=1| A = 0)} \quad \quad \quad \quad \quad \quad \quad \quad \quad \text{ (Associational risk ratio)} \\ \\ & \frac{ P(Y=1| A=1) / P(Y=0| A = 1)}{ P(Y=1| A=0) / P(Y=0| A = 0)} \quad \quad \text{ (Associational odds ratio)} \end{align*}\]
Or, as stated in Hernán and Robins (2010, 12) :
“The question is then under which conditions real world data can be used for causal inference.”
Causal Inference & DCL