SIMPSON’S PARADOX EXPLAINED
Normal Deviate 2013-06-20
SIMPSON’S PARADOX EXPLAINED
Imagine a treatment with the following properties:
The treatment is good for men (E1) The treatment is good for women (E2) The treatment bad overall (E3)That’s the essence of Simpson’s paradox. But there is no such treatment. Statements (E1), (E2) and (E3) cannot all be true simultaneously.
Simpson’s paradox occurs when people equate three probabilistic statements (P1), (P2), (P3) described below, with the statements (E1), (E2), (E3) above. It turns out that (P1), (P2), (P3) can all be true. But, to repeat: (E1), (E2), (E3) cannot all be true.
The paradox is NOT that (P1), (P2), (P3) are all true. The paradox only occurs if you mistakenly equate (P1-P3) with (E1-E3).
1. Details
Throughout this post I’ll assume we have essentially an infinite sample size. The confusion about Simpson’s paradox is about population quantities so we needn’t focus on sampling error.
Assume that is binary. The key probability statements are:
(P1) (P2) (P3)Here, is the outcome ( means success, means failure), is treatment ( means treated, means not-treated) and is sex ( means male, means female).
It is easy to construct numerical examples where (P1), (P2) and (P3) are all true. The confusion arises if we equate the three probability statements (P1-P3) with the English sentences (E1-E3).
To summarize: it is possible for (P1), (P2), (P3) to all be true. It is NOT possible for (E1), (E2), (E3) to all be true. The error is in equating (P1-P3) with (E1-E3).
To capture the English statements above, we need causal language, either counterfactuals or causal directed graphs. Either will do. I’ll use counterfactuals. (For an equivalent explanation using causal graphs, see Pearl 2000). Thus, we introduce where is your outcome if treated and is your outcome if not treated. We observe
In other words, if we observe and if we observe . We never observe both and on any person. The correct translation of (E1), (E2) and (E3) is:
(C1) (C2) (C3)These three statements cannot simultaneously be true. Indeed, if the first two statements hold then
Thus, (C1)+(C2) implies (not C3). If the treatment is good for mean and good for women then of course it is good overall.
To summarize, in general we have
and, moreover (E3) cannot hold if both (E1) and (E2) hold.
The key is that, in general,
and
In other words, correlation (left hand side) is not equal to causation (right hand side).
Now, if treatment is randomly assigned, then is independent of and
and so we will not observe the reversal, even for the correlation statements. That is, when is randomly assigned, (P1-P3) cannot all hold.
In the non-randomized case, we can only recover the causal effect by conditioning on all possible confounding variables . (Recall a confounding variable is a variable that affects both and .) This is because is independent of conditional on (that’s what it means to control for confounders) and we have
and similarly, and so
which reduces the causal effect into a formula involving only observables. This is usually called the adjusted treatment effect. Now, if it should happen that there is only one confounding variable and it happens to be our variable then
In this case we get the correct causal conclusion by conditioning on . That’s why people usually call the conditional answer correct and the unconditional statement misleading. But this is only true if is a confounding variable and, in fact, is the only confounding variable.
2. What’s the Right Answer?
Some texts make it seem as if the conditional answers (P2) and (P3) are correct and (P1) is wrong. This is not necessarily true. There are several possibilities:
- is a confounder and is the only confounder. Then (P3) is misleading and (P1) and (P2) are correct causal statements.
- There is no confounder. Moreover, conditioning on causes confounding. Yes, contrary to popular belief, conditioning on a non-confounder can sometimes cause confounding. (I discuss this more below.) In this case, (P3) is correct and (P1) and (P2) are misleading.
- is a confounder but there are other unobserved confounders. In this case, none of (P1), (P2) or (P3) are causally meaningful.
Without causal language— counterfactuals or causal graphs— it is impossible to describe Simpson’s paradox correctly. For example, Lindley and Novick (1981) tried to explain Simpson’s paradox using exchangeability. It doesn’t work. This is not meant to impugn Lindley or Novick— known for their important and influential work— but just to point out that you need the right language to correctly resolve a paradox. In this case, you need the language of causation.
3. Conditioning on Nonconfounders
I mentioned that conditioning on a non-confounder can actually create confounding. Pearl calls this -bias. (For those familar with causal graphs, this is bascially the fact that conditioning on a collider creates dependence.)
To elaborate, suppose I want to estimate the causal effect
If is a confounder (and is the only confounder) then we have the identity
that is, the causal effect is equal to the adjusted treatment effect. Let us write
to indicate that the formula for is a function of the distributions and .
But, if is not a confounder, does the equality still hold? To simplify the discussion assume that there are no other confounders. Either is a confounder or there are no confounders. What is the correct identity for ? Is it
or
The answer (under these assumptions) is this: if is not a confounder then the first identity is correct and if is a confounder then the second identity is correct. In the first case .
Now, when there are no confounders, the first identity is correct. But is the second actually incorrect or will it give the same answer as the first? The answer is: sometimes they gave the same answer but it is possible to construct situations where
(This is something that Judea Pearl has often pointed out.) In these cases, the correct formula for the causal effect is the first one and it does not involve conditioning on . Put simply, conditioning on a non-confounder can (in certain situations) actually cause confounding.
4. Continuous Version
A continuous version of Simpson’s paradox, sometimes called the ecological fallacy, looks like this:
Here we see that increasing doses of drug lead to poorer outcomes (left plot). But when we separate the data by sex () the drug shows better outcomes for higher doses for both males and females.
5. A Blog Argument Resolved?
We saw that in some cases is a function of but in other cases it is only a function of . This fact led to an interesting exchange between Andrew Gelman and Judea Pearl and, later, Pearl’s student Elias Bareinboim. See, for example, here and here and here.
As I recall (Warning! my memory could be wrong), Pearl and Bareinboim were arguing that in some cases, the correct formula for the causal effect was the first one above which does not involve conditioning on . Andrew was arguing that conditioning was a good thing to do. This led to a heated exchange.
But I think they were talking past each other. When they said that one should not condition, they meant that the formula for the causal effect does not involve the conditional distribution . Andrew was talking about conditioning as a tool in data analysis. They were each using the word conditioning but they were referring to two different things. At least, that’s how it appeared to me.
6. References
For numerical examples of Simpson’s paradox, see the Wikipedia article.
Lindley, Dennis V and Novick, Melvin R. (1981). The role of exchangeability in inference. The Annals of Statistics, 9, 45-58.
Pearl, J. (2000). Causality: models, reasoning and inference, {Cambridge Univ Press}.