Ignorability
In statistics, ignorability is a feature of an experiment design whereby the method of data collection (and the nature of missing data) do not depend on the missing data. A missing data mechanism such as a treatment assignment or survey sampling strategy is "ignorable" if the missing data matrix, which indicates which variables are observed or missing, is independent of the missing data conditional on the observed data.
This idea is part of the Rubin Causal Inference Model, developed by Donald Rubin in collaboration with Paul Rosenbaum in the early 1970s. The exact definition differs between their articles in that period. In one of Rubins articles from 1978 Rubin discuss ignorable assignment mechanisms[1], which can be understood as the way individuals are assigned to treatment groups is irrelevant for the data analysis, given everything that is recorded about that individual. Later, in 1983 [2] Rubin and Rosenbaum rather define strongly ignorable treatment assignment which is a stronger condition, mathematically formulated as , where is a potential outcome given treatment , is some covariates and is the actual treatment.
Pearl [2000] devised a simple graphical criterion, called back-door, that entails ignorability and identifies sets of covariates that achieve this condition.
Ignorability (better called exogeneity) simply means we can ignore how one ended up in one vs. the other group (‘treated’ Tx = 1, or ‘control’ Tx = 0) when it comes to the potential outcome (say Y). It was also called unconfoundedness, selection on the observables, or no omitted variable bias.[3]
Formally it has been written as [Yi1, Yi0] ⊥ Txi, or in words the potential Y outcome of person i had they been treated or not does not depend on whether they have really been (observable) treated or not. We can ignore in other words how people ended up in one vs. the other condition, and treat their potential outcomes as exchangeable. While this seems thick, it becomes clear if we add subscripts for the ‘realized’ and superscripts for the ‘ideal’ (potential) worlds (notation suggested by David Freedman; a visual can help here: potential outcomes simplified). So: Y11/*Y01 are potential Y outcomes had the person been treated (superscript 1), when in reality they have actually been (Y11, subscript 1), or not (*Y01: the * signals this quantity can never be realized or observed, or is fully contrary-to-fact or counterfactual, CF).
Similarly, *Y10/Y00 are potential Y outcomes had the person not been treated (superscript 0), when in reality they have been (*Y10, subscript 1), or not actually (Y00).
Only one of each potential outcome (PO) can be realized, the other cannot, for the same assignment to condition, so when we try to estimate treatment effects, we need something to replace the fully contrary-to-fact ones with observables (or estimate them). When ignorability/exogeneity holds, like when people are randomized to be treated or not, we can ‘replace’ *Y01 with its observable counterpart Y11, and *Y10 with its observable counterpart Y00, not at the individual level Yi’s, but when it comes to averages like E[Yi1 – Yi0], which is exactly the causal treatment effect (TE) one tries to recover.
Because of the ‘consistency rule’, the potential outcomes are the values actually realized, so we can write Yi0 = Yi00 and Yi1 = Yi11 (“the consistency rule states that an individual’s potential outcome under a hypothetical condition that happened to materialize is precisely the outcome experienced by that individual”,[4] p. 872). Hence TE = E[Yi1 – Yi0] = E[Yi11 – Yi00]. Now, by simply adding and subtracting the same fully counterfactual quantity *Y10 we get: E[Yi11 – Yi00] = E[Yi11 –*Y10 +*Y10 - Yi00] = E[Yi11 –*Y10] + E[*Y10 - Yi00] = ATT + {Selection Bias}, where ATT = average treatment effect on the treated [5] and the second term is the bias introduced when people have the choice to belong to either the ‘treated’ or the ‘control’ group. Ignorability, either plain or conditional on some other variables, implies that such selection bias can be ignored, so one can recover (or estimate) the causal effect.
See also
- Missing at random
References
- Rubin, Donald (1978). "Bayesian Inference for Causal Effects: The Role of Randomization". The Annals of Statistics. 6 (1): 34–58. doi:10.1214/aos/1176344064.
- Rubin, Donald B.; Rosenbaum, Paul R. (1983). "The Central Role of the Propensity Score in Observational Studies for Causal Effects". Biometrika. 70 (1): 41–55. doi:10.2307/2335942. JSTOR 2335942.
- Yamamoto, Teppei (2012). "Understanding the Past: Statistical Analysis of Causal Attribution". Journal of Political Science. 56 (1): 237–256. doi:10.1111/j.1540-5907.2011.00539.x. hdl:1721.1/85887.
- Pearl, Judea (2010). "On the consistency rule in causal inference: axiom, definition, assumption, or theorem?". Epidemiology. 21 (6): 872–875. doi:10.1097/EDE.0b013e3181f5d3fd. PMID 20864888.
- Imai, Kosuke (2006). "Misunderstandings between experimentalists and observationalists about causal inference". Journal of the Royal Statistical Society: Series A (Statistics in Society). 171 (2): 481–502. doi:10.1111/j.1467-985X.2007.00527.x.
Further reading
- Gelman, Andrew; Carlin, John B.; Stern, Hal S.; Rubin, Donald B. (2004). Bayesian Data Analysis. New York: Chapman & Hall/CRC.
- Pearl, Judea (2000). Causality: Models, Reasoning, and Inference. Cambridge University Press.
- Jaeger, Manfred (2011). "Ignorability in Statistical and Probabilistic Inference". Journal of Artificial Intelligence Research. 24: 889–917. arXiv:1109.2143. Bibcode:2011arXiv1109.2143J. doi:10.1613/jair.1657.