A Causal Calculus for Statistical Research

Judea Pearl
Pre-proceedings of the Fifth International Workshop on Artificial Intelligence and Statistics, PMLR R0:430-449, 1995.

Abstract

Many statisticians are reluctant to deal with problems involving causal considerations because we lack the mathematical notation for distinguishing causal influence from statistical association. To address this problem, a notation is proposed that admits two conditioning operators: ordinary Bayes conditioning, $P(y \mid X=x)$, and causal conditioning, $P(y \mid \operatorname{set}(X=x))$, that is, conditioning $P(y)$ on holding $X$ constant (at $x)$ by external intervention. This distinction, which will be supported by three rules of inference, will permit us to derive probability expressions for the combined effect of observations and interventions. The resulting calculus yields simple solutions to a number of interesting problems in causal inference and should allow rank-and-file researchers to tackle practical problems that are generally considered too hard, or impossible. Examples are: 1. Deciding whether the information available in a given observational study is sufficient for obtaining consistent estimates of causal effects. 2. Deriving algebraic expressions for causal effect estimands. 3. Selecting measurements that would render randomized experiments unnecessary. 4. Selecting a set of indirect (randomized) experiments to replace direct experiments that are either infeasible or too expensive. 5. Predicting (or bounding) the efficacy of treatments from randomized trials with imperfect compliance. Starting with nonparametric specification of structural equations, the paper establishes the semantics necessary for a theory of interventions, presents the three rules of inference, demonstrates the use of the resulting calculus on a number of examples, and establishes an operational definition of structural equations.

Cite this Paper


BibTeX
@InProceedings{pmlr-vR0-pearl95a, title = {A Causal Calculus for Statistical Research}, author = {Pearl, Judea}, booktitle = {Pre-proceedings of the Fifth International Workshop on Artificial Intelligence and Statistics}, pages = {430--449}, year = {1995}, editor = {Fisher, Doug and Lenz, Hans-Joachim}, volume = {R0}, series = {Proceedings of Machine Learning Research}, month = {04--07 Jan}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/r0/pearl95a/pearl95a.pdf}, url = {https://proceedings.mlr.press/r0/pearl95a.html}, abstract = {Many statisticians are reluctant to deal with problems involving causal considerations because we lack the mathematical notation for distinguishing causal influence from statistical association. To address this problem, a notation is proposed that admits two conditioning operators: ordinary Bayes conditioning, $P(y \mid X=x)$, and causal conditioning, $P(y \mid \operatorname{set}(X=x))$, that is, conditioning $P(y)$ on holding $X$ constant (at $x)$ by external intervention. This distinction, which will be supported by three rules of inference, will permit us to derive probability expressions for the combined effect of observations and interventions. The resulting calculus yields simple solutions to a number of interesting problems in causal inference and should allow rank-and-file researchers to tackle practical problems that are generally considered too hard, or impossible. Examples are: 1. Deciding whether the information available in a given observational study is sufficient for obtaining consistent estimates of causal effects. 2. Deriving algebraic expressions for causal effect estimands. 3. Selecting measurements that would render randomized experiments unnecessary. 4. Selecting a set of indirect (randomized) experiments to replace direct experiments that are either infeasible or too expensive. 5. Predicting (or bounding) the efficacy of treatments from randomized trials with imperfect compliance. Starting with nonparametric specification of structural equations, the paper establishes the semantics necessary for a theory of interventions, presents the three rules of inference, demonstrates the use of the resulting calculus on a number of examples, and establishes an operational definition of structural equations.}, note = {Reissued by PMLR on 01 May 2022.} }
Endnote
%0 Conference Paper %T A Causal Calculus for Statistical Research %A Judea Pearl %B Pre-proceedings of the Fifth International Workshop on Artificial Intelligence and Statistics %C Proceedings of Machine Learning Research %D 1995 %E Doug Fisher %E Hans-Joachim Lenz %F pmlr-vR0-pearl95a %I PMLR %P 430--449 %U https://proceedings.mlr.press/r0/pearl95a.html %V R0 %X Many statisticians are reluctant to deal with problems involving causal considerations because we lack the mathematical notation for distinguishing causal influence from statistical association. To address this problem, a notation is proposed that admits two conditioning operators: ordinary Bayes conditioning, $P(y \mid X=x)$, and causal conditioning, $P(y \mid \operatorname{set}(X=x))$, that is, conditioning $P(y)$ on holding $X$ constant (at $x)$ by external intervention. This distinction, which will be supported by three rules of inference, will permit us to derive probability expressions for the combined effect of observations and interventions. The resulting calculus yields simple solutions to a number of interesting problems in causal inference and should allow rank-and-file researchers to tackle practical problems that are generally considered too hard, or impossible. Examples are: 1. Deciding whether the information available in a given observational study is sufficient for obtaining consistent estimates of causal effects. 2. Deriving algebraic expressions for causal effect estimands. 3. Selecting measurements that would render randomized experiments unnecessary. 4. Selecting a set of indirect (randomized) experiments to replace direct experiments that are either infeasible or too expensive. 5. Predicting (or bounding) the efficacy of treatments from randomized trials with imperfect compliance. Starting with nonparametric specification of structural equations, the paper establishes the semantics necessary for a theory of interventions, presents the three rules of inference, demonstrates the use of the resulting calculus on a number of examples, and establishes an operational definition of structural equations. %Z Reissued by PMLR on 01 May 2022.
APA
Pearl, J.. (1995). A Causal Calculus for Statistical Research. Pre-proceedings of the Fifth International Workshop on Artificial Intelligence and Statistics, in Proceedings of Machine Learning Research R0:430-449 Available from https://proceedings.mlr.press/r0/pearl95a.html. Reissued by PMLR on 01 May 2022.

Related Material