37
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Causal Diagrams and the Identification of Causal Effects A presentation of Chapters 3 and 4 of: Judea Pearl Causality: Models, Reasoning and Inference Cambridge University Press, 2000

Causal Diagrams and the Identification of Causal Effects

  • Upload
    maree

  • View
    106

  • Download
    1

Embed Size (px)

DESCRIPTION

Causal Diagrams and the Identification of Causal Effects. A presentation of Chapters 3 and 4 of: Judea Pearl Causality: Models, Reasoning and Inference Cambridge University Press, 2000. An Example. This is example is due to Cochran through Pearl [2000, p.66] - PowerPoint PPT Presentation

Citation preview

Page 1: Causal Diagrams and the Identification of Causal Effects

UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINADepartment of Computer Science and

Engineering

Department of Computer Science and Engineering

Causal Diagrams and the Identification of Causal

EffectsA presentation of Chapters 3 and 4 of:

Judea PearlCausality: Models, Reasoning and Inference

Cambridge University Press, 2000

Page 2: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

2UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

An ExampleThis is example is due to Cochran through Pearl [2000, p.66]

“Soil fumigants (X) are used to increase oat crop yields (Y) by controlling the eelworm population (Z).”

Last year’s eelworm population (Z0) is an unknown quantity that is strongly correlated with this year’s population.

Through laboratory analysis of soil samples, we can determine the eelworm populations before and after the treatments (Z1 and Z2). Furthermore , we assume that the fumigants do not affect the growth of eelworms surviving the treatment. Instead, eelworm’s growth depends on the population of birds (B), which is correlated with last year’s eelworm population and hence with the treatment itself. Z3 here represents the eelworm population at the end of the season.”

We wish to assess the total effect of the fumigants on yields. But, controlled randomized experiment are unfeasible and Z0 is unknown.

If we got a correct model, can we obtain consistent estimate of the target quantity – the total effect of the fumigants on yields – through observations?

Page 3: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

3UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Graphs as Models of Interventions

• A causal diagram is a directed acyclic graph G that identifies the causal connections among the variables of interest.

• The causal reading of a DAG is in terms of functional, rather than probabilistic relationships

where pai are the parents of variable Xi in G. ui are mutually independent and represent unobserved factors, including random disturbances.

• We have the same recursive decomposition as in a Bayesian network:

),4.3(,,1),,( niupafX iiii

n

iiin paxPxxP

11 )5.3()|(),,(

Page 4: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

4UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

External Intervention & Causal Effect

• The simplest type of external intervention is one in which a single variable, say Xi, is forced to take on some fixed value xi. We call such an intervention “atomic” and denote it by do(Xi =xi) or do(xi) for short.

• Definition 3.2.1(Causal Effect): Given two disjoint sets of variables, X and Y, the causal

effect of X on Y, denoted either as or as , is a function from X to the space of probability distributions on Y. For each realization x of X, gives the probability of Y = y induced by deleting from the model of (3.4) all equations corresponding to variables in X and substituting X=x in the remaining equations.

)ˆ|( xyP ))(|( xdoyP

)ˆ|( xyP

Page 5: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

5UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Correlation vs. Causation• The genotype theory (Fisher, 1958)

of smoking and lung cancer: smoking and lung cancer are both effects of a genetic predisposition

• Three node network• X( smoking) and Y( lung cancer) are

in lockstep• X precedes Y in time (smoke before

cancer)• But, X does not cause Y, because if

we set X, Y does not change: Y only changes according to the value of U (the genotype)

X Y

U

Page 6: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

6UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Effect of Interventions• From definition 3.2.1, we have:

• Theorem 3.2.2(adjustment for Direct causes)Let PAi denote the set of direct causes of variable Xi, and let Y be any set of variables disjoint of . The effect of the intervention on Y is given by

Where and represent preintervention probabilities.

)12.3(0

)(),|,,()ˆ|,,(

)11.3(0

)|(

)ˆ|,,()ˆ|,,(

)10.3(0

)|()ˆ|,,(

11

1

1

1

ii

iiiiinin

ii

ii

ii

n

in

ii

iijjijin

xxif

xxifpaPpaxxxPxxxP

xxif

xxifpaxP

xxxPxxxP

xxif

xxifpaxpxxxP

),13.3()(),|()ˆ|( ipa

iiii paPpaxyPxyP

),|( ii paxyP )( ipaP

ii PAX )( ii xXdo

Page 7: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

7UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Causal Effect Identifiably• Definition 3.2.4(Causal Effect Identifiability)

• The causal effect of X on Y is identifiable from a graph G if the quantity can be computed uniquely from any positive probability of the observed variables – that is , if for every pair of models M1 and M2 with the same probability distribution for the set of observed variables (v), i.e. and the same graph, i.e.,

• Theorem 3.2.5

• Given a causal diagram G of any Markovian model in which a subset V of variables are measured, the causal effect is identifiable whenever , that is, whenever X, Y, and all parents of variables in X are measured. The expression for is then obtained by adjusting for PAx, as in 3.14

• A special case of Theorem 3.2.5 holds when all variables are assumed to be observed.

)ˆ|()ˆ|(21

xyPxyP MM

)ˆ|( xyP

,0)()(21

vPvP MM

GMGMG )()( 21

)ˆ|( xyP VPAYX X

)ˆ|( xyP

Page 8: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

8UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Example of Nonidentifiability

• The identifiablility of ensures that it is possible to infer the effect of action do(X=x) on Y from passive observations and the causal graph G, which specifies which variables participate in the determination of each variable in the domain.

• To prove nonidentifiability, it is sufficient to present two sets of structural equations that induce identical distributions over observed variables but have different causal effects.

• X,Y is observable, U is not. All of them are binary variables.

• If P(X=0|U) = (0.5,0.5)• P(Y=0|X,U) =

• But we don’t know P(U)• When P(U=0) = 0.5, P(Y|X=0) =(.45,.55)• When P(U=0) = 0.1, P(Y|X=0) =(.73,.27)• So, P(Y|do(X)) is non-identifiable

Y=0 X =0

X= 1

U =0 0.1 0.2

U=1 0.8 0.7

X Y

U

Page 9: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

9UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Example of Nonidentifiability

• The identifiablility of ensures that it is possible to infer the effect of action do(X=x) on Y from passive observations and the causal graph G, which specifies which variables participate in the determination of each variable in the domain.

• To prove nonidentifiability, it is sufficient to present two sets of conditional probability tables that induce identical distributions over observed variables but have different causal effects.

• X,Y is observable, U is not. All of them are binary variables.

• Let P(X,Y,U) be If P(X=0|U) = (0.5,0.5)• P(Y=0|X,U) =

• But we don’t know P(U)• When P(U=0) = 0.5, P(Y|X=0) =(.45,.55)• When P(U=0) = 0.1, P(Y|X=0) =(.73,.27)• So, P(Y|do(X)) is non-identifiable

(U=0, U=1)

X =0 X= 1

Y =0 0.5,0.5 0,0

Y=1 0.0 0.5, 0.5

X Y

U

Page 10: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

10UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Fisher vs. the Surgeon General

X Y

U

Only X and Y are observable

U

X Y X Y

U

Fisher’s Genotype Theory Surgeon General’s Opinion

Page 11: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

11UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Intervention as surgically modified DAGs

Page 12: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

12UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Different Causal Effects for the Same Observations

Page 13: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

13UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Interventions as Variables

The effect of an atomic intervention do(Xi=xi’) is encoded by adding to G a link Fi Xi, where Fi is a new variable taking values in {do(xi’),idle}, xi’ ranges over the domain of Xi, and “idle” represents no intervention. Then we define:

)8.3(

,)(

,)(

,

1

0

)|(

)|(

iiii

iiii

iii

ii

xxandxdoFif

xxandxdoFif

idleFifpaxP

apxP

)9.3())(|,,()ˆ|,,( 11 iinin xdoFxxPxxxP

Page 14: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

14UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Backdoor Criterion• If we can get P’(y|z,x,Fx) = P(y|x,z) and P’(z|Fx) =P(z), then we know is identifiable. • Definition 3.3.1(Back-Door)

A set of variables Z satisfies that back-door criterion relative to an ordered pair of variables(Xi,Xj) in a DAG G if:

(i) no node in Z is a descendant of Xi; and(ii) Z blocks every path between Xi and Xj that contains an arrow into Xi.

Similarly, if X and Y are two disjoint subsets of nodes in G, then Z is said to satisfy that back-door criterion relative to (X,Y) if it satisfies that criterion relative to any pair (Xi,Xj) such that Xi X and Xj Y.

Theorem 3.3.2(Back-Door Adjustment)If a set of variables Z satisfies the back-door criterion relative to (X,Y), then the causal effect of X on Y is identifiable and is given by the formula:

z

xxz

xxx FzPFxzyPFzPFzyPFyPxyP )|(),,|()|(),|()|()ˆ|(

)ˆ|( xyP

)19.3()(),|()|()ˆ|( z

x zPzxyPFyPxyP

Page 15: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

15UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Keshan Disease

Diet Region of China

Family History

Serum Selenium

Genotype

Keshan Disease Congenital

Arrythmis

Enlarged Heart

ECG

To determining the causal effect of Selenium on Keshan Disease, we need to find the variable set Z, called concomitants, which satisfies the back-door criteria.

Z={ Region of China} is a answer.

Z={Genotype} also, but this variable maybe is not observable.

Backdoor Path

Page 16: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

16UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Smoking and the genotype theory

• Consider the relation between smoking(X) and lung cancer(Y).

• the tobacco industry has managed to forestall antismoking legislation by arguing that observed correlation between smoking and lung cancer could be explained by some sort of carcinogenic genotype(U) that involves inborn carving for nicotine

• here, Z is the amount of tar deposited in a person's lungs.

• Can we get ?)ˆ|( xyP

Page 17: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

17UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Effect of Smoking on Lung Cancer in the Presence of Tar Deposits

• We compute

• But • And, by the

backdoor criterion,

xzxPzxyPzyP )|(),|()|(

z

zyPxzPxyP )|()|()|(

)|()|( xzPxzP

Page 18: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

18UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Front-Door Criterion• Definition 3.3.3(Front-Door)

A set of variables Z is said to satisfy the front-door criterion relative to an ordered pair of variables (X,Y) if:(i) Z intercepts all directed paths from X to Y;(ii) there is no back-door path from X to Z; and(iii) all back-door paths from Z to Y are blocked by X.

• Theorem 3.3.4 (Front-Door Adjustment)if Z satisfies the front-door criterion relative to (X,Y) and if P(x,z) >0, then the causal effect of X on Y is identifiable and is given by the formula:

)29.3()(),|()|()ˆ|(

xz

xPzxyPxzPxyP

Page 19: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

19UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Proof of Front_DoorAssume u is the parent set of x, from (i), we have

)22.3(),|()|()|()(),,,( uzyPxzPuxPuPuzyxP

The intervention do(x) removes the factor P(x|u) and induces the post intervention distribution. So

)23.3(),|()|()()ˆ|,,( uzyPxzPuPxuzyP

)24.3()(),|()|()ˆ|( uPuzyPxzPxyPuz

From (ii) and (iii), we also have:)26.3,25.3(),|(),,|()|(),|( uzyPuzxyPandxuPxzuP

x uu

xPxuPuzyPuPuzyP )()|(),|()(),|(

)27.3()(),|()(),|(),,|( x u x

xPzxyPxPzxuPuzxyP

So: )29.3()(),|()|()ˆ|(

xz

xPzxyPxzPxyP

Another proof will be given after the three rules.

Page 20: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

20UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Pearl’s Calculus of Interventions

• Let X,Y and Z be arbitrary disjoint sets of nodes in a causal DAG G. We denote by the graph obtained by deleting from G all arrows pointing to nodes in X. Likewise, we denote by Gx the graph obtained by deleting from G all arrows emerging from nodes in X. to represent the deletion of both incoming and outgoing arrows, we use the notation ( see Figure 3.6 for an illustration). Finally, the expression

stands for the probability of Y=y given that X is held constant at x and that (under this condition) Z=z is observed.

Each of these inference rules follows from the basic interpretation of the “hat” operator as a replacement of the causal mechanism that connects X to its preaction parents by a new mechanism X =x introduced by the intervening force. The result is a sub model characterized by the subgraph Gx.

)ˆ|(/)ˆ|,(),ˆ|( xzPxzyPzxyP

XG

ZXG

Page 21: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

21UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

The Three RulesTheorem 3.4.1 ( Rules of Do calculus)Let G be the directed acyclic graph associated with a causal model as defined in(3.2), and let P(.) stand for the probability distribution induced by that model. For any disjoint subsets of variables X, Y, Z, and W, we have the following rules.

Rule 1 (Insertion/deletion of observations):

)31.3(),|(),ˆ|(),,ˆ|(XGWXZYifwxyPwzxyP

Rule 1 reaffirms d-separation as a valid test for conditional independence in the distribution resulting from the intervention do(X=x), hence the graph . This rule follows from the fact that deleting equations from the system does not introduce any dependencies among the remaining disturbance terms.

XG

Page 22: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

22UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

The Three Rules

Rule 2 provides a condition for an external intervention do(Z=z) to have the same effect on Y as the passive observation Z=z, The condition amounts to {XW} blocking all back-door paths form Z to Y( in ), since retains all (and only ) such paths.

XG ZX

G

Rule 2 ( Action /observation exchange):

)32.3(.),|(),,ˆ|(),ˆ,ˆ|(ZXGWXZYifwzxyPwzxyP

Page 23: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

23UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

The Three RulesRule 3 (Insertion/deletion of actions): ,where Z(W) is

the set of Z-nodes that are not ancestors of any W-node in

)33.3(.),|(),ˆ|(),ˆ,ˆ|()(, WZXGWXZYifwxyPwzxyP

Consider G’ with intervention arcs FZZ added.implies that

So, any path from Z to Y that is not blocked by {X,W} in must end in an arrow pointing to Z, otherwise would not hold. In addition, if there is a path from some Z’ of Z to Y that does and in an arrow pointing to Z’, then W must not be a descendant of Z’, otherwise would not hold. Thus the only paths from Y to Z must end in an arrow pointing at Z, and must end in some member of Z(W). Thus,

),ˆ|(),ˆ,ˆ|( wxyPwzxyP .),|(XGZ WXFY

XGZ WXFY ),|(

XGZ WXFY ),|(

)(,),|(

WZXGWXZY

XG

.X

G

Page 24: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

24UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Usage of the Three Rules Corollary 3.4.2

A causal effect Is identifiable in a model characterized by a graph G if there exists an finite sequence of transformations, each conforming to one of the inference rules in Theorem 3.4.1, that reduces q into a standard (i.e., “hat”-free) probability expression involving observed quantities.

Whether Rules 1-3 are sufficient for deriving all identifiable causal effects remains an open question

)ˆ,,ˆ|,,( 11 mk xxyyPq

Page 25: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

25UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

The Smoking Example

1, Based on rule 2, we have )|()ˆ|( xzPxzP

Page 26: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

26UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

The Smoking Example• 2

Note: we can use the same process to prove back-door formula

x

G

G

x

xPzxyPzyPSo

ruleXYZifzxyPzxyP

ruleXZifxPzxP

zxPzxyPzyP

Z

Z

)(),|()ˆ|(,

2,)|(),|()ˆ,|(

3)()()ˆ|(

).ˆ|()ˆ,|()ˆ|(

Page 27: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

27UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

The Smoking Example• 3

Based on 1,2, and above, we get:

This is also a proof of front door formula

3),ˆ|()ˆ,|(

2),ˆ,ˆ|()ˆ,|(

),ˆ|()ˆ,|()ˆ|(

rulezyPxzyP

rulexzyPxzyP

xzPxzyPxyPz

xz

xPzxyPxzPxyP )(),|()|()ˆ|(

Page 28: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

28UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

The Smoking Example• 4

See 1 and the third formula in 3, we have:

5

See 2

).ˆ|()ˆ,|()ˆ|,( xzPxzyPxzyP

).(),|()ˆ|()ˆ,|()ˆ|,( xPzxyPzxPzxyPzyxP

xxPzxyPxzPxzPzyPxzyP )(),|()|()|()ˆ|()ˆ|,(

Page 29: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

29UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Some Nonidentifiable Models

Why c is not identifiable?

Page 30: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

30UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Some Identifiable Models•

Page 31: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

31UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Why They are Identifiable• (a), (b) rule 2,• (c), (d) back-door• (e) front-door• (f)

• (g)

21 , 221221 )(),|()|(),,|()ˆ|(zz x

zPzxzPzxPxzzyPxyP

).(),|()|(),,|()ˆ|(

2),,|()ˆ,,|(

)ˆ|,()ˆ,,|()ˆ|(

12,

121

2121

,2121

21

21

xPxzzPxzPxzzyPxyP

rulexzzyPxzzyP

xzzPxzzyPxyP

zz x

zz

Page 32: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

32UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Why g?• Z1 block all

directed paths from X to Y

• Z2 blocks all back-door paths between Y and Z1 in

• Putting the pieces together, we obtain the claimed result ),|()(),ˆ|()(

),ˆ|()ˆ|()ˆ|,(

)|'(),ˆ|'(

),,'|(),ˆ,'|(

),ˆ|'(),ˆ,'|(),ˆ|(

),ˆ|(),ˆ,ˆ|(

),ˆ,ˆ|(),,ˆ|(

)ˆ|,(),,ˆ|()ˆ|(

212

2

212

3

21221

2

3

21

21

2

21

2' 12121

21

3

21

21

2

21

, 212121

zxzPzPzxzPzP

zxzPxzPxzzP

zxPzzxP

zzxyPzzxyP

zzxPzzxyPzzyP

zzyPzzxyP

zzxyPzzxyP

xzzPzzxyPxyP

r

rdef

r

r

x

def

r

r

zz

XG

Page 33: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

33UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Several More Nonidentifiable Models

Page 34: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

34UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Completeness of the Three Rules

• Completeness is conjectured by Pearl• Scheme for finding a counterexample

– generate graphs (models)– filter identifiable models by using 3 rules– filter unidentifiable models by using edge

subgraph algorithm with figure 3.9 patterns– rewrite P(y|do(x)) using the rules of

probability – If some formula for P(y|do(x)) without U

exists, a counterexample has been found– If no formula for P(y|do(x)) without U exists,

we add the model to the patterns of figure 3.9

Page 35: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

35UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Identify

Theorem 4.3.1(Galles and Pearl 1995)Let X and Y denote two singleton variables in a semi_Markovian Model characterized by graph G. A sufficient condition for the identifiability of is that G satisfy one of the following four conditions.

1, there is no back-door path from X to Y in G.2,There is no directed path from X to Y in G.3,There exists a set of nodes B that bocks all back-door paths from X to Y so that

is identifiable. ( A special case of this condition occurs when B consists entirely of nondescendents of X, in which case reduces immediately to P(b).)

4,There exist sets of nodes Z1 and Z2 such that:(i) Z1 blocks every directed path form X to Y;(ii) Z2 blocks all back-door paths between Z1 and Y;(iii) Z2 blocks all back-door paths between X and Z1;(iv) Z2 does not activate any back-door paths from X to Y. this condition holds if (i)-(iii) are met and no member of Z2 is a descendant of X.(A special case of condition 4 occurs when Z2 = and there is no back-door path

from X to Z1 or from Z1 to Y.)Theorem 4.3.2 the four conditions of Theorem 4.3.1 are necessary for identifiability

in do calculus.

)ˆ|( xyP

XGY)(X

)ˆ|( xbP

)ˆ|( xbP

XZGZXY1

)|( 11

)|( 21 ZXGZZY

XGZZX )|( 21

)2(1

),|( 21 ZXZGZZYX

Page 36: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

36UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Remarks on Efficiency• Theorem 4.3.3 if is identifiable for one minimal blocking set Bi,

then is identifiable for any other minimal set Bj. ( for condition 3)

• Lemma 4.3.4 If the query is identifiable and if a set of nodes Z lies on a directed path form X to Y, then the query is identifiable.

• Theorem 4.3.5 let Y1 and Y2 be two subsets of nodes such that either (i) no nodes Y1 are descendants of X or (ii) all nodes Y1 and Y2 are descendants of X and all nodes Y1 are nondescendants of Y2. A reducing sequence for exists if and only if there are reducing sequences for both and .

• Theorem 4.3.6 If there exists a set Z1 that meets all of the requirements for Z1 in condition 4, then the set consisting of the children of X intersected with the ancestors of Y will also meet all of the requirements for Z1 in condition 4.

)ˆ|( xbP j

)ˆ|( xyP

)ˆ|( xzP

)ˆ|,( 22 xyyP

)ˆ|( 1 xyP ),ˆ|( 12 yxyP

)ˆ|( xbP i

Page 37: Causal Diagrams and the Identification of Causal Effects

Causality: Models, Reasoning and Inference Chapter 3

37UNIVERSITY OF SOUTH CAROLINAUNIVERSITY OF SOUTH CAROLINA

Department of Computer Science and Engineering

Department of Computer Science and Engineering

Closed-Form Expression for Control Queries

• Function: ClosedForm( ).• Input:Control query of the form .• Output: either a closed-form expression for ,in terms of observed

variables only, or FAIL when query is not identifiable.• 1, if then return P(y).• 2, Otherwise, if then return P(y|x).• 3, Otherwise, let B=BlockingSet(X,Y), and Pb=ClosedForm( )

;if Pb!=FAIL, then return .• 4. Otherwise, let Z1=Children(X) (Y Ancestors(Y)),

Z3=BlockingSet(X,Z1), Z4 = BlockingSet(Z1,Y), and Z2 = Z3 Z4; if Y Z1 and X Z2 then return

• 5, Otherwise return FAILSteps 3 and 4 invoke the function BlockingSet(X,Y), which selects a set of

nodes Z that d-separate X form Y. Step 3 contains a recursive call to the algorithm ClosedForm( ) itself, in order to obtain an expression for causal effect .

)ˆ|( xyP

)ˆ|( xyP

)ˆ|( xyP

XGYXP )( XGYXP )(

xb ˆ|

bPbxbyP *),|(

21 , 221221 )(),|()|(),,|(

zz xzPzxzPzxPxzzyP

)ˆ|( xbP

xb ˆ|