The Assumptions a Causal DAG encodes Michael Rosenblum March 16, 2010.

Slides:



Advertisements
Similar presentations
Introduction to Graphical Models Brookes Vision Lab Reading Group.
Advertisements

CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Jan, 29, 2014.
A Tutorial on Learning with Bayesian Networks
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Discovering Cyclic Causal Models by Independent Components Analysis Gustavo Lacerda Peter Spirtes Joseph Ramsey Patrik O. Hoyer.
Knowledge Representation and Reasoning University "Politehnica" of Bucharest Department of Computer Science Fall 2010 Adina Magda Florea
Estimating Direct Effects of New HIV Prevention Methods. Focus: the MIRA Trial UCSF: Helen Cheng, Nancy Padian, Steve Shiboski, Ariane van der Straten.
Reasoning Under Uncertainty: Bayesian networks intro Jim Little Uncertainty 4 November 7, 2014 Textbook §6.3, 6.3.1, 6.5, 6.5.1,
Belief Networks Done by: Amna Omar Abdullah Fatima Mahmood Al-Hajjat Najwa Ahmad Al-Zubi.
Bayesian Networks VISA Hyoungjune Yi. BN – Intro. Introduced by Pearl (1986 ) Resembles human reasoning Causal relationship Decision support system/ Expert.
Introduction of Probabilistic Reasoning and Bayesian Networks
Probabilistic Reasoning (2)
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Part II: Graphical models
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Review: Bayesian learning and inference
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
CPSC 322, Lecture 28Slide 1 Reasoning Under Uncertainty: More on BNets structure and construction Computer Science cpsc322, Lecture 28 (Textbook Chpt 6.3)
Today Logistic Regression Decision Trees Redux Graphical Models
Bayesian Networks Alan Ritter.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Bayes Nets. Bayes Nets Quick Intro Topic of much current research Models dependence/independence in probability distributions Graph based - aka “graphical.
Part II: How to make a Bayesian model. Questions you can answer… What would an ideal learner or observer infer from these data? What are the effects of.
Causal Models, Learning Algorithms and their Application to Performance Modeling Jan Lemeire Parallel Systems lab November 15 th 2006.
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
CPS 270: Artificial Intelligence Bayesian networks Instructor: Vincent Conitzer.
Bayes Net Perspectives on Causation and Causal Inference
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
Summary of the Bayes Net Formalism David Danks Institute for Human & Machine Cognition.
A Brief Introduction to Graphical Models
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Chapter 10 The Normal and t Distributions. The Normal Distribution A random variable Z (-∞ ∞) is said to have a standard normal distribution if its probability.
Probabilistic graphical models. Graphical models are a marriage between probability theory and graph theory (Michael Jordan, 1998) A compact representation.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Reasoning Under Uncertainty: Bayesian networks intro CPSC 322 – Uncertainty 4 Textbook §6.3 – March 23, 2011.
Introduction to Bayesian Networks
Bayes Nets recitation Bayes Nets Represent dependencies between variables Compact representation of probability distribution Flu Allergy.
Learning Linear Causal Models Oksana Kohutyuk ComS 673 Spring 2005 Department of Computer Science Iowa State University.
Selected Topics in Graphical Models Petr Šimeček.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2008 Readings: K&F: 3.1, 3.2, –  Carlos.
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 2, 2015.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Machine Learning – Lecture 11
Causal Model Ying Nian Wu UCLA Department of Statistics July 13, 2007 IPAM Summer School.
Reasoning Under Uncertainty: Independence and Inference CPSC 322 – Uncertainty 5 Textbook §6.3.1 (and for HMMs) March 25, 2011.
Mediation: The Causal Inference Approach David A. Kenny.
Belief Networks Kostas Kontogiannis E&CE 457. Belief Networks A belief network is a graph in which the following holds: –A set of random variables makes.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2006 Readings: K&F: 3.1, 3.2, 3.3.
PROBABILISTIC REASONING Heng Ji 04/05, 04/08, 2016.
Variable selection in Regression modelling Simon Thornley.
1 Day 2: Search June 9, 2015 Carnegie Mellon University Center for Causal Discovery.
1Causal Inference and the KMSS. Predicting the future with the right model Stijn Meganck Vrije Universiteit Brussel Department of Electronics and Informatics.
Integrative Genomics I BME 230. Probabilistic Networks Incorporate uncertainty explicitly Capture sparseness of wiring Incorporate multiple kinds of data.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 12
CS 4/527: Artificial Intelligence
CAP 5636 – Advanced Artificial Intelligence
Correlation for a pair of relatives
CAP 5636 – Advanced Artificial Intelligence
CS 188: Artificial Intelligence
Additional notes on random variables
CS 188: Artificial Intelligence
Additional notes on random variables
The Geometric Distributions
Bayesian networks (2) Lirong Xia. Bayesian networks (2) Lirong Xia.
Bayesian networks (2) Lirong Xia.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 12
Presentation transcript:

The Assumptions a Causal DAG encodes Michael Rosenblum March 16, 2010

Overview I describe the set of assumptions encoded by a causal directed acyclic graph (DAG). I use an example from page 15 of the book Causality by Judea Pearl (2009). This presentation includes animations, so it’s best to watch it as a slideshow. (It may not make sense otherwise).

Causal Directed Acyclic Graphs (DAGs) Causal DAG encodes: 1. Assumptions about distribution generating observed data 2. How distribution under hypothetical intervention can be computed from distribution generating observed data T T W W S S R R A A Time of Year Accident Wet Sidewalk RainSprinkler Causal

Causal Directed Acyclic Graphs (DAGs) 1. Assumptions about distribution generating observed data (Markov assumption): Each node is conditionally independent of non-descendents given its parents. E.g.P(R|S,T) = P(R|T), P(W|R,S,T) = P(W|R,S), P(A|W,R,S,T) = P(A|W). Furthermore, these conditional independences hold under any interventions. T T W W S S R R A A Time of Year Accident Wet Sidewalk Rain Sprinkler

Causal Directed Acyclic Graphs (DAGs) 2. Assumptions about distribution under hypothetical interventions: Except for intervened-on nodes, probability of node given its parents is unchanged by interventions. E.g. Under intervention: do[Sprinkler = off] P(T|do[S=off]) = P(T), P(R|T,do[S=off]) = P(R|T), S=off w.p. 1, P(W|R,do[S=off]) = P(W|R,S=off), P(A|W,do[S=off]) = P(A|W). T T W W S S R R A A Time of Year Accident Wet Sidewalk Rain off Sprinkler

Causal Directed Acyclic Graphs (DAGs) 2. Assumptions about distribution under hypothetical interventions: Except for intervened-on nodes, probability of node given its parents is unchanged by interventions. E.g. Under intervention: “do[Wet Sidewalk = wet]” P(T|do[W=wet]) = P(T) P(R|T,do[W=wet]) = P(R|T), P(S|T,do[W=wet]) = P(S|T), W = wet w.p. 1. P(A|do[W=wet]) = P(A|W=wet). T T W W S S R R A A Time of Year Accident Wet Sidewalk Rain wet Sprinkler

Causal Directed Acyclic Graphs (DAGs) Structural Equation Model representation: For u 1,…,u 5 independent, unmeasured variables, and f T, f R, f S, f W, f A unknown functions, we have T=f T (u 1 ), R=f R (T,u 2 ), S=f S (T,u 3 ), W=f W (R,S,u 4 ), A=f A (W,u 5 ). T T W W S S R R A A Time of Year Accident Wet Sidewalk Rain Sprinkler

Causal Directed Acyclic Graphs (DAGs) Structural Equation Model representation: For u 1,…,u 5 independent, unmeasured variables, and f T, f R, f S, f W, f A unknown functions, we have T=f T (u 1 ), R=f R (T,u 2 ), S=f S (T,u 3 ), S=off, W=f W (R,S,u 4 ), W=f W (R,off,u 4 ) A=f A (W,u 5 ). Setting S=off gives mutilated set of equations. T T W W S S R R A A Time of Year Accident Wet Sidewalk Rain off Sprinkler

Causal Directed Acyclic Graphs (DAGs) Structural Equation Model representation: For u 1,…,u 5 independent, unmeasured variables, and f T, f R, f S, f W, f A unknown functions, we have T=f T (u 1 ), R=f R (T,u 2 ), S=f S (T,u 3 ), W=f W (R,S,u 4 ), A=f A (W,u 5 ). T T W W S S R R A A Time of Year Accident Wet Sidewalk Rain Sprinkler

Causal Directed Acyclic Graphs (DAGs) Structural Equation Model representation: For u 1,…,u 5 independent, unmeasured variables, and f T, f R, f S, f W, f A unknown functions, we have T=f T (u 1 ), R=f R (T,u 2 ), S=f S (T,u 3 ), W=f W (R,S,u 4 ), W=wet A=f A (W,u 5 ), A=f A (wet,u 5 ). Setting W=wet gives mutilated set of equations. T T W W S S R R A A Time of Year Accident Wet Sidewalk Rain wet Sprinkler

Counterfactuals Can Represent Counterfactuals using Structural Eqn. Models: T=f T (u 1 ), R=f R (T,u 2 ), S=f S (T,u 3 ), W=f W (R,S,u 4 ), A=f A (W,u 5 ). E.g. Counterfactual value of A setting W=wet is f A (wet,u 5 ); Counterfactual value of W setting S=off is f W (R,off,u 4 ). T T W W S S R R A A Time of Year Accident Wet Sidewalk Rain Sprinkler

MIRA Trial Example Randomized trial 2 study arms (diaphragm arm, control arm) Intensive condom counseling and provision to both arms We want to estimate effect of intervention assignment on HIV infection, holding condom use fixed. That is, we want: P(H=1|do[R=1,C=never])- P(H=1|do[R=0,C=never]). R R C C H H Study Arm Condom Use HIV Infection never

MIRA Trial Example We want to estimate effect of intervention assignment on HIV infection, holding condom use fixed. That is, we want: P(H=1|do[R=1,C=never])- P(H=1|do[R=0,C=never]). This causal DAG would imply: P(H=1|do[R=1,C=never]) =P(H=1|R=1,C=never). R R C C H H Study Arm Condom Use HIV Infection never

MIRA Trial Example Potential Confounders of effect of condom use on HIV infection: N = Number of Partners Then causal DAG implies: P(H=1|N,do[R=1,C=never]) =P(H=1|N,R=1,C=never). Can multiply each side by P(N) and sum over values of N to get P(H=1|do[R=1,C=never]). R R C C H H Study Arm Condom Use HIV Infection N N

MIRA Trial Example Potential Confounders of effect of condom use on HIV infection: N = Number of Partners P = Main Partner Seropositive Then P(H=1|N,P,do[R=1,C=never]) =P(H=1|N,P,R=1,C=never). But we don’t observe P!  R R C C H H Study Arm Condom Use HIV Infection N N P P

MIRA Trial Example Unmeasured (hidden) variables represented by dashed circle and dashed lines. In determining what assumptions a Causal DAG encodes, unmeasured variables treated just like measured variables. E.g. P(R|N) = P(R), P(H|N,P,R,do[C=never]) = P(H|N,P,R,C=never). R R C C H H Study Arm Condom Use HIV Infection N N P P