Exploratory studies: you have empirical data and you want to know what sorts of causal models are consistent with it. Confirmatory tests: you have a causal.

Slides:



Advertisements
Similar presentations
Inferential Statistics and t - tests
Advertisements

Structural Equation Modeling
1. Person 1 1.Stress 2.Depression 3. Religious Coping Task: learn causal model 2 Data from Bongjae Lee, described in Silva et al
Topic Outline Motivation Representing/Modeling Causal Systems
Weakening the Causal Faithfulness Assumption
Bayesian Networks, Winter Yoav Haimovitch & Ariel Raviv 1.
Outline 1)Motivation 2)Representing/Modeling Causal Systems 3)Estimation and Updating 4)Model Search 5)Linear Latent Variable Models 6)Case Study: fMRI.
Peter Spirtes, Jiji Zhang 1. Faithfulness comes in several flavors and is a kind of principle that selects simpler (in a certain sense) over more complicated.
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
Omitted Variable Bias Methods of Economic Investigation Lecture 7 1.
Learning Causality Some slides are from Judea Pearl’s class lecture
AP Statistics – Chapter 9 Test Review
Correlation Chapter 9.
1 By Gil Kalai Institute of Mathematics and Center for Rationality, Hebrew University, Jerusalem, Israel presented by: Yair Cymbalista.
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Review: What influences confidence intervals?
Discussion #36 Spanning Trees
NP-Complete Problems Reading Material: Chapter 10 Sections 1, 2, 3, and 4 only.
Bayesian Network Representation Continued
Evaluation.
Michael Bender - SUNY Stony Brook Dana Ron - Tel Aviv University Testing Acyclicity of Directed Graphs in Sublinear Time.
Inferring Causal Graphs Computing 882 Simon Fraser University Spring 2002.
Biol 500: basic statistics
1 gR2002 Peter Spirtes Carnegie Mellon University.
Experimental Evaluation
Causal Modeling for Anomaly Detection Andrew Arnold Machine Learning Department, Carnegie Mellon University Summer Project with Naoki Abe Predictive Modeling.
Expressions and Equations
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
1 Day 2: Search June 9, 2015 Carnegie Mellon University Center for Causal Discovery.
CAUSAL SEARCH IN THE REAL WORLD. A menu of topics  Some real-world challenges:  Convergence & error bounds  Sample selection bias  Simpson’s paradox.
Bayes Net Perspectives on Causation and Causal Inference
1 Part 2 Automatically Identifying and Measuring Latent Variables for Causal Theorizing.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Bayes’ Nets  A Bayes’ net is an efficient encoding of a probabilistic model of a domain  Questions we can ask:  Inference: given a fixed BN, what is.
Demonstrate Basic Algebra Skills
Methods and Solving Equations
Bill Shipley, département de biologie Université de Sherbrooke Sherbrooke (Qc) Canada
Homework Review notes Complete Worksheet #1. Homework Let A = {a,b,c,d}, B = {a,b,c,d,e}, C = {a,d}, D = {b, c} Describe any subset relationships. 1.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Hypotheses tests for means
Comp. Genomics Recitation 3 The statistics of database searching.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 1): Two-tail Tests & Confidence Intervals Fall, 2008.
DIRECTIONAL HYPOTHESIS The 1-tailed test: –Instead of dividing alpha by 2, you are looking for unlikely outcomes on only 1 side of the distribution –No.
Course files
Spanning Trees. A spanning tree for a connected, undirected graph G is a graph S consisting of the nodes of G together with enough edges of G such that:
Linear Equation The equation Which express the real or complex quantity b in terms of the unknowns and the real or complex constants is called a linear.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Lesson 3: More with Combining Like Terms
3-1 MGMG 522 : Session #3 Hypothesis Testing (Ch. 5)
NP-COMPLETE PROBLEMS. Admin  Two more assignments…  No office hours on tomorrow.
Review I A student researcher obtains a random sample of UMD students and finds that 55% report using an illegally obtained stimulant to study in the past.
Chapter 12: Correlation and Linear Regression 1.
Hypothesis Testing and the T Test. First: Lets Remember Z Scores So: you received a 75 on a test. How did you do? If I said the mean was 72 what do you.
1 BN Semantics 2 – The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 20 th, 2006 Readings: K&F:
The Visual Causality Analyst: An Interactive Interface for Causal Reasoning Jun Wang, Stony Brook University Klaus Mueller, Stony Brook University, SUNY.
Ordering of Events in Distributed Systems UNIVERSITY of WISCONSIN-MADISON Computer Sciences Department CS 739 Distributed Systems Andrea C. Arpaci-Dusseau.
1 BN Semantics 2 – Representation Theorem The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 17.
Hypothesis test flow chart
Graphs + Shortest Paths David Kauchak cs302 Spring 2013.
Chapter 7: Hypothesis Testing. Learning Objectives Describe the process of hypothesis testing Correctly state hypotheses Distinguish between one-tailed.
1 Day 2: Search June 9, 2015 Carnegie Mellon University Center for Causal Discovery.
1Causal Inference and the KMSS. Predicting the future with the right model Stijn Meganck Vrije Universiteit Brussel Department of Electronics and Informatics.
1 Day 2: Search June 14, 2016 Carnegie Mellon University Center for Causal Discovery.
David Stotts Computer Science Department UNC Chapel Hill.
Simple Linear Regression and Correlation (Continue..,) Reference: Chapter 17 of Statistics for Management and Economics, 7 th Edition, Gerald Keller. 1.
Spanning Trees.
Review: What influences confidence intervals?
Center for Causal Discovery: Summer Short Course/Datathon
Graphing Linear Equations
Presentation transcript:

Exploratory studies: you have empirical data and you want to know what sorts of causal models are consistent with it. Confirmatory tests: you have a causal hypothesis and you want to see if the empirical data agree with it. Do the data agree with my hypothesis? I think that: A B C (hypothesis) My data has this pattern of correlation within it What causal processes could have generated this pattern?

A BC D E “3-D” causal process“2-D” correlational shadow Hypothesis generation B & C independent given A A & D independent given B & C B & D independent given D and so on...

Besides the notion of d-separation, we need one other notion: faithfulness of data to a causal graph Is there one Bighorn Sheep in this picture, or are there two, except that the second is hidden behind the first? Both cases are possible, but the second case requires a very special combination of factors, i.e. that the second animal is positioned so that it gives the illusion of being absent. If the second case happens, then we can say that this is unfaithful to our normal experience.

AB C Overall effect of A on B: (-2*5)=0 Because the two paths exactly cancel out, the overall correlation between A and B is zero; i.e. uncorrelated! The joint probability distribution over A,B & C is unfaithful to the graph because it gives the illusion of of independence between A and B contrary to d-separation. AB C This will only occur when positive and negative values exactly cancel out (very special conditions) like seeing one sheep because the other one is hiding behind the first! Unfaithfulness

Obtaining the undirected dependency graph A B CD E True process which we can’t see! Step 1: create a saturated undirected dependency graph. A B CD E

Obtaining the undirected dependency graph A B CD E True process which we can’t see! Step 2: let the order n (i.e the number) of conditioning variables be zero (i.e. no conditioning variables) - For each unique pair of variables (X,Y) that are still adjacent in the graph…) - For each unique set Q of the n other variables in the graph (in this case none…) Test the data to see if variables X and Y are independent given the conditioning set Q. A B CD E If X and Y are independent in the data, remove the line between them in the graph

Obtaining the undirected dependency graph A B CD E True process which we can’t see! Step 2: let the order n (i.e the number) of conditioning variables be zero (i.e. no conditioning variables) A B CD E Is A & B independent given no others? Is A & C independent given no others? No; don’t remove the line And so on... Result: we don’t remove any lines at this stage.

Obtaining the undirected dependency graph A B CD E True process which we can’t see! Step 3: let the order n (i.e the number) of conditioning variables be one (i.e. one conditioning variable) A B CD E Is A & B independent give C? No. Is A & B independent given D? E? No. Is A & C independent given B? Yes. Therefore, remove the line between A and C and go to next pair (A,D)

Obtaining the undirected dependency graph A B CD E True process which we can’t see! Step 3: let the order n (i.e the number) of conditioning variables be one (i.e. one conditioning variable) A B CD E Is A & B independent give C? No. Is A & B independent given D? E? No. Is A & C independent given B? Yes. Therefore, remove the line between A and C and go to next pair (A,D) Is A & D independent given B? Yes. Therefore, remove the line between A and D and go to next pair (A,E)

Obtaining the undirected dependency graph A B CD E True process which we can’t see! Step 3: let the order n (i.e the number) of conditioning variables be one (i.e. one conditioning variable) A B CD E Is A & B independent give C? No. Is A & B independent given D? E? No. Is A & C independent given B? Yes. Therefore, remove the line between A and C and go to next pair (A,D) Is A & D independent given B? Yes. Therefore, remove the line between A and D and go to next pair (A,E) And so on for each unique pair of variables and each unique conditioning set.

Obtaining the undirected dependency graph A B CD E True process which we can’t see! Step 4: let the order n (i.e the number) of conditioning variables be two (i.e. two conditioning variables) A B CD E Is A & B independent give any two others? No. Therefore, remove the line between B and E and go to next pair (A,D) Is B & C independent given any two others? No. Is B & D independent given any two others? No. Is B & E independent given any two others? Yes (C & D).

Obtaining the undirected dependency graph A B CD E True process which we can’t see! Step 4: let the order n (i.e the number) of conditioning variables be two (i.e. two conditioning variables) A B CD E Is A & B independent give any two others? No. Is A & E independent given any two others? No. Therefore, remove the line between B and E and go to next pair (A,D) Is B & C independent given any two others? No. Is B & D independent given any two others? No. Is B & E independent given any two others? Yes (C & D).

Obtaining the undirected dependency graph This algorithm is provably correct for any probability distribution, and for any functional relationship between variables, and for both cyclic and acyclic causal structures assuming: 1. Faithfulness; 3. No incorrect statistical decisions have been made when deciding upon statistical independence between variables in the data (i.e. lots of data and tests appropriate to the variables in question) The fewer data you have, the greater the chance of missing small, but real, statistical dependencies (statistical power) 2. All data are generated by the same causal process;

Interpreting the undirected dependency graph A B CD E If there is a line between two variables in this undirected dependency graph then: 1. There is a direct causal relationship between the two and/or There is a latent variable that is a common cause of the two and/or There is a more complicated type of undirected path between the two (an inducing path) A B CD A B CD latent

Orienting the undirected dependency graph C D E E Shielded collidersUnshielded collider C D E E E E Unshielded non-colliders XYZ Unshielded pattern A B CD E True process which we can’t see! A B CD E We can’t see this! We’ve learned this!

Orienting the undirected dependency graph In an unshielded collider, C & D will never be independent conditional on E plus every possible combination of remaining variables. C D E In an unshielded non-collider, C & D must be independent conditional on E plus (possibly) some other combination of remaining variables; this is why the line between C & D was removed in the undirected dependency graph! C D E E E C & D dependent given all of Q={ E, E+A, E+B,E+A+B} C & D independent given one of Q= { E, E+A,E+B,E+A+B}

Orienting the undirected dependency graph A B CD E True process which we can’t see! A B CD E We can’t see this! We’ve learned this! A B CA B D C B DB C E B D EC E D A B CD E

Orienting the undirected dependency graph A B CD E True process which we can’t see! We can’t see this! We’ve learned this! A B CD E A B CD E We can’t learn any more by just looking at the data. We can orient the rest of the edges any way we want, so long as we don’t: Partially-oriented acyclic graph - create or destroy any unshielded colliders that are found in the partially-oriented graph - create any cycles in the graph. All such graphs are statistically equivalent and we can’t test between them.

There are some further algorithms that can sometimes allow us to orient more lines, but they are more complicated and require more specialized patterns. There are also algorithms for oriented cyclic causal processes, but these are even more complicated and require stronger assumptions (linearity of relationships and continuous variables). There are also algorithms for detecting latent variables, but these assume both linearity and normality. The TETRAD Project: Causal Models and Statistical Data Causal toolbox: