Lecture 5: Causality and Feature Selection Isabelle Guyon

Slides:



Advertisements
Similar presentations
Alexander Statnikov1, Douglas Hardin1,2, Constantin Aliferis1,3
Advertisements

The Simple Linear Regression Model Specification and Estimation Hill et al Chs 3 and 4.
Making Inferences about Causality In general, children who watch violent television programs tend to behave more aggressively toward their peers and siblings.
Making Inferences about Causality In general, children who watch violent television programs tend to behave more aggressively toward their peers and siblings.
Deriving Biological Inferences From Epidemiologic Studies.
Causality Inferences. Objectives: 1. To understand the concept of risk factors and outcome in a scientific way. 2. To understand and comprehend each and.
Research Methods: How We Do Psychology Forming and Testing Hypotheses Theory Integrated set of principles that explain and predict observed events Hypotheses.
Parameter Learning in MN. Outline CRF Learning CRF for 2-d image segmentation IPF parameter sharing revisited.
Can causal models be evaluated? Isabelle Guyon ClopiNet / ChaLearn
Lecture 5: Causality and Feature Selection Isabelle Guyon
BIAS AND CONFOUNDING Nigel Paneth. HYPOTHESIS FORMULATION AND ERRORS IN RESEARCH All analytic studies must begin with a clearly formulated hypothesis.
Chance, bias and confounding
Lecture 4: Embedded methods
Using Markov Blankets for Causal Structure Learning Jean-Philippe Pellet Andre Ellisseeff Presented by Na Dai.
The bumpy road of the search for a (good) cause
Causality Workbenchclopinet.com/causality Results of the Causality Challenge Isabelle Guyon, Clopinet Constantin Aliferis and Alexander Statnikov, Vanderbilt.
The dynamics of iterated learning Tom Griffiths UC Berkeley with Mike Kalish, Steve Lewandowsky, Simon Kirby, and Mike Dowman.
Lecture 3: Introduction to Feature Selection Isabelle Guyon
Causality Workbenchclopinet.com/causality The LOCANET task (Pot-luck challenge, NIPS 2008) Isabelle Guyon, Clopinet Alexander Statnikov, Vanderbilt Univ.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
PGM 2003/04 Tirgul 3-4 The Bayesian Network Representation.
Feature selection methods from correlation to causality Isabelle Guyon NIPS 2008 workshop on kernel learning.
Lecture 2: Introduction to Feature Selection Isabelle Guyon
Review: The Logic Underlying ANOVA The possible pair-wise comparisons: X 11 X 12. X 1n X 21 X 22. X 2n Sample 1Sample 2 means: X 31 X 32. X 3n Sample 3.
Lecture 6: Causal Discovery Isabelle Guyon
Feature selection methods Isabelle Guyon IPAM summer school on Mathematics in Brain Imaging. July 2008.
Feature selection and causal discovery fundamentals and applications Isabelle Guyon
Linear Regression Analysis
Image Analysis and Markov Random Fields (MRFs) Quanren Xiong.
A REVIEW OF FEATURE SELECTION METHODS WITH APPLICATIONS Alan Jović, Karla Brkić, Nikola Bogunović {alan.jovic, karla.brkic,
Feature Selection and Causal discovery Isabelle Guyon, Clopinet André Elisseeff, IBM Zürich Constantin Aliferis, Vanderbilt University.
RESEARCH METHODS.
Addressing Alternative Explanations: Multiple Regression
315 Feature Selection. 316 Goals –What is Feature Selection for classification? –Why feature selection is important? –What is the filter and what is the.
Week 6: Model selection Overview Questions from last week Model selection in multivariable analysis -bivariate significance -interaction and confounding.
Causality challenge #2: Pot-Luck
Research Methods in Psychology Descriptive Methods Naturalistic observation Intensive individual case study Surveys/questionnaires/interviews Correlational.
1 Effective Feature Selection Framework for Cluster Analysis of Microarray Data Gouchol Pok Computer Science Dept. Yanbian University China Keun Ho Ryu.
S TATISTICS Dr. Omar Al Jadaan Assistant Professor – Computer Science & Mathematics.
Case-control study Chihaya Koriyama August 17 (Lecture 1)
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 5.2: Recap on Probability Theory Jürgen Sturm Technische Universität.
Jun-16H.S.1 Confounding and DAGs (Directed Acyclic Graphs) Hein Stigum.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
273 Discovery of Causal Structure Using Causal Probabilistic Network Induction AMIA 2003, Machine Learning Tutorial Constantin F. Aliferis & Ioannis Tsamardinos.
Scientific Inquiry SCI Correlation Vs. Causation.
NIPS 2001 Workshop on Feature/Variable Selection Isabelle Guyon BIOwulf Technologies.
Lost but not forgotten : attrition in the Étude longitudinale du développement des enfants du Québec (ÉLDEQ), Julien BÉRARD-CHAGNON and Simona.
Challenges in causality: Results of the WCCI 2008 challenge Isabelle Guyon, Clopinet Constantin Aliferis and Alexander Statnikov, Vanderbilt Univ. André.
276 Causal Discovery Methods Using Causal Probabilistic Networks MEDINFO 2004, T02: Machine Learning Methods for Decision Support and Discovery Constantin.
Aim: What factors must we consider to make an experimental design?
Chapter Two Methods in the Study of Personality. Gathering Information About Personality Informal Sources of Information: Observations of Self—Introspection,
MACHINE LEARNING 3. Supervised Learning. Learning a Class from Examples Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Making Inferences about Causality In general, children who watch violent television programs tend to behave more aggressively toward their peers and siblings.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
Psychology Research Methods. Characteristics of Good Psychological Research © 2002 John Wiley & Sons, Inc.
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
Using Directed Acyclic Graphs (DAGs) to assess confounding Glenys Webster & Anne Harris May 14, 2007 St Paul’s Hospital Statistical “Rounds”
Inferring Regulatory Networks from Gene Expression Data BMI/CS 776 Mark Craven April 2002.
Proving Causation Why do you think it was me?!.
Lecture 3: Causality and Feature Selection
Lecture 3: Introduction to confounding (part 1)
High level GWAS analysis
Feature Selection Ioannis Tsamardinos Machine Learning Course, 2006
Areas of Research … Causal Discovery Application Integration
Improving Overlap Farrokh Alemi, Ph.D.
Quantitative Research
Searching for Graphical Causal Models of Education Data
Types of Research Questions
Quantitative Research Methods
Presentation transcript:

Lecture 5: Causality and Feature Selection Isabelle Guyon

Variable/feature selection Remove features X i to improve (or least degrade) prediction of Y. X Y

What can go wrong? Guyon-Aliferis-Elisseeff, 2007

What can go wrong?

Guyon-Aliferis-Elisseeff, 2007 X2X2 Y X1X1 Y X1X1 X2X2

X Y Causal feature selection Uncover causal relationships between X i and Y. Y

Lung cancer Causal feature relevance

Lung cancer Causal feature relevance

Lung cancer Causal feature relevance

Lung cancer Markov Blanket Strongly relevant features (Kohavi-John, 1997)  Markov Blanket (Tsamardinos-Aliferis, 2003)

Feature relevance Surely irrelevant feature X i : P(X i, Y |S \i ) = P(X i |S \i )P(Y |S \i ) for all S \i  X \i and all assignment of values to S \i Strongly relevant feature X i : P(X i, Y |X \i )  P(X i |X \i )P(Y |X \i ) for some assignment of values to X \i Weakly relevant feature X i : P(X i, Y |S \i )  P(X i |S \i )P(Y |S \i ) for some assignment of values to S \i  X \i

Lung cancer Markov Blanket Strongly relevant features (Kohavi-John, 1997)  Markov Blanket (Tsamardinos-Aliferis, 2003)

Lung cancer Strongly relevant features (Kohavi-John, 1997)  Markov Blanket (Tsamardinos-Aliferis, 2003) PARENTS Markov Blanket

Lung cancer Strongly relevant features (Kohavi-John, 1997)  Markov Blanket (Tsamardinos-Aliferis, 2003) CHILDREN Markov Blanket

Lung cancer Strongly relevant features (Kohavi-John, 1997)  Markov Blanket (Tsamardinos-Aliferis, 2003) SPOUSES Markov Blanket

Causal relevance Surely irrelevant feature X i : P(X i, Y |S \i ) = P(X i |S \i )P(Y |S \i ) for all S \i  X \i and all assignment of values to S \i Causally relevant feature X i : P(X i,Y| do( S \i ) )  P(X i | do( S \i ) )P(Y| do( S \i ) ) for some assignment of values to S \i Weak/strong causal relevance: –Weak=ancestors, indirect causes –Strong=parents, direct causes.

Lung cancer Examples

Smoking Lung cancer Immediate causes (parents) Genetic factor1

Smoking Anxiety Lung cancer Non-immediate causes (other ancestors)

Genetic factor1 Other cancers Lung cancer Non causes (e.g. siblings)

Y X || Y | C FORK CHAIN X X C C

Smoking Anxiety Lung cancer Hidden more direct cause Tar in lungs

Smoking Lung cancer Confounder Genetic factor2

Coughing Metastasis Lung cancer Biomarker1 Immediate consequences (children)

Lung cancer Strongly relevant features (Kohavi-John, 1997)  Markov Blanket (Tsamardinos-Aliferis, 2003) X C X C X C X || Y but X || Y | C

Lung cancer Bio- marker2 Biomarker1 Non relevant spouse (artifact)

Lung cancer Bio- marker2 Biomarker1 Another case of confounder Systematic noise

Coughing Allergy Lung cancer Truly relevant spouse

Hormonal factor Metastasis Lung cancer Sampling bias

Coughing Allergy Smoking Anxiety Genetic factor1 Hormonal factor Metastasis (b) Other cancers Lung cancer Genetic factor2 Tar in lungs Bio- marker2 Biomarker1 Systematic noise Causal feature relevance

Conclusion Feature selection focuses on uncovering subsets of variables X 1, X 2, … predictive of the target Y. Multivariate feature selection is in principle more powerful than univariate feature selection, but not always in practice. Taking a closer look at the type of dependencies in terms of causal relationships may help refining the notion of variable relevance.

1)Feature Extraction, Foundations and Applications I. Guyon et al, Eds. Springer, ) Causal feature selection I. Guyon, C. Aliferis, A. Elisseeff In “Computational Methods of Feature Selection”, Huan Liu and Hiroshi Motoda Eds., Chapman and Hall/CRC Press, Acknowledgements and references