Data Analysis with Bayesian Networks: A Bootstrap Approach Nir Friedman, Moises Goldszmidt, and Abraham Wyner, UAI99.

Slides:



Advertisements
Similar presentations
ABSTRACT: We examine how to determine the number of states of a hidden variables when learning probabilistic models. This problem is crucial for improving.
Advertisements

ABSTRACT: We examine how to detect hidden variables when learning probabilistic models. This problem is crucial for for improving our understanding of.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
Probabilistic modelling in computational biology Dirk Husmeier Biomathematics & Statistics Scotland.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Graduate Center/City University of New York University of Helsinki FINDING OPTIMAL BAYESIAN NETWORK STRUCTURES WITH CONSTRAINTS LEARNED FROM DATA Xiannian.
. Inferring Subnetworks from Perturbed Expression Profiles D. Pe’er A. Regev G. Elidan N. Friedman.
Learning Module Networks Eran Segal Stanford University Joint work with: Dana Pe’er (Hebrew U.) Daphne Koller (Stanford) Aviv Regev (Harvard) Nir Friedman.
© 1998, Nir Friedman, U.C. Berkeley, and Moises Goldszmidt, SRI International. All rights reserved. Learning I Excerpts from Tutorial at:
Nir Friedman, Iftach Nachman, and Dana Peer Announcer: Kyu-Baek Hwang
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Cs726 Modeling regulatory networks in cells using Bayesian networks Golan Yona Department of Computer Science Cornell University.
PGM: Tirgul 11 Na?ve Bayesian Classifier + Tree Augmented Na?ve Bayes (adapted from tutorial by Nir Friedman and Moises Goldszmidt.
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
Required Sample size for Bayesian network Structure learning
Artificial Intelligence Term Project #3 Kyu-Baek Hwang Biointelligence Lab School of Computer Science and Engineering Seoul National University
Probabilistic Graphical models for molecular networks Sushmita Roy BMI/CS 576 Nov 11 th, 2014.
Cristina Manfredotti D.I.S.Co. Università di Milano - Bicocca An Introduction to the Use of Bayesian Network to Analyze Gene Expression Data Cristina Manfredotti.
Inferring subnetworks from perturbed expression profiles Dana Pe’er, Aviv Regev, Gal Elidan and Nir Friedman Bioinformatics, Vol.17 Suppl
Bayes Net Perspectives on Causation and Causal Inference
Bump Hunting The objective PRIM algorithm Beam search References: Feelders, A.J. (2002). Rule induction by bump hunting. In J. Meij (Ed.), Dealing with.
Bayesian integration of biological prior knowledge into the reconstruction of gene regulatory networks Dirk Husmeier Adriano V. Werhli.
Using Bayesian Networks to Analyze Expression Data N. Friedman, M. Linial, I. Nachman, D. Hebrew University.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Using Bayesian Networks to Analyze Expression Data By Friedman Nir, Linial Michal, Nachman Iftach, Pe'er Dana (2000) Presented by Nikolaos Aravanis Lysimachos.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Michael Baron + * Department of Computer Science, University of Texas at Dallas + Department of Mathematical.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Reverse engineering gene regulatory networks Dirk Husmeier Adriano Werhli Marco Grzegorczyk.
Collective Classification A brief overview and possible connections to -acts classification Vitor R. Carvalho Text Learning Group Meetings, Carnegie.
Using Bayesian Networks to Analyze Whole-Genome Expression Data Nir Friedman Iftach Nachman Dana Pe’er Institute of Computer Science, The Hebrew University.
Non-Informative Dirichlet Score for learning Bayesian networks Maomi Ueno and Masaki Uto University of Electro-Communications, Japan 1.Introduction: Learning.
Learning Bayesian Networks with Local Structure by Nir Friedman and Moises Goldszmidt.
Learning Linear Causal Models Oksana Kohutyuk ComS 673 Spring 2005 Department of Computer Science Iowa State University.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Problem Limited number of experimental replications. Postgenomic data intrinsically noisy. Poor network reconstruction.
Learning Bayesian networks from postgenomic data with an improved structure MCMC sampling scheme Dirk Husmeier Marco Grzegorczyk 1) Biomathematics & Statistics.
Dependency Networks for Collaborative Filtering and Data Visualization UAI-2000 발표 : 황규백.
Bayesian Networks for Regulatory Network Learning BMI/CS 576 Colin Dewey Fall 2015.
Dimensionality Reduction in Unsupervised Learning of Conditional Gaussian Networks Authors: Pegna, J.M., Lozano, J.A., Larragnaga, P., and Inza, I. In.
Reverse engineering of regulatory networks Dirk Husmeier & Adriano Werhli.
Objective Evaluation of Intelligent Medical Systems using a Bayesian Approach to Analysis of ROC Curves Julian Tilbury Peter Van Eetvelt John Curnow Emmanuel.
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, March 8, 2000 Jincheng.
Using Bayesian Networks to Analayze Expression Data
Gaussian Process Networks Nir Friedman and Iftach Nachman UAI-2K.
Project Plan Task 8 and VERSUS2 Installation problems Anatoly Myravyev and Anastasia Bundel, Hydrometcenter of Russia March 2010.
Computational methods for inferring cellular networks II Stat 877 Apr 17 th, 2014 Sushmita Roy.
Bayesian Networks in Bioinformatics Kyu-Baek Hwang Biointelligence Lab School of Computer Science and Engineering Seoul National University
Inferring Regulatory Networks from Gene Expression Data BMI/CS 776 Mark Craven April 2002.
Standard Errors Beside reporting a value of a point estimate we should consider some indication of its precision. For this we usually quote standard error.
A Simple Approach to Ranking Differentially Expressed Gene Expression Time Courses through Gaussian Process Regression By Alfredo A Kalaitzis and Neil.
Bayesian Networks Applied to Modeling Cellular Networks
Bud Mishra Professor of Computer Science and Mathematics 12 ¦ 3 ¦ 2001
Irina Rish IBM T.J.Watson Research Center
Markov Properties of Directed Acyclic Graphs
Test for Mean of a Non-Normal Population – small n
Simulation: Sensitivity, Bootstrap, and Power
Bootstrap - Example Suppose we have an estimator of a parameter and we want to express its accuracy by its standard error but its sampling distribution.
Feature Selection Ioannis Tsamardinos Machine Learning Course, 2006
Integration and Graphical Models
Evaluation of inferred networks
Regulation Analysis using Restricted Boltzmann Machines
An Algorithm for Bayesian Network Construction from Data
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
A Direct Measure for the Efficacy of Bayesian Network Structures Learned from Data Gary F. Holness.
Robust inference of biological Bayesian networks
Presentation transcript:

Data Analysis with Bayesian Networks: A Bootstrap Approach Nir Friedman, Moises Goldszmidt, and Abraham Wyner, UAI99

Abstract Confidence on learned Bayesian networks  Edges How can we believe that the presence of an edge is true?  Markov blankets The Markov blanket of a variable is true?  Order relations The variable Y is ancestor of the variable X? Especially for small datasets, this problem is so crucial. Efron’s Bootstrap approach was used in this paper.

Sparse datasets An application of Bayesian networks to molecular biology Thousands of attributes and at most hundreds of samples How can we separate the measurable “signal” from the “noise”?

Learning Bayesian networks Given data D, find the network structure with high score.  Bde score and MDL score Search space is so large.  Exponential order Greedy hill-climbing with restart can be used.

Partially Directed Acyclic Graphs (PDAGs) The network structure with directed and undirected edges.  The undirected edge allows both directions. X – Y represents both X  Y and Y  X.  In the case that both directions have the same score, we only have to allow both directions in the network. The accurate causal relationship can not be guaranteed by the dataset.

The Confidence Level of Features in the Network Edges, Markov blankets, and order relations  Above quantity can be regarded as the probability of the feature f’s presence in the Bayesian network induced from the samples of size N.

Non-Parametric Bootstrap For i = 1, 2, …, m  Re-sample, with replacement, N instances from D. Denote the resulting dataset by D i.  Apply the learning procedure on D i to induce a network structure For each feature of interest, define

Parametric Bootstrap Induce a network B from D For i = 1, 2, …, m  Sample N instances from B. Denote the resulting dataset by D i.  Apply the learning procedure on D i to induce a network structure For each feature of interest, define

Empirical Evaluation Synthetic datasets from alarm, gene, text networks were used. N = 100, 250, 500, 1000 Bootstrap sampling size was 10 and the number of re- sampling, m was 100.

Results on the Alarm Network

Threshold Setting The appropriate threshold setting is due to the problem domain.  0.8 was best to the alarm network and 0.65 was best to the text network.

Robust features Order relations and Markov blankets were robust to small dataset, but edges were sensitive to the sample size.

The Comparison of Parametric and Non- Parametric Bootstrap

Bootstrap for Network Induction Some constraints according to the threshold values from bootstrapping.

Conclusions The bootstrap estimates are quite cautious. Features induce with high confidence are rarely false positive. The Markov blanket and partial ordering amongst variables are more robust than the existence of edges.