Directed - Bayes Nets Undirected - Markov Random Fields Gibbs Random Fields Causal graphs and causality GRAPHICAL MODELS.

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

1 Undirected Graphical Models Graphical Models – Carlos Guestrin Carnegie Mellon University October 29 th, 2008 Readings: K&F: 4.1, 4.2, 4.3, 4.4,
Markov Networks Alan Ritter.
Graphical Models BRML Chapter 4 1. the zoo of graphical models Markov networks Belief networks Chain graphs (Belief and Markov ) Factor graphs =>they.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
Introduction to Markov Random Fields and Graph Cuts Simon Prince
Exact Inference in Bayes Nets
1 Knowledge Engineering for Bayesian Networks. 2 Probability theory for representing uncertainty l Assigns a numerical degree of belief between 0 and.
Undirected Probabilistic Graphical Models (Markov Nets) (Slides from Sam Roweis)
Introduction of Probabilistic Reasoning and Bayesian Networks
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Markov Networks.
Chapter 8-3 Markov Random Fields 1. Topics 1. Introduction 1. Undirected Graphical Models 2. Terminology 2. Conditional Independence 3. Factorization.
Junction Tree Algorithm Brookes Vision Reading Group.
Junction Trees: Motivation Standard algorithms (e.g., variable elimination) are inefficient if the undirected graph underlying the Bayes Net contains cycles.
From Variable Elimination to Junction Trees
CS774. Markov Random Field : Theory and Application Lecture 06 Kyomin Jung KAIST Sep
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Belief Propagation, Junction Trees, and Factor Graphs
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Bayesian Networks Alan Ritter.
CPSC 422, Lecture 18Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Feb, 25, 2015 Slide Sources Raymond J. Mooney University of.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Image Analysis and Markov Random Fields (MRFs) Quanren Xiong.
Computer vision: models, learning and inference
A Brief Introduction to Graphical Models
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Markov Random Fields Probabilistic Models for Images
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.
Slides for “Data Mining” by I. H. Witten and E. Frank.
An Introduction to Variational Methods for Graphical Models
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
1 Mean Field and Variational Methods finishing off Graphical Models – Carlos Guestrin Carnegie Mellon University November 5 th, 2008 Readings: K&F:
Probabilistic Graphical Models seminar 15/16 ( ) Haim Kaplan Tel Aviv University.
Lecture 2: Statistical learning primer for biologists
Indexing Correlated Probabilistic Databases Bhargav Kanagal, Amol Deshpande University of Maryland, College Park, USA SIGMOD Presented.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Inference Algorithms for Bayes Networks
CPSC 422, Lecture 17Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17 Oct, 19, 2015 Slide Sources D. Koller, Stanford CS - Probabilistic.
1 Param. Learning (MLE) Structure Learning The Good Graphical Models – Carlos Guestrin Carnegie Mellon University October 1 st, 2008 Readings: K&F:
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Pattern Recognition and Machine Learning
Today Graphical Models Representing conditional dependence graphically
Dynamic Programming & Hidden Markov Models. Alan Yuille Dept. Statistics UCLA.
Bayesian Belief Propagation for Image Understanding David Rosenberg.
Markov Random Fields in Vision
Ch 6. Markov Random Fields 6.1 ~ 6.3 Adaptive Cooperative Systems, Martin Beckerman, Summarized by H.-W. Lim Biointelligence Laboratory, Seoul National.
Probabilistic Reasoning Inference and Relational Bayesian Networks.
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation Yee W. Teh, David Newman and Max Welling Published on NIPS 2006 Discussion.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Inference in Bayesian Networks
Exact Inference Continued
Prof. Adriana Kovashka University of Pittsburgh April 4, 2017
Markov Networks.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
CSCI 5822 Probabilistic Models of Human and Machine Learning
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Markov Networks.
Markov Random Fields Presented by: Vladan Radosavljevic.
Graduate School of Information Sciences, Tohoku University
Readings: K&F: 5.1, 5.2, 5.3, 5.4, 5.5, 5.6, 5.7 Markov networks, Factor graphs, and an unified view Start approximate inference If we are lucky… Graphical.
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Markov Networks.
Sequential Learning with Dependency Nets
Presentation transcript:

Directed - Bayes Nets Undirected - Markov Random Fields Gibbs Random Fields Causal graphs and causality GRAPHICAL MODELS

Graphical Model Technology B-Course: Server at Helsinki University Bayes Net: Kevin Murphy’s package (EM oriented) Course code directory (MCMC-oriented) Genie (Carnegie- Mellon University) Numerous commercial and academic packages of varying quality

Graphical Model Technology Inference model, directed: The pdf of a variable C is chosen from CPT (Conditional Probability Table) set depending on values of parent variables. Inference model, undirected: The pdf of a variable is chosen from CPT set depending on values of all neighbours... Inference model, GRF The pdf of a variable is the product of a number of ‘energy functions’ involving cliques containing node.

Bayes’ Net (BN) Directed Graph, No cycles. Sample can be generated along edges, each variable can be sampled when its parents are sampled Useful for engineered systems, diagnostic systems… Conditional Probability Table-- CPT

Complex BN application: Situation Awareness R Suzic, thesis 2006

A part of the human fysiology: Bayes net describing

Markov Random Fields Common in Imaging: Distribution of node conditional on the rest of the nodes is the same as distribution of node conditional on neighbors-- MARKOV property Gibbs sampling: sample unknown nodes conditional on neighbors: MCMC with acceptance probability 1!

Segmentation in MR- MRF sampling Clustering (in spectrum)to get prel segmentation. Smoothing with MRF removes ‘pepper and salt’.

From BN to MRF A BN can be changed to an equivalent MRF by MORALIZATION: Find unmarried parents and marry them. The MRF graph can however describe a larger set of pdfs. The opposite way is not possible: Many MRFs have no equivalent BN

From BN to MRF Exact inference in BN is possible by transforming moralized graph to junction tree: Every edge in left graph must live in some node of right tree. Feasible only if node sets are small Moralized graph from BN Junction tree or tree-decomposition

Gibbs Random Field (GRF) Maximal cliques C of G: {i,h,g,e}, {e,f,d}, {e,c},{d,b},{c,b,a}. A GRF is a probability distribution over node values that falls apart into CLIQUE ENERGY FUNCTIONS G:

GRF:s are MRF:s!! (see proof by Cheung in course pack)

Graphical Model Technology Train model using historic/simulation data Where are the edges? Which are the dependencies? Use model: From partial set of variables, infer values of missing variables Flexible, Intuitive -- but Error Prone!

Learning Graphical model from Data, MCMC style Decide on model type (BN, MRF, Chain Graph) If directed, decide ordering of Nodes to prevent comparing equivalent models Find appropriate formula for computing Bayes’ factor in favor of edge present. Run MCMC: in each step propose to delete/add edge, decide acceptance or not of proposal. Trace is sample of posterior graph structure.

Learning Graphical model from Data, EM style Inference of most likely tree model is easy (Chow Liu, 1968): Use Dirichlet Prior dependency test, select largest Bayes’ factor edge which does not create cycle, and include it. This is essentially Kruskal’s algorithm for shortest spanning tree.

Graph learning: M3 vs M4, M3’ vs M4’ Causal Inference: M4’ vs M4’’ (Unreliable!)

Causality Reasoning M4’: A dependent on B, but given C, A and B are independent A  B|C, not A  B M4’’: A independent of B, but given C, A dependent on B not A  B|C, A  B Does this suggest that C causes A and B in M4’, and C is caused by A and B in M4’’?? Can be decided from observational data!

Testing Treatment T --> R Simple model. Is there an edge between T (treatment) and R (recovery)? But what about Gender perspective?

Simpson’s Paradox A more useful model was: S -> T -> R <- S, where S is sex