CHAPTER 16: Graphical Models

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Markov Networks Alan Ritter.
A Tutorial on Learning with Bayesian Networks
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for 1 Lecture Notes for E Alpaydın 2010.
Graphical Models BRML Chapter 4 1. the zoo of graphical models Markov networks Belief networks Chain graphs (Belief and Markov ) Factor graphs =>they.
An Introduction to Variational Methods for Graphical Models.
Conditional Random Fields - A probabilistic graphical model Stefan Mutter Machine Learning Group Conditional Random Fields - A probabilistic graphical.
Introduction of Probabilistic Reasoning and Bayesian Networks
An introduction to Bayesian networks Stochastic Processes Course Hossein Amirkhani Spring 2011.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Introduction to probability theory and graphical models Translational Neuroimaging Seminar on Bayesian Inference Spring 2013 Jakob Heinzle Translational.
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Today Logistic Regression Decision Trees Redux Graphical Models
Bayesian Networks Alan Ritter.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
A Brief Introduction to Graphical Models
Probabilistic graphical models. Graphical models are a marriage between probability theory and graph theory (Michael Jordan, 1998) A compact representation.
EE462 MLCV Lecture (1.5 hours) Segmentation – Markov Random Fields Tae-Kyun Kim 1.
第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models
1 CS 391L: Machine Learning: Bayesian Learning: Beyond Naïve Bayes Raymond J. Mooney University of Texas at Austin.
Introduction to Bayesian Networks
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 28 of 41 Friday, 22 October.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS.
LAC group, 16/06/2011. So far...  Directed graphical models  Bayesian Networks Useful because both the structure and the parameters provide a natural.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
Slides for “Data Mining” by I. H. Witten and E. Frank.
An Introduction to Variational Methods for Graphical Models
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Lecture 2: Statistical learning primer for biologists
Functional Data Graphical Models Hongxiao Zhu Virginia Tech July 2, 2015 BIRS Workshop 1 (Joint work with Nate Strawn and David B. Dunson)
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Pattern Recognition and Machine Learning
Introduction on Graphic Models
Today Graphical Models Representing conditional dependence graphically
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Bayesian Decision Theory Introduction to Machine Learning (Chap 3), E. Alpaydin.
INTRODUCTION TO Machine Learning 2nd Edition
CS 2750: Machine Learning Review
CS 2750: Machine Learning Directed Graphical Models
CHAPTER 3: Bayesian Decision Theory
INTRODUCTION TO Machine Learning
Artificial Intelligence Chapter 19
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Bayesian Networks: Motivation
CSCI 5822 Probabilistic Models of Human and Machine Learning
Propagation Algorithm in Bayesian Networks
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17
Markov Networks.
Markov Random Fields Presented by: Vladan Radosavljevic.
Class #19 – Tuesday, November 3
Graduate School of Information Sciences, Tohoku University
Class #16 – Tuesday, October 26
A Direct Measure for the Efficacy of Bayesian Network Structures Learned from Data Gary F. Holness.
INTRODUCTION TO Machine Learning
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Chapter 14 February 26, 2004.
Presentation transcript:

CHAPTER 16: Graphical Models

Graphical Models Aka Bayesian networks, probabilistic networks Nodes are hypotheses (random vars) and the probabilities corresponds to our belief in the truth of the hypothesis Arcs are direct influences between hypotheses The structure is represented as a directed acyclic graph (DAG) The parameters are the conditional probabilities in the arcs (Pearl, 1988, 2000; Jensen, 1996; Lauritzen, 1996) Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Conditional Independence X and Y are independent if P(X,Y)=P(X)P(Y) X and Y are conditionally independent given Z if P(X,Y|Z)=P(X|Z)P(Y|Z) or P(X|Y,Z)=P(X|Z) Three canonical cases: Head-to-tail, Tail-to-tail, head-to- head Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Case 1: Head-to-Tail P(X,Y,Z)=P(X)P(Y|X)P(Z|Y) P(W|C)=P(W|R)P(R|C)+P(W|~R)P(~R|C) Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Case 2: Tail-to-Tail P(X,Y,Z)=P(X)P(Y|X)P(Z|X) Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Case 3: Head-to-Head P(X,Y,Z)=P(X)P(Y)P(Z|X,Y) Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Hidden Markov Model as a Graphical Model Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Belief Propagation (Pearl, 1988) Chain: Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Trees Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Polytrees Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Undirected Graphs: Markov Random Fields In a Markov random field, dependencies are symmetric, for example, pixels in an image In an undirected graph, A and B are independent if removing C makes them unconnected. Potential function yc(Xc) shows how favorable is the particular configuration X over the clique C The joint is defined in terms of the clique potentials Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Factor Graphs Define new factor nodes and write the joint in terms of them Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Learning a Graphical Model Learning the conditional probabilities, either as tables (for discrete case with small number of parents), or as parametric functions Learning the structure of the graph: Doing a state-space search over a score function that uses both goodness of fit to data and some measure of complexity Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)