Graphical Models - Inference -

Slides:



Advertisements
Similar presentations
ABSTRACT: We examine how to detect hidden variables when learning probabilistic models. This problem is crucial for for improving our understanding of.
Advertisements

Weight Annealing Data Perturbation for Escaping Local Maxima in Learning Gal Elidan, Matan Ninio, Nir Friedman Hebrew University
Learning with Missing Data
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
Graphical Models - Inference - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP.
. Exact Inference in Bayesian Networks Lecture 9.
CSCE 582 Computation of the Most Probable Explanation in Bayesian Networks using Bucket Elimination -Hareesh Lingareddy University of South Carolina.
Exact Inference in Bayes Nets
An introduction to machine learning and probabilistic graphical models
Learning: Parameter Estimation
From Variable Elimination to Junction Trees
Graphical Models - Learning -
Bayesian Networks - Intro - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP HREKG.
Graphical Models - Modeling - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP.
3/24 Project 3 released; Due in two weeks. Blog Questions You have been given the topology of a bayes network, but haven't yet gotten the conditional.
Graphical Models - Learning - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP.
CPSC 322, Lecture 30Slide 1 Reasoning Under Uncertainty: Variable elimination Computer Science cpsc322, Lecture 30 (Textbook Chpt 6.4) March, 23, 2009.
Graphical Models - Inference - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP.
. PGM: Tirgul 10 Learning Structure I. Benefits of Learning Structure u Efficient learning -- more accurate models with less data l Compare: P(A) and.
Structure Learning in Bayesian Networks
. Bayesian Networks For Genetic Linkage Analysis Lecture #7.
. Bayesian Networks Lecture 9 Edited from Nir Friedman’s slides by Dan Geiger from Nir Friedman’s slides.
Belief Propagation, Junction Trees, and Factor Graphs
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
. Inference I Introduction, Hardness, and Variable Elimination Slides by Nir Friedman.
Bayesian Networks Russell and Norvig: Chapter 14 CMCS424 Fall 2003 based on material from Jean-Claude Latombe, Daphne Koller and Nir Friedman.
PGM 2002/03 Tirgul5 Clique/Junction Tree Inference.
. Learning Bayesian Networks from Data Nir Friedman Daphne Koller Hebrew U. Stanford.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
1 Midterm Exam Mean: 72.7% Max: % Kernel Density Estimation.
Bayesian Networks Russell and Norvig: Chapter 14 CMCS421 Fall 2006.
Kansas State University Department of Computing and Information Sciences KDD Presentations: Fractals and BBN Inference Saturday, Oct. 12, 2001 Haipeng.
Uncertainty Russell and Norvig: Chapter 14, 15 Koller article on BNs CMCS424 Spring 2002 April 23.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Introduction to Probabilistic Graphical Models Eran Segal Weizmann Institute.
Automated Planning and Decision Making Prof. Ronen Brafman Automated Planning and Decision Making 2007 Bayesian networks Variable Elimination Based on.
Generalizing Variable Elimination in Bayesian Networks 서울 시립대학원 전자 전기 컴퓨터 공학과 G 박민규.
COMP 538 Reasoning and Decision under Uncertainty Introduction Readings: Pearl (1998, Chapter 1 Shafer and Pearl, Chapter 1.
UIUC CS 598: Section EA Graphical Models Deepak Ramachandran Fall 2004 (Based on slides by Eyal Amir (which were based on slides by Lise Getoor and Alvaro.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Real-Time AI Wednesday, March 14, 2001 Haipeng Guo.
Notes on Graphical Models Padhraic Smyth Department of Computer Science University of California, Irvine.
1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.
CS498-EA Reasoning in AI Lecture #10 Instructor: Eyal Amir Fall Semester 2009 Some slides in this set were adopted from Eran Segal.
Intro to Junction Tree propagation and adaptations for a Distributed Environment Thor Whalen Metron, Inc.
Describing Data The canonical descriptive strategy is to describe the data in terms of their underlying distribution As usual, we have a p-dimensional.
Inference Algorithms for Bayes Networks
Some Neat Results From Assignment 1. Assignment 1: Negative Examples (Rohit)
Guidance: Assignment 3 Part 1 matlab functions in statistics toolbox  betacdf, betapdf, betarnd, betastat, betafit.
1 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart CSC384: Intro to Artificial Intelligence  Fahiem Bacchus Office.
1 CMSC 671 Fall 2010 Class #18/19 – Wednesday, November 3 / Monday, November 8 Some material borrowed with permission from Lise Getoor.
Bayesian Networks Russell and Norvig: Chapter 14 CS440 Fall 2005.
A quick run-through about Machine Learning. Classification (Supervised)
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk
Rina Dechter Bren school of ICS University of California, Irvine
Introduction to Artificial Intelligence
Exact Inference ..
Bell & Coins Example Coin1 Bell Coin2
Introduction to Artificial Intelligence
Bayesian Networks (Directed Acyclic Graphical Models)
CSCI 5822 Probabilistic Models of Human and Machine Learning
Read R&N Ch Next lecture: Read R&N
Bayesian Networks Read R&N Ch. 13.6,
Introduction to Artificial Intelligence
Professor Marie desJardins,
Exact Inference ..
Markov Random Fields Presented by: Vladan Radosavljevic.
Class #16 – Tuesday, October 26
Class #22/23 – Wednesday, November 12 / Monday, November 17
Clique Tree Algorithm: Computation
Elimination in Chains A B C E D.
Presentation transcript:

Graphical Models - Inference - Mainly based on F. V. Jensen, „Bayesian Networks and Decision Graphs“, Springer-Verlag New York, 2001. Advanced I WS 06/07 PCWP CO HRBP HREKG HRSAT ERRCAUTER HR HISTORY CATECHOL SAO2 EXPCO2 ARTCO2 VENTALV VENTLUNG VENITUBE DISCONNECT MINVOLSET VENTMACH KINKEDTUBE INTUBATION PULMEMBOLUS PAP SHUNT ANAPHYLAXIS MINOVL PVSAT FIO2 PRESS INSUFFANESTH TPR LVFAILURE ERRBLOWOUTPUT STROEVOLUME LVEDVOLUME HYPOVOLEMIA CVP BP Graphical Models - Inference - Variable Elimination Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany

Reminder: Probability theory Basics of Bayesian Networks Outline Introduction Reminder: Probability theory Basics of Bayesian Networks Modeling Bayesian networks Inference Excourse: Markov Networks Learning Bayesian networks Relational Models

Elimination in Chains Forward pass Using definition of probability, we have A B C E D - Inference (Variable Elimination)

Elimination in Chains By chain decomposition, we get A B C E D - Inference (Variable Elimination)

Elimination in Chains Rearranging terms ... A B C E D - Inference (Variable Elimination)

X Elimination in Chains Now we can perform innermost summation A B C E A has been eliminated A B C E D Now we can perform innermost summation - Inference (Variable Elimination)

X X … Elimination in Chains Rearranging and then summing again, we get B has been eliminated A B C E D Rearranging and then summing again, we get - Inference (Variable Elimination) …

Elimination in Chains with Evidence Similar due to Bayes’ Rule We write the query in explicit form A B C E D - Inference (Variable Elimination)

Variable Elimination Write query in the form Iteratively Move all irrelevant terms outside of innermost sum Perform innermost sum, getting a new term Insert the new term into the product Eliminate irrelevant variables Semantic Bayesian Network - Inference (Variable Elimination)

Asia - A More Complex Example Visit to Asia Smoking Lung Cancer Tuberculosis Abnormality in Chest Bronchitis X-Ray Dyspnoea - Inference (Variable Elimination)

We want to compute P(d) Need to eliminate: v,s,x,t,l,a,b Initial factors - Inference (Variable Elimination)

We want to compute P(d) Need to eliminate: v,s,x,t,l,a,b Initial factors Compute: Eliminate: v - Inference (Variable Elimination) but in general the result of eliminating a variable is not necessarily a probability term !

We want to compute P(d) Need to eliminate: s,x,t,l,a,b V S L T A B X D We want to compute P(d) Need to eliminate: s,x,t,l,a,b Initial factors Compute: Eliminate: s - Inference (Variable Elimination) Summing on s results in a factor with two arguments fs(b,l). In general, result of elimination may be a function of several variables

We want to compute P(d) Need to eliminate: x,t,l,a,b V S L T A B X D We want to compute P(d) Need to eliminate: x,t,l,a,b Initial factors - Inference (Variable Elimination) Compute: Eliminate: x Note that fx(a)=1 for all values of a

We want to compute P(d) Need to eliminate: t,l,a,b V S L T A B X D We want to compute P(d) Need to eliminate: t,l,a,b Initial factors - Inference (Variable Elimination) Compute: Eliminate: t

We want to compute P(d) Need to eliminate: l,a,b V S L T A B X D We want to compute P(d) Need to eliminate: l,a,b Initial factors - Inference (Variable Elimination) Compute: Eliminate: l

We want to compute P(d) Need to eliminate: a,b V S L T A B X D We want to compute P(d) Need to eliminate: a,b Initial factors - Inference (Variable Elimination) Compute: Eliminate: a,b

Variable Elimination (VE) We now understand VE as a sequence of rewriting operations Actual computation is done in elimination step Exactly the same procedure applies to Markov networks Computation depends on order of elimination - Inference (Variable Elimination)

Dealing with Evidence How do we deal with evidence? S L T A B X D Dealing with Evidence How do we deal with evidence? Suppose get evidence V = t, S = f, D = t We want to compute P(L, V = t, S = f, D = t) - Inference (Variable Elimination)

Dealing with Evidence We start by writing the factors: V S L T A B X D - Inference (Variable Elimination)

Dealing with Evidence We start by writing the factors: X D Dealing with Evidence We start by writing the factors: Since we know that V = t, we don’t need to eliminate V Instead, we can replace the factors P(V) and P(T|V) with - Inference (Variable Elimination)

Dealing with Evidence We start by writing the factors: X D Dealing with Evidence We start by writing the factors: Since we know that V = t, we don’t need to eliminate V Instead, we can replace the factors P(V) and P(T|V) with This “selects” the appropriate parts of the original factors given the evidence Note that fP(v) is a constant, and thus does not appear in elimination of other variables - Inference (Variable Elimination)

Dealing with Evidence Initial factors, after setting evidence Given evidence V = t, S = f, D = t, compute P(L, V = t, S = f, D = t ) V S L T A B X D Dealing with Evidence Initial factors, after setting evidence - Inference (Variable Elimination)

Dealing with Evidence Initial factors, after setting evidence Given evidence V = t, S = f, D = t, compute P(L, V = t, S = f, D = t ) V S L T A B X D Dealing with Evidence Initial factors, after setting evidence Eliminating x, we get - Inference (Variable Elimination)

Dealing with Evidence Initial factors, after setting evidence Given evidence V = t, S = f, D = t, compute P(L, V = t, S = f, D = t ) V S L T A B X D Dealing with Evidence Initial factors, after setting evidence Eliminating x, we get Eliminating t, we get - Inference (Variable Elimination)

Dealing with Evidence Initial factors, after setting evidence Given evidence V = t, S = f, D = t, compute P(L, V = t, S = f, D = t ) V S L T A B X D Dealing with Evidence Initial factors, after setting evidence Eliminating x, we get Eliminating t, we get Eliminating a, we get - Inference (Variable Elimination)

Dealing with Evidence Initial factors, after setting evidence Given evidence V = t, S = f, D = t, compute P(L, V = t, S = f, D = t ) V S L T A B X D Dealing with Evidence Initial factors, after setting evidence Eliminating x, t, a, we get Finally, when eliminating b, we get - Inference (Variable Elimination)

Dealing with Evidence Initial factors, after setting evidence Given evidence V = t, S = f, D = t, compute P(L, V = t, S = f, D = t ) V S L T A B X D Dealing with Evidence Initial factors, after setting evidence Eliminating x, t, a, we get Finally, when eliminating b, we get - Inference (Variable Elimination)

Reminder: Probability theory Basics of Bayesian Networks Outline Introduction Reminder: Probability theory Basics of Bayesian Networks Modeling Bayesian networks Inference (VE, Junction Tree) Excourse: Markov Networks Learning Bayesian networks Relational Models