Belief Propagation: An Extremely Rudimentary Discussion

Slides:



Advertisements
Similar presentations
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Advertisements

ETHEM ALPAYDIN © The MIT Press, Lecture Slides for 1 Lecture Notes for E Alpaydın 2010.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
Classification. Introduction A discriminant is a function that separates the examples of different classes. For example – IF (income > Q1 and saving >Q2)
1 Chapter 5 Belief Updating in Bayesian Networks Bayesian Networks and Decision Graphs Finn V. Jensen Qunyuan Zhang Division. of Statistical Genomics,
Graphical Models BRML Chapter 4 1. the zoo of graphical models Markov networks Belief networks Chain graphs (Belief and Markov ) Factor graphs =>they.
Exact Inference in Bayes Nets
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
Dynamic Bayesian Networks (DBNs)
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Hidden Markov Models M. Vijay Venkatesh. Outline Introduction Graphical Model Parameterization Inference Summary.
Junction Trees: Motivation Standard algorithms (e.g., variable elimination) are inefficient if the undirected graph underlying the Bayes Net contains cycles.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
M.I. Jaime Alfonso Reyes ´Cortés.  The basic task for any probabilistic inference system is to compute the posterior probability distribution for a set.
Midterm Review. The Midterm Everything we have talked about so far Stuff from HW I won’t ask you to do as complicated calculations as the HW Don’t need.
Inference in Bayesian Nets
December Marginal and Joint Beliefs in BN1 A Hybrid Algorithm to Compute Marginal and Joint Beliefs in Bayesian Networks and its complexity Mark.
Propagation in Poly Trees Given a Bayesian Network BN = {G, JDP} JDP(a,b,c,d,e) = p(a)*p(b|a)*p(c|e,b)*p(d)*p(e|d) a d b e c.
Belief Propagation, Junction Trees, and Factor Graphs
Exact Inference: Clique Trees
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Aspects of Bayesian Inference and Statistical Disclosure Control in Python Duncan Smith Confidentiality and Privacy Group CCSR University of Manchester.
Computer vision: models, learning and inference
CSC2535 Spring 2013 Lecture 2a: Inference in factor graphs Geoffrey Hinton.
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Belief Propagation. What is Belief Propagation (BP)? BP is a specific instance of a general class of methods that exist for approximate inference in Bayes.
Probability and Measure September 2, Nonparametric Bayesian Fundamental Problem: Estimating Distribution from a collection of Data E. ( X a distribution-valued.
1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.
Lecture: Forensic Evidence and Probability Characteristics of evidence Class characteristics Individual characteristics  features that place the item.
Introduction to Probabilistic Analysis Introduction to Probabilistic Analysis The third phase of the cycle incorporates uncertainty into the analysis.
Belief Propagation and its Generalizations Shane Oldenburger.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Pattern Recognition and Machine Learning
Today Graphical Models Representing conditional dependence graphically
Belief propagation with junction trees Presented by Mark Silberstein and Yaniv Hamo.
Distributed cooperation and coordination using the Max-Sum algorithm
Bayesian Belief Propagation for Image Understanding David Rosenberg.
Qian Liu CSE spring University of Pennsylvania
Inference in Bayesian Networks
Today.
Exam Preparation Class
Graphical Models in Brief
Data Mining Lecture 11.
CAP 5636 – Advanced Artificial Intelligence
Bayesian Models in Machine Learning
CSCI 5822 Probabilistic Models of Human and Machine Learning
Structure and Semantics of BN
Propagation Algorithm in Bayesian Networks
Pattern Recognition and Image Analysis
Lecture: Forensic Evidence and Probability Characteristics of evidence
CS 188: Artificial Intelligence
Inferring Edges by Using Belief Propagation
Class #19 – Tuesday, November 3
Solving Linear Equations
Graduate School of Information Sciences, Tohoku University
Propagation Algorithm in Bayesian Networks
Class #16 – Tuesday, October 26
Structure and Semantics of BN
Clique Tree Algorithm: Computation
Junction Trees 3 Undirected Graphical Models
Multiply and divide Expressions
Measures of Dispersion
Presentation transcript:

Belief Propagation: An Extremely Rudimentary Discussion McLean & Pavel

The Problem When we have a tree structure describing dependencies between variable, we want to update our probability distributions based on evidence Trees are nice to work with, as the probability distribution can be expressed as the product of edge marginals divided by the product of separator node marginals Simple case: if risk of heart disease depends on your father’s risk of heart disease, what can you say about your own risk if you know your grandfather’s risk? Essentially, given a prior distribution on the tree, find the posterior distribution given some observed evidence

The Solution Belief Propagation is an algorithm to incorporate evidence into a tree distribution Non-iterative method: it only requires two passes through the tree to get an updated distribution

The Algorithm First, incorporate evidence: for each observed variable, find one edge that it is a part of, and set all entries in the edge table that do not correspond to the observed value to zero Next, choose some edge as a root Collect evidence in from every direction to the root Normalize the root edge table Distribute evidence out in every direction from the root

Okay, but what do you mean by “collect evidence”? Well, we want to propagate evidence through the system This is fairly simple for singly-linked items: just update marginal based on joint, then update the next joint based on that marginal, and so on: So if we observed x, Txy becomes T*xy, then we get T*y by summing T*xy over all x Then, T*yz = (Tyz)(T*y/Ty) z x y

And if we have multiply linked items? Then it’s slightly (but only slightly) more complicated: Now if we observe x1 and x2, we get T*x1y and T*x2y We then calculate T1y and T2y (the equivalents of T*y from before, but each using only the information from one of the Xs) Now, T*yz=(Tyz)(T1y/Ty)(T2y/Ty) See Pavel’s handout for a complete workthrough using this graph, and a justification of the calculation of T*yz x1 y z x2

But you’ve got two different marginals for Y! That can’t be right! Patience. All will work out in time. After we have finished collecting evidence, we normalize our root table – in this case, the root would be T*yz Now we distribute evidence – this is the same process as collecting evidence, but in the opposite direction Note that now we are only ever having a single “input” edge going to any given node, so we can come up with proper marginals When we’ve finished distributing evidence, we will have a probability distribution over the tree that reflects the incorporated evidence.