1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.

Slides:



Advertisements
Similar presentations
1 Bayes nets Computing conditional probability Polytrees Probability Inferences Bayes nets Computing conditional probability Polytrees Probability Inferences.
Advertisements

Exact Inference in Bayes Nets
Identifying Conditional Independencies in Bayes Nets Lecture 4.
For Monday Read chapter 18, sections 1-2 Homework: –Chapter 14, exercise 8 a-d.
For Monday Finish chapter 14 Homework: –Chapter 13, exercises 8, 15.
Probabilistic Reasoning (2)
Artificial Intelligence Chapter 19 Reasoning with Uncertain Information Biointelligence Lab School of Computer Sci. & Eng. Seoul National University.
Pearl’s Belief Propagation Algorithm Exact answers from tree-structured Bayesian networks Heavily based on slides by: Tomas Singliar,
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
M.I. Jaime Alfonso Reyes ´Cortés.  The basic task for any probabilistic inference system is to compute the posterior probability distribution for a set.
Bayesian network inference
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
10/24  Exam on 10/26 (Lei Tang and Will Cushing to proctor)
Inference in Bayesian Nets
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
CS 561, Session 29 1 Belief networks Conditional independence Syntax and semantics Exact inference Approximate inference.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
10/22  Homework 3 returned; solutions posted  Homework 4 socket opened  Project 3 assigned  Mid-term on Wednesday  (Optional) Review session Tuesday.
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
1 Midterm Exam Mean: 72.7% Max: % Kernel Density Estimation.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
Approximate Inference 2: Monte Carlo Markov Chain
Bayesian networks Chapter 14. Outline Syntax Semantics.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
For Wednesday Read Chapter 11, sections 1-2 Program 2 due.
2 Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions Exact inference by enumeration Exact.
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 11 th, 2006 Readings: K&F: 8.1, 8.2, 8.3,
Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Jim Little Uncertainty 2 Nov 3, 2014 Textbook §6.1.3.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 2, 2015.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Inference Algorithms for Bayes Networks
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
1 CMSC 671 Fall 2010 Class #18/19 – Wednesday, November 3 / Monday, November 8 Some material borrowed with permission from Lise Getoor.
Reasoning Under Uncertainty: Independence and Inference CPSC 322 – Uncertainty 5 Textbook §6.3.1 (and for HMMs) March 25, 2011.
Slide 1 Directed Graphical Probabilistic Models: inference William W. Cohen Machine Learning Feb 2008.
Artificial Intelligence Chapter 19 Reasoning with Uncertain Information Biointelligence Lab School of Computer Sci. & Eng. Seoul National University.
Web-Mining Agents Data Mining Prof. Dr. Ralf Möller Universität zu Lübeck Institut für Informationssysteme Karsten Martiny (Übungen)
CS 541: Artificial Intelligence Lecture VII: Inference in Bayesian Networks.
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
CS 541: Artificial Intelligence
Inference in Bayesian Networks
Today.
Reasoning Under Uncertainty: Conditioning, Bayes Rule & Chain Rule
Artificial Intelligence Chapter 19
Bayesian Network Reasoning with Gibbs Sampling
CSCI 5822 Probabilistic Models of Human and Machine Learning
Instructors: Fei Fang (This Lecture) and Dave Touretzky
CAP 5636 – Advanced Artificial Intelligence
Professor Marie desJardins,
Bayesian Statistics and Belief Networks
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11
Class #19 – Tuesday, November 3
CS 188: Artificial Intelligence
Biointelligence Lab School of Computer Sci. & Eng.
Directed Graphical Probabilistic Models: the sequel
CS 188: Artificial Intelligence Fall 2008
Class #16 – Tuesday, October 26
Class #22/23 – Wednesday, November 12 / Monday, November 17
Markov Networks.
Presentation transcript:

1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13

2 Today’s class Conditional independence Bayesian networks –Network structure –Conditional probability tables –Conditional independence –Inference in Bayesian networks

3 Inference in Bayesian Networks Chapter 14 (2/e)

4 Approaches to inference Exact inference –Enumeration –Belief propagation in polytrees –Variable elimination –Clustering / join tree algorithms Approximate inference –Stochastic simulation / sampling methods –Markov chain Monte Carlo methods –Genetic algorithms –Neural networks –Simulated annealing –Mean field theory

5 Exact inference: Reminder P(x i ) = Σ πi P(x i | π i ) P(π i ) Suppose we want P(D=true), and only the value of E is given as true P (d|e) =  Σ ABC P(a, b, c, d, e) =  Σ ABC P(a) P(b|a) P(c|a) P(d|b,c) P(e|c) With simple iteration to compute this expression, there’s going to be a lot of repetition (e.g., P(e|c) has to be recomputed every time we iterate over C=true) a b c d e

6 Efficient inference in polytrees Polytree: at most one undirected path between any two nodes in the network (i.e., no loops) Query variable X Evidence variables E –Causal support comes from “ancestor” variables, E + –Evidential support comes from “descendant” variables, E - P(X|E) = P(X|E +,E - ) = P(E - |X, E + ) P(X|E + ) / P(E - | E + ) =  P(E - |X) P(X|E + ) Causal support Evidential support 

7 Polytree Causal support (from ancestors and their other descendants) Evidential support (from descendants and their other ancestors)

8 Computing causal support P(X|E) = P(X|E +,E-) = P(E-|X, E + ) P(X|E + ) / P(E- | E + ) =  P(E-|X) P(X|E + ) P(X|E + ) = Σ p = π(X) P(X|p,E + ) P(p|E + ) = Σ p = π(X) P(X|p)  i P(p i |E + i ) When you recurse, exclude X, so each node is visited once X is conditionally independent of E + given X’s parents; this is just a CPT lookup term p i is the i th parent; E + i is the set of evidence variables that are “connected” to that parent; recurse to compute this, unless p i is in E + i (then we’re done)

9 Computing evidential support P(X|E) = P(X|E +,E - ) = P(X, E + ) P(X|E + ) / P(E - |E + ) =  P(E - |X) P(X|E + ) Evidential support is trickier than causal support P(E - |X) =  C  Children(X) P(E - C |X) =  C  Children(X) Σ C Σ p=π(C) P(E - C |X, c, p) P(c, p|X) =  C Σ C Σ p P(E - (E - C ) |X, c, p) P(E + (E - C ) |X, c, p) P(c, p|X) =  C Σ C P(E - (E - C ) | c) Σ p P(E + (E - C ) | p) P(c, p|X) product of evidence propagated via each child of X sum out (marginalize) over values of child and its parents separate into causal and evidential support for the child apply independence to reduce terms

10 Computing evidential support II  C Σ C P(E - (E - C ) | c) Σ p P(E + (E - C ) | p) P(c, p|X) =  C Σ C P(E - (E - C ) | c) Σ p P(p|E + (E- C )) P(E + (E - C ))/P(p) P(c|X,p) P(p|X) =   C Σ C P(E - (E - C ) | c) Σ p P(p|E + (E - C )) P(c|X,p) =   C Σ C P(E - (E - C ) | c) Σ p P(c|X,p)  i P(p i |E + (E - C )) apply Bayes’ rule to the first term and the product rule to the 2nd P(p)=P(p|X), pull out normalizing constant (P(Ec(Ee C )) Parents p i are conditionally independent given evidence; compute this term recursively (causal support for the child’s parents) This term is just a CPT lookup Compute this recursively (evidential support for the child)

11 Exact inference with non-polytrees Conditioning: Find the network’s smallest cutset S (a set of nodes whose removal renders the network singly connected) –In this network, S = {A} or {B} or {C} or {D} For each instantiation of S, compute the belief update with the polytree algorithm Combine the results from all instantiations of S Computationally expensive (finding the smallest cutset is in general NP- hard, and the total number of possible instantiations of S is O(2 }S| )) a b c d e

12 Stochastic simulation Suppose you are given values for some subset of the variables, G, and want to infer values for unknown variables, U Randomly generate a very large number of instantiations from the BN –Generate instantiations for all variables – start at root variables and work your way “forward” Only keep those instantiations that are consistent with the values for G Use the frequency of values for U to get estimated probabilities Accuracy of the results depends on the size of the sample (asymptotically approaches exact results)

13 Markov chain Monte Carlo methods So called because –Markov chain – each instance generated in the sample is dependent on the previous instance –Monte Carlo – statistical sampling method Perform a random walk through variable assignment space, collecting statistics as you go –Start with a random instantiation, consistent with evidence variables –At each step, for some nonevidence variable, randomly sample its value, consistent with the other current assignments Given enough samples, MCMC gives an accurate estimate of the true distribution of values

14 Evidential reasoning: Dempster-Shafer