1 Bayesian Networks Chapter 14.1-14.2; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.

Slides:



Advertisements
Similar presentations
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
Advertisements

Exact Inference in Bayes Nets
Identifying Conditional Independencies in Bayes Nets Lecture 4.
Generative Models. Announcements Probability Review (Friday, 1:15 Gates B03) Late days… To be fair… Start the p-set early double late days.
Bayesian Networks Chapter 14 Section 1, 2, 4. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
Probabilistic Reasoning (2)
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
Bayesian Network. Introduction Independence assumptions Seems to be necessary for probabilistic inference to be practical. Naïve Bayes Method Makes independence.
Review: Bayesian learning and inference
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Bayesian Networks. Motivation The conditional independence assumption made by naïve Bayes classifiers may seem to rigid, especially for classification.
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
Bayesian networks Chapter 14 Section 1 – 2.
Bayesian Belief Networks
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Example applications of Bayesian networks
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
CS 561, Session 29 1 Belief networks Conditional independence Syntax and semantics Exact inference Approximate inference.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
1 Midterm Exam Mean: 72.7% Max: % Kernel Density Estimation.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
Bayesian networks Chapter 14. Outline Syntax Semantics.
1 Abduction, Uncertainty, and Probabilistic Reasoning Chapters 13, 14, and more.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
Probabilistic Reasoning ECE457 Applied Artificial Intelligence Spring 2007 Lecture #9.
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
Introduction to Bayesian Networks
Generalizing Variable Elimination in Bayesian Networks 서울 시립대학원 전자 전기 컴퓨터 공학과 G 박민규.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Announcements Project 4: Ghostbusters Homework 7
1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Abduction, Uncertainty, and Probabilistic Reasoning
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Bayesian Networks CSE 473. © D. Weld and D. Fox 2 Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Inference Algorithms for Bayes Networks
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
1 CMSC 671 Fall 2010 Class #18/19 – Wednesday, November 3 / Monday, November 8 Some material borrowed with permission from Lise Getoor.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
CS 541: Artificial Intelligence Lecture VII: Inference in Bayesian Networks.
Artificial Intelligence Bayes’ Nets: Independence Instructors: David Suter and Qince Li Course Harbin Institute of Technology [Many slides.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11 CS479/679 Pattern Recognition Dr. George Bebis.
CS 541: Artificial Intelligence
CS 2750: Machine Learning Directed Graphical Models
Bayesian networks Chapter 14 Section 1 – 2.
Qian Liu CSE spring University of Pennsylvania
Read R&N Ch Next lecture: Read R&N
Bayesian Networks Probability In AI.
Logic for Artificial Intelligence
Markov Networks.
Read R&N Ch Next lecture: Read R&N
Bayesian Network Reasoning with Gibbs Sampling
Instructors: Fei Fang (This Lecture) and Dave Touretzky
CAP 5636 – Advanced Artificial Intelligence
Professor Marie desJardins,
Bayesian Statistics and Belief Networks
Class #19 – Tuesday, November 3
CS 188: Artificial Intelligence
Class #16 – Tuesday, October 26
Class #22/23 – Wednesday, November 12 / Monday, November 17
Bayesian networks Chapter 14 Section 1 – 2.
Read R&N Ch Next lecture: Read R&N
Chapter 14 February 26, 2004.
Presentation transcript:

1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.

2 Outline Bayesian networks –Network structure –Conditional probability tables –Conditional independence Inference in Bayesian networks –Exact inference –Approximate inference

3 Bayesian Belief Networks (BNs) Definition: BN = (DAG, CPD) –DAG: directed acyclic graph (BN’s structure) Nodes: random variables (typically binary or discrete, but methods also exist to handle continuous variables) Arcs: indicate probabilistic dependencies between nodes (lack of link signifies conditional independence) –CPD: conditional probability distribution (BN’s parameters) Conditional probabilities at each node, usually stored as a table (conditional probability table, or CPT) –Root nodes are a special case – no parents, so just use priors in CPD:

4 Example BN a b c d e P(C|A) = 0.2 P(C|  A) = P(B|A) = 0.3 P(B|  A) = P(A) = P(D|B,C) = 0.1 P(D|B,  C) = 0.01 P(D|  B,C) = 0.01 P(D|  B,  C) = P(E|C) = 0.4 P(E|  C) = Note that we only specify P(A) etc., not P(¬A), since they have to add to one

5 Conditional independence assumption – where q is any set of variables (nodes) other than and its successors – blocks influence of other nodes on and its successors (q influences only through variables in ) –With this assumption, the complete joint probability distribution of all variables in the network can be represented by (recovered from) local CPDs by chaining these CPDs: q Conditional independence and chaining

6 Chaining: Example Computing the joint probability for all variables is easy: P(a, b, c, d, e) = P(e | a, b, c, d) P(a, b, c, d) by the product rule =P(e | c) P(a, b, c, d) by cond. indep. assumption = P(e | c) P(d | a, b, c) P(a, b, c) =P(e | c) P(d | b, c) P(c | a, b) P(a, b) =P(e | c) P(d | b, c) P(c | a) P(b | a) P(a) a b c d e

7 Topological semantics A node is conditionally independent of its non- descendants given its parents A node is conditionally independent of all other nodes in the network given its parents, children, and children’s parents (also known as its Markov blanket) The method called d-separation can be applied to decide whether a set of nodes X is independent of another set Y, given a third set Z

8 Inference tasks Simple queries: Computer posterior marginal P(X i | E=e) –E.g., P(NoGas | Gauge=empty, Lights=on, Starts=false) Conjunctive queries: –P(X i, X j | E=e) = P(X i | e=e) P(X j | X i, E=e) Optimal decisions: Decision networks include utility information; probabilistic inference is required to find P(outcome | action, evidence) Value of information: Which evidence should we seek next? Sensitivity analysis: Which probability values are most critical? Explanation: Why do I need a new starter motor?

9 Approaches to inference Exact inference –Enumeration –Belief propagation in polytrees –Variable elimination –Clustering / join tree algorithms Approximate inference –Stochastic simulation / sampling methods –Markov chain Monte Carlo methods –Genetic algorithms –Neural networks –Simulated annealing –Mean field theory

10 Direct inference with BNs Instead of computing the joint, suppose we just want the probability for one variable Exact methods of computation: –Enumeration –Variable elimination Join trees: get the probabilities associated with every query variable

11 Inference by enumeration Add all of the terms (atomic event probabilities) from the full joint distribution If E are the evidence (observed) variables and Y are the other (unobserved) variables, then: P(X|e) = α P(X, E) = α ∑ P(X, E, Y) Each P(X, E, Y) term can be computed using the chain rule Computationally expensive!

12 Example: Enumeration P(x i ) = Σ πi P(x i | π i ) P(π i ) Suppose we want P(D=true), and only the value of E is given as true P (d|e) =  Σ ABC P(a, b, c, d, e) =  Σ ABC P(a) P(b|a) P(c|a) P(d|b,c) P(e|c) With simple iteration to compute this expression, there’s going to be a lot of repetition (e.g., P(e|c) has to be recomputed every time we iterate over C=true) a b c d e

13 Exercise: Enumeration smartstudy preparedfair pass p(smart)=.8p(study)=.6 p(fair)=.9 p(prep|…)smart  smart study.9.7  study.5.1 p(pass|…) smart  smart prep  prep prep  prep fair  fair.1 Query: What is the probability that a student studied, given that they pass the exam?

14 Summary Bayes nets –Structure –Parameters –Conditional independence –Chaining BN inference –Enumeration –Variable elimination –Sampling methods