CSE 415 -- (c) S. Tanimoto, 2007 Bayes Nets 1 Bayes Networks Outline: Why Bayes Nets? Review of Bayes’ Rule Combining independent items of evidence General.

Slides:



Advertisements
Similar presentations
Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
Advertisements

Identifying Conditional Independencies in Bayes Nets Lecture 4.
For Monday Read chapter 18, sections 1-2 Homework: –Chapter 14, exercise 8 a-d.
Bayesian Networks CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
For Monday Finish chapter 14 Homework: –Chapter 13, exercises 8, 15.
1 22c:145 Artificial Intelligence Bayesian Networks Reading: Ch 14. Russell & Norvig.
5/17/20151 Probabilistic Reasoning CIS 479/579 Bruce R. Maxim UM-Dearborn.
CS 484 – Artificial Intelligence1 Announcements Homework 8 due today, November 13 ½ to 1 page description of final project due Thursday, November 15 Current.
Toothache  toothache catch  catch catch  catch cavity  cavity Joint PDF.
Probability Concepts Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Bayesian Belief Networks
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Naïve Bayes Model. Outline Independence and Conditional Independence Naïve Bayes Model Application: Spam Detection.
CSE (c) S. Tanimoto, 2008 Bayes Nets 1 Probabilistic Reasoning With Bayes’ Rule Outline: Motivation Generalizing Modus Ponens Bayes’ Rule Applying.
P Fig. 6-1, p. 193 Fig. 6-2, p. 193 Fig. 6-3, p. 195.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
CSc411Artificial Intelligence1 Chapter 5 STOCHASTIC METHODS Contents The Elements of Counting Elements of Probability Theory Applications of the Stochastic.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Does Naïve Bayes always work?
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Artificial Intelligence CS 165A Tuesday, November 27, 2007  Probabilistic Reasoning (Ch 14)
Lab Assignment 1 Environments Search Bayes Nets. Problem 1: Peg Solitaire Is Peg Solitaire: Partially observable? Stochastic? Continuous? Adversarial?
Bayes for Beginners Presenters: Shuman ji & Nick Todd.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
For Wednesday Read Chapter 11, sections 1-2 Program 2 due.
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
1 Reasoning Under Uncertainty Artificial Intelligence Chapter 9.
CSE PR 1 Reasoning - Rule-based and Probabilistic Representing relations with predicate logic Limitations of predicate logic Representing relations.
Generalizing Variable Elimination in Bayesian Networks 서울 시립대학원 전자 전기 컴퓨터 공학과 G 박민규.
AP STATISTICS LESSON 6.3 (DAY 1) GENERAL PROBABILITY RULES.
1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
12/7/20151 Math b Conditional Probability, Independency, Bayes Theorem.
Education as a Signaling Device and Investment in Human Capital Topic 3 Part I.
Today’s Topics Graded HW1 in Moodle (Testbeds used for grading are linked to class home page) HW2 due (but can still use 5 late days) at 11:55pm tonight.
Fall 2002Biostat Probability Probability - meaning 1) classical 2) frequentist 3) subjective (personal) Sample space, events Mutually exclusive,
Inference Algorithms for Bayes Networks
1 Bayesian Networks: A Tutorial. 2 Introduction Suppose you are trying to determine if a patient has tuberculosis. You observe the following symptoms:
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
STATISTICS 6.0 Conditional Probabilities “Conditional Probabilities”
Probability Theory. Topics Basic Probability Concepts: Sample Spaces and Events, Simple Probability, and Joint Probability, Conditional Probability Bayes’
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Chapter 4 Probability.
Weng-Keen Wong, Oregon State University © Bayesian Networks: A Tutorial Weng-Keen Wong School of Electrical Engineering and Computer Science Oregon.
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Chapter 2: Probability CIS Computational Probability and Statistics.
Recuperação de Informação B Modern Information Retrieval Cap. 2: Modeling Section 2.8 : Alternative Probabilistic Models September 20, 1999.
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Chapter 2: Probability CIS Computational Probability and.
CS 2750: Machine Learning Directed Graphical Models
Does Naïve Bayes always work?
Qian Liu CSE spring University of Pennsylvania
Inference in Bayesian Networks
Chapter 4 Probability.
Read R&N Ch Next lecture: Read R&N
Bayesian Networks: A Tutorial
Representing Uncertainty
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Chapter 2: Probability CIS Computational Probability and.
CAP 5636 – Advanced Artificial Intelligence
Bayesian Statistics and Belief Networks
Class #19 – Tuesday, November 3
CS 188: Artificial Intelligence
Class #16 – Tuesday, October 26
Probabilistic Reasoning With Bayes’ Rule
General Probability Rules
Probabilistic Reasoning With Bayes’ Rule
Probabilistic Reasoning With Bayes’ Rule
Chapter 5 – Probability Rules
Presentation transcript:

CSE (c) S. Tanimoto, 2007 Bayes Nets 1 Bayes Networks Outline: Why Bayes Nets? Review of Bayes’ Rule Combining independent items of evidence General combination of evidence Benefits of Bayes nets for expert systems

CSE (c) S. Tanimoto, 2007 Bayes Nets 2 Why Bayes Networks? Reasoning about events involving many parts or contingencies generally requires that a joint probability distribution be known. Such a distribution might require thousands of parameters. Modeling at this level of detail is typically not practical. Bayes Nets require making assumptions about the relevance of some conditions to others. Once the assumptions are made, the joint distribution can be “factored” so that there are many fewer separate parameters that must be specified.

CSE (c) S. Tanimoto, 2007 Bayes Nets 3 Review of Bayes’ Rule E: Some evidence exists, i.e., a particular condition is true H: some hypothesis is true. P(E|H) = probability of E given H. P(E|~H) = probability of E given not H. P(H) = probability of H, independent of E. P(E|H) P(H) P(H|E) = P(E) P(E) = P(E|H) P(H) + P(E|~H)(1 - P(H))

CSE (c) S. Tanimoto, 2007 Bayes Nets 4 Combining Independent Items of Evidence E 1 : The patient’s white blood cell count exceeds 110% of average. E 2 : The patient’s body temperature is above 101 o F. H: The patient is infected with tetanus. O(H) = 0.01/0.99 O(H|E 1 ) = λ 1 O(H) sufficiency factor for high white cell count. O(H|E 2 ) = λ 2 O(H) sufficiency factor for high body temp. Assuming E1 and E2 are independent: O(H|E 1  E 2 ) = λ 1 λ 2 O(H)

CSE (c) S. Tanimoto, 2007 Bayes Nets 5 Bayes Net Example A: Accident (An accident blocked traffic on the highway.) B: Barb Late (Barbara is late for work). C: Chris Late (Christopher is late for work). BC A P(A) = 0.2 P(B|A) = 0.5 P(B|~A) = 0.15 P(C|A) = 0.3 P(C|~A) = 0.1

CSE (c) S. Tanimoto, 2007 Bayes Nets 6 Forward Propagation (from causes to effects) BC A P(A) = 0.2 P(B|A) = 0.5 P(B|~A) = 0.15 P(C|A) = 0.3 P(C|~A) = 0.1 Suppose A (there is an accident): Then P(B|A) = 0.5 P(C|A) = 0.3 Suppose ~A (no accident): Then P(B|~A) = 0.15 P(C|A) = 0.1 (These come directly from the given information.)

CSE (c) S. Tanimoto, 2007 Bayes Nets 7 Marginal Probabilities (using forward propagation) BC A P(A) = 0.2 P(B|A) = 0.5 P(B|~A) = 0.15 P(C|A) = 0.3 P(C|~A) = 0.1 Then P(B) = probability Barb is late in any situation = P(B|A) P(A) + P(B|~A) P(~A) = (0.5)(0.2) + (0.15)(0.8) = 0.22 Similarly P(C) = probability Chris is late in any situation = P(C|A) P(A) + P(C|~A) P(~A) = (0.3)(0.2) + (0.1)(0.8) = 0.14 Marginalizing means eliminating a contingency by summing the probabilities for its different cases (here A and ~A).

CSE (c) S. Tanimoto, 2007 Bayes Nets 8 Backward Propagation: “diagnosis” (from effects to causes) BC A P(A) = 0.2 P(B|A) = 0.5 P(B|~A) = 0.15 P(C|A) = 0.3 P(C|~A) = 0.1 Suppose B (Barb is late) What’s the probability of an accident on the highway? Use Bayes’ rule: Then P(A|B) = P(B|A) P(A) / P(B) = 0.5 * 0.2 / (0.5 * * 0.8) = 0.1 / 0.22 =

CSE (c) S. Tanimoto, 2007 Bayes Nets 9 Revising Probabilities of Consequences BC A P(A|B) = P(B|A) = 0.5 P(B|~A) = 0.15 P(C|A) = 0.3 P(C|~A) = 0.1 P(C|B) = ??? Suppose B (Barb is late). What’s the probability that Chris is also late, given this information? We already figured that P(A|B) = P(C|B) = P(C|A) P(A|B) + P(C|~A) P(~A|B) = (0.3)(0.4545) + (0.1)(0.5455) = somewhat higher than P(C)=0.14

CSE (c) S. Tanimoto, 2007 Bayes Nets 10 Handling Multiple Causes BC A P(B|A^D) = 0.9 P(B|A^~D) = 0.45 P(B|~A^D) = 0.75 P(B|~A^~D) = 0.1 D: Disease (Barb has the flu). P(D) = (These values are consistent with P(B|A) = 0.5. ) D

CSE (c) S. Tanimoto, 2007 Bayes Nets 11 Explaining Away BC A P(B|A^D) = 0.9 P(B|A^~D) = 0.45 P(B|~A^D) = 0.75 P(B|~A^~D) = 0.1 Suppose B (Barb is late). This raises the probability for each cause: P(A|B) = , P(D|B) = P(B|D) P(D)/ P(B) = Now, in addition, suppose C (Chris is late). C makes it more likely that A is true, “And this explains B.” D is now a less probable. P(B|D) = P(B|A^D)P(A) + P(B|~A^D)P(~A) = 0.78 D

CSE (c) S. Tanimoto, 2007 Bayes Nets 12 Benefits of Bayes Nets The joint probability distribution normally requires 2 n – 1 independent parameters. With Bayes Nets we only specify these parameters: 1.“root” node probabilities. e. g., P(A=true) = 0.2; P(A=false)= For each non-root node, a table of 2 k values, where k is the number of parents of that node. Typically k < Propagating probabilities happens along the paths in the net. With a full joint prob. dist., many more computations may be needed.