Reasoning with Bayesian Belief Networks

Slides:



Advertisements
Similar presentations
Making Simple Decisions Chapter 16 Some material borrowed from Jean-Claude Latombe and Daphne Koller by way of Marie desJadines,
Advertisements

Making Simple Decisions
Artificial Intelligence Universitatea Politehnica Bucuresti Adina Magda Florea
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Probability: Review The state of the world is described using random variables Probabilities are defined over events –Sets of world states characterized.
PROBABILITY. Uncertainty  Let action A t = leave for airport t minutes before flight from Logan Airport  Will A t get me there on time ? Problems :
1 Knowledge Engineering for Bayesian Networks. 2 Probability theory for representing uncertainty l Assigns a numerical degree of belief between 0 and.
Bayesian Classification
IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
Probabilities Random Number Generators –Actually pseudo-random –Seed Same sequence from same seed Often time is used. Many examples on web. Custom random.
Bayesian Networks. Motivation The conditional independence assumption made by naïve Bayes classifiers may seem to rigid, especially for classification.
Bayesian Belief Networks
Proactivity using Bayesian Methods and Learning 3rd UK-UbiNet Workshop, Bath Lukas Sklenar Computing Laboratory, University of Kent.
KI2 - 2 Kunstmatige Intelligentie / RuG Probabilities Revisited AIMA, Chapter 13.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Representing Uncertainty CSE 473. © Daniel S. Weld 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one.
Bayesian networks practice. Semantics e.g., P(j  m  a   b   e) = P(j | a) P(m | a) P(a |  b,  e) P(  b) P(  e) = … Suppose we have the variables.
Uncertainty Logical approach problem: we do not always know complete truth about the environment Example: Leave(t) = leave for airport t minutes before.
Bayesian Networks Alan Ritter.
Bayesian Reasoning. Tax Data – Naive Bayes Classify: (_, No, Married, 95K, ?)
Judgment and Decision Making in Information Systems Computing with Influence Diagrams and the PathFinder Project Yuval Shahar, M.D., Ph.D.
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Adapted from slides by Tim Finin
Reasoning with Bayesian Networks. Overview Bayesian Belief Networks (BBNs) can reason with networks of propositions and associated probabilities Useful.
Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models.
Bayesian networks Chapter 14 Section 1 – 2. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Reasoning with Bayesian Belief Networks. Overview Bayesian Belief Networks (BBNs) can reason with networks of propositions and associated probabilities.
Introduction to Bayesian Networks
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Jim Little Uncertainty 2 Nov 3, 2014 Textbook §6.1.3.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
Computing & Information Sciences Kansas State University Data Sciences Summer Institute Multimodal Information Access and Synthesis Learning and Reasoning.
Chapter 6 Bayesian Learning
Uncertainty Chapter 13. Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes' Rule.
Uncertainty Let action A t = leave for airport t minutes before flight Will A t get me there on time? Problems: 1.partial observability (road state, other.
Bayesian Learning Bayes Theorem MAP, ML hypotheses MAP learners
Introduction on Graphic Models
Conditional Probability, Bayes’ Theorem, and Belief Networks CISC 2315 Discrete Structures Spring2010 Professor William G. Tanner, Jr.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
Making Simple Decisions Chapter 16 Some material borrowed from Jean-Claude Latombe and Daphne Koller by way of Marie desJadines,
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Bayesian Decision Theory Introduction to Machine Learning (Chap 3), E. Alpaydin.
INTRODUCTION TO Machine Learning 2nd Edition
Bayesian networks Chapter 14 Section 1 – 2.
Presented By S.Yamuna AP/CSE
Qian Liu CSE spring University of Pennsylvania
Making Simple Decisions
Read R&N Ch Next lecture: Read R&N
Uncertainty Chapter 13.
Reasoning Under Uncertainty: Conditioning, Bayes Rule & Chain Rule
Conditional Probability, Bayes’ Theorem, and Belief Networks
Learning Bayesian Network Models from Data
INTRODUCTION TO Machine Learning
Read R&N Ch Next lecture: Read R&N
Representing Uncertainty
Professor Marie desJardins,
CS 188: Artificial Intelligence Fall 2008
CS 188: Artificial Intelligence Fall 2007
Class #21 – Monday, November 10
Class #16 – Tuesday, October 26
Uncertainty Logical approach problem: we do not always know complete truth about the environment Example: Leave(t) = leave for airport t minutes before.
Making Simple Decisions
Bayesian networks Chapter 14 Section 1 – 2.
Probabilistic Reasoning
Read R&N Ch Next lecture: Read R&N
Intro. to Data Mining Chapter 6. Bayesian.
Presentation transcript:

Reasoning with Bayesian Belief Networks

Overview Bayesian Belief Networks (BBNs) can reason with networks of propositions and associated probabilities Useful for many AI problems Diagnosis Expert systems Planning Learning

BBN Definition AKA Bayesian Network, Bayes Net A graphical model (as a DAG) of probabilistic relationships among a set of random variables Links represent direct influence of one variable on another source

Recall Bayes Rule Note the symmetry: we can compute the probability of a hypothesis given its evidence and vice versa

Simple Bayesian Network Smoking Cancer P( S=no) 0.80 S=light) 0.15 S=heavy) 0.05 Smoking= no light heavy P( C=none) 0.96 0.88 0.60 C=benign) 0.03 0.08 0.25 C=malig) 0.01 0.04 0.15

More Complex Bayesian Network Age Gender Exposure to Toxics Smoking Cancer Serum Calcium Lung Tumor

More Complex Bayesian Network Nodes represent variables Age Gender Exposure to Toxics Smoking Links represent “causal” relations Cancer Does gender cause smoking? Influence might be a more appropriate term Serum Calcium Lung Tumor

More Complex Bayesian Network predispositions Age Gender Exposure to Toxics Smoking Cancer Serum Calcium Lung Tumor

More Complex Bayesian Network Age Gender Exposure to Toxics Smoking condition Cancer Serum Calcium Lung Tumor

More Complex Bayesian Network Age Gender Exposure to Toxics Smoking Cancer observable symptoms Serum Calcium Lung Tumor

Independence Age and Gender are independent. P(A,G) = P(G) * P(A) P(A |G) = P(A) P(G |A) = P(G) P(A,G) = P(G|A) P(A) = P(G)P(A) P(A,G) = P(A|G) P(G) = P(A)P(G)

Conditional Independence Cancer is independent of Age and Gender given Smoking Age Gender Smoking P(C | A,G,S) = P(C | S) Cancer

Conditional Independence: Naïve Bayes Serum Calcium and Lung Tumor are dependent Cancer Serum Calcium is indepen-dent of Lung Tumor, given Cancer P(L | SC,C) = P(L|C) P(SC | L,C) = P(SC|C) Serum calcium is a blood test to measure the amount of calcium in the blood Elevated levels of calcium in the blood is associated with some cancers. Serum Calcium Lung Tumor Naïve Bayes assumption: evidence (e.g., symptoms) is indepen-dent given the disease. This make it easy to combine evidence

Explaining Away Exposure to Toxics and Smoking are independent Exposure to Toxics Smoking Exposure to Toxics is dependent on Smoking, given Cancer Cancer Occam’s Razor : among competing hypotheses, the one with the fewest assumptions should be selected P(E=heavy | C=malignant) > P(E=heavy | C=malignant, S=heavy) Explaining away: reasoning pattern where confirma-tion of one causereduces need to invoke alternatives Essence of Occam’s Razor (prefer hypothesis with fewest assumptions) Relies on independence of causes

Conditional Independence A variable (node) is conditionally independent of its non-descendants given its parents Age Gender Non-Descendants Exposure to Toxics Smoking Parents Cancer is independent of Age and Gender given Exposure to Toxics and Smoking. Cancer Serum Calcium Lung Tumor Descendants

Another non-descendant A variable is conditionally independent of its non-descendants given its parents Age Gender Exposure to Toxics Smoking Diet Cancer Cancer is independent of Diet given Exposure to Toxics and Smoking Serum Calcium Lung Tumor

BBN Construction The knowledge acquisition process for a BBN involves three steps KA1: Choosing appropriate variables KA2: Deciding on the network structure KA3: Obtaining data for the conditional probability tables

KA1: Choosing variables Variable values can be integers, reals or enumerations Variable should have collectively exhaustive, mutually exclusive values Error Occurred No Error They should be values, not probabilities Risk of Smoking Smoking

Heuristic: Knowable in Principle Example of good variables Weather: {Sunny, Cloudy, Rain, Snow} Gasoline: Cents per gallon {0,1,2…} Temperature: {  100°F , < 100°F} User needs help on Excel Charting: {Yes, No} User’s personality: {dominant, submissive}

KA2: Structuring Network structure corresponding Gender Age Network structure corresponding to “causality” is usually good. Smoking Exposure to Toxic Cancer Genetic Damage Initially this uses the designer’s knowledge but can be checked with data Lung Tumor

KA3: The Numbers For each variable we have a table of probability of its value for values of its parents For variables w/o parents, we have prior probabilities Cancer Smoking smoking cancer no light heavy none 0.96 0.88 0.60 benign 0.03 0.08 0.25 malignant 0.01 0.04 0.15 smoking priors no 0.80 light 0.15 heavy 0.05

KA3: The numbers Second decimal usually doesn’t matter Relative probabilities are important Zeros and ones are often enough Order of magnitude is typical: 10-9 vs 10-6 Sensitivity analysis can be used to decide accuracy needed

Three kinds of reasoning BBNs support three main kinds of reasoning: Predicting conditions given predispositions Diagnosing conditions given symptoms (and predisposing) Explaining a condition by one or more predispositions To which we can add a fourth: Deciding on an action based on probabilities of the conditions

Predictive Inference How likely are elderly males Age Gender How likely are elderly males to get malignant cancer? Exposure to Toxics Smoking Cancer P(C=malignant | Age>60, Gender=male) Serum Calcium Lung Tumor

Predictive and diagnostic combined Age Gender How likely is an elderly male patient with high Serum Calcium to have malignant cancer? Exposure to Toxics Smoking Cancer P(C=malignant | Age>60, Gender= male, Serum Calcium = high) Serum Calcium Lung Tumor

Explaining away Age Gender If we see a lung tumor, the probability of heavy smoking and of exposure to toxics both go up Exposure to Toxics Smoking If we then observe heavy smoking, the probability of exposure to toxics goes back down Smoking Cancer Serum Calcium Lung Tumor

Decision making A decision is a medical domain might be a choice of treatment (e.g., radiation or chemotherapy) Decisions should be made to maximize expected utility View decision making in terms of Beliefs/Uncertainties Alternatives/Decisions Objectives/Utilities

A Decision Problem Should I have my party inside or outside? Regret Relieved Perfect! Disaster dry wet

Value Function A numerical score over all possible states of the world allows BBN to be used to make decisions

Two software tools Netica: Windows app for working with Bayes-ian belief networks and influence diagrams A commercial product but free for small networks Includes a graphical editor, compiler, inference engine, etc. Samiam: Java system for modeling and reasoning with Bayesian networks Includes a GUI and reasoning engine

Predispositions or causes

Conditions or diseases

Functional Node

Symptoms or effects Dyspnea is shortness of breath

Decision Making with BBNs Today’s weather forecast might be either sunny, cloudy or rainy Should you take an umbrella when you leave? Your decision depends only on the forecast The forecast “depends on” the actual weather Your satisfaction depends on your decision and the weather Assign a utility to each of four situations: (rain|no rain) x (umbrella, no umbrella)

Decision Making with BBNs Extend the BBN framework to include two new kinds of nodes: Decision and Utility A Decision node computes the expected utility of a decision given its parent(s), e.g., forecast, an a valuation A Utility node computes a utility value given its parents, e.g. a decision and weather We can assign a utility to each of four situations: (rain|no rain) x (umbrella, no umbrella) The value assigned to each is probably subjective