Foundations of Artificial Intelligence 1 Bayes’ Rule - Example  Medical Diagnosis  suppose we know from statistical data that flu causes fever in 80%

Slides:



Advertisements
Similar presentations
Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
Advertisements

Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Probability: Review The state of the world is described using random variables Probabilities are defined over events –Sets of world states characterized.
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
FT228/4 Knowledge Based Decision Support Systems
Evaluating Diagnostic Accuracy of Prostate Cancer Using Bayesian Analysis Part of an Undergraduate Research course Chantal D. Larose.
For Monday Finish chapter 14 Homework: –Chapter 13, exercises 8, 15.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
M.I. Jaime Alfonso Reyes ´Cortés.  The basic task for any probabilistic inference system is to compute the posterior probability distribution for a set.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 12 Jim Martin.
Visual Recognition Tutorial
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Inductive Reasoning Bayes Rule. Urn problem (1) A B A die throw determines from which urn to select balls. For outcomes 1,2, and 3, balls are picked from.
Section 9.3 Testing a Proportion p. 2 Focus Points Identify the components needed for testing a proportion. Compute the sample test statistic. Find the.
5-3 Inference on the Means of Two Populations, Variances Unknown
Statistical Analysis. Purpose of Statistical Analysis Determines whether the results found in an experiment are meaningful. Answers the question: –Does.
Probability, Bayes’ Theorem and the Monty Hall Problem
Systems of Equations and Inequalities
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
PBG 650 Advanced Plant Breeding
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
P Values Robin Beaumont 10/10/2011 With much help from Professor Chris Wilds material University of Auckland.
Lab 4 1.Get an image into a ROS node 2.Find all the orange pixels (suggest HSV) 3.Identify the midpoint of all the orange pixels 4.Explore the findContours.
1 As we have seen in section 4 conditional probability density functions are useful to update the information about an event based on the knowledge about.
Statistical Decision Theory
AP Statistics Section 8.1: The Binomial Distribution.
Introduction to Probability Theory March 24, 2015 Credits for slides: Allan, Arms, Mihalcea, Schutze.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Digital Statisticians INST 4200 David J Stucki Spring 2015.
Sampling Distribution of the Sample Mean. Example a Let X denote the lifetime of a battery Suppose the distribution of battery battery lifetimes has 
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
1 Reasoning Under Uncertainty Artificial Intelligence Chapter 9.
CSE PR 1 Reasoning - Rule-based and Probabilistic Representing relations with predicate logic Limitations of predicate logic Representing relations.
NLP. Introduction to NLP Formula for joint probability –p(A,B) = p(B|A)p(A) –p(A,B) = p(A|B)p(B) Therefore –p(B|A)=p(A|B)p(B)/p(A) Bayes’ theorem is.
Reasoning with Uncertainty; Probabilistic Reasoning Foundations of Artificial Intelligence.
Uncertainty in Expert Systems
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 5.3: Reasoning with Bayes Law Jürgen Sturm Technische Universität.
4 Proposed Research Projects SmartHome – Encouraging patients with mild cognitive disabilities to use digital memory notebook for activities of daily living.
P Values Robin Beaumont 8/2/2012 With much help from Professor Chris Wilds material University of Auckland.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
Naïve Bayes Classification Material borrowed from Jonathan Huang and I. H. Witten’s and E. Frank’s “Data Mining” and Jeremy Wyatt and others.
P Values - part 2 Samples & Populations Robin Beaumont 2011 With much help from Professor Chris Wilds material University of Auckland.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Conditional Probability, Bayes’ Theorem, and Belief Networks CISC 2315 Discrete Structures Spring2010 Professor William G. Tanner, Jr.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Chapter 8 – Naïve Bayes DM for Business Intelligence.
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 8: Estimating with Confidence Section 8.2 Estimating a Population Proportion.
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
Bayes for Beginners Anne-Catherine Huys M. Berk Mirza Methods for Dummies 20 th January 2016.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Bayesian Classification 1. 2 Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership.
Statistics 200 Objectives:
Probability Axioms and Formulas
Chapter 10: Using Uncertain Knowledge
Today.
Constraints on Credence
Markov ó Kalman Filter Localization
Chapter 8: Estimating with Confidence
CS 188: Artificial Intelligence Fall 2008
CS 188: Artificial Intelligence Fall 2007
Estimating a Population Proportion
Chapter 8: Estimating with Confidence
Wellcome Trust Centre for Neuroimaging
LECTURE 07: BAYESIAN ESTIMATION
11. Conditional Density Functions and Conditional Expected Values
2/3/ Estimating a Population Proportion.
11. Conditional Density Functions and Conditional Expected Values
MAS2317 Presentation Question 1
Presentation transcript:

Foundations of Artificial Intelligence 1 Bayes’ Rule - Example  Medical Diagnosis  suppose we know from statistical data that flu causes fever in 80% of cases, approximately 1 in 10,000 people have flu at a given time, and approximately 1 out of every 50 people is suffering from fever: Pr(fever | flu) = 0.8Pr(flu) = Pr(fever) = 0.02  Given a patient with fever, does she have flu? Answer by applying Bayes’ rule: Pr(flu | fever) = [ Pr(fever | flu). Pr(flu) ] / Pr(fever) = 0.8 x / 0.02 = 0.004

Foundations of Artificial Intelligence 2 Bayes’ Rule & Normalization  Direct assessments of prior probabilities for evidence (or symptoms in diagnosis problems) may not be possible  we can avoid direct assessment using normalization:  now since Pr(H | E) + Pr(  H | E) = 1, we obtain  substituting in the equation for Pr(H | E) we get:  this allows for conditional terms to sum to 1  so at the cost of assessing Pr(E |  H), we can avoid assessing Pr(E), and still obtain exact probabilities using the Bayes’ rule.

Foundations of Artificial Intelligence 3 Example: Normalization  Suppose A blood test is 90% effective in detecting a disease. It also falsely diagnoses that a healthy person has the disease 3% of the time. If 10% of those tested have the disease, what is the probability that a person who tests positive will actually have the disease? ( i.e. find P (disease | positive) )  P(disease) = 0.10  P(  disease) = 0.90  P(positive | disease) = 0.90  P(positive |  disease) = 0.03   P (disease | positive) = ––––––––––––––––––––––––––––––––––––––––––––––––––––  = (0.90)(0.10) / ( (0.90)(0.10) + (0.03)(0.90) )  = 0.77 P(positive | disease) * P(disease) P(positive|disease)*P(disease) + P(positive|  disease)*P(  disease)

Foundations of Artificial Intelligence 4 Bayesian Updating  As each new evidence is observed, the belief in the unknown variable is multiplied by a factor that depends on the new evidence  suppose we have obtained, using Bayes’ rule and evidence E 1 a probability for our diagnosis:  now a new evidence E 2 is observed; we can apply Bayes’ rule with E 1 as the constant conditioning context:  If we assume the pieces of evidence are conditionally independent given H, we have:  then: Bayesian updating formula

Foundations of Artificial Intelligence 5 Bayesian Updating- Example  In the previous example, we had: Pr(fever | flu) = 0.8, Pr(flu) = , Pr(fever) = 0.02 allowing us to conclude that Pr(flu | fever) = using Bayes’ rule.  Observe the new symptom, soar-throat. We know that flu causes soar-throat 50% of the time (i.e., Pr(soar | flu) = 0.5) and 1/200 = people have soar throats. We can now update our diagnosis that she has flu.  Assuming that symptoms are conditionally independent given flu, the Bayesian updating gives:  But, we don’t have Pr(soar | fever). What can we do?  Note that  So:  i.e.,

Foundations of Artificial Intelligence 6 Bayesian Updating- Example  So, we have:  From before, we know:  Pr(flu | fever) = and so Pr(  flu | fever) =  Pr(soar | flu) = 0.5  We don’t know Pr(soar |  flu), but let’s assume that this is about the same as Pr(soar):  Pr(soar |  flu) = Pr(soar) = 1/200 =  Note: this seems reasonable, since few people have flu anyway, and so the probability of having soar throat at random would not be very different from the probability of a person who does not have the flu, having soar throat. In other words, it is reasonable to assume that having soar throat and not having flu are independent.  Putting things together:  Pr(flu | fever /\ soar)= [0.004 * 0.5] / [ (0.004 * 0.5) + (0.996 * 0.005) =0.002 / ( ) =