Probability, Bayes’ Theorem and the Monty Hall Problem

Slides:



Advertisements
Similar presentations
Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
Advertisements

Probability Monty Hall, Bayes Law, and why Anya probably isn’t autistic Arup Guha 1/14/2013.
Probability Distributions CSLU 2850.Lo1 Spring 2008 Cameron McInally Fordham University May contain work from the Creative Commons.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
Chapter 4 Probability and Probability Distributions
Beyond Null Hypothesis Testing Supplementary Statistical Techniques.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Lec 18 Nov 12 Probability – definitions and simulation.
Probability & Certainty: Intro Probability & Certainty.
Business and Economics 7th Edition
Visual Recognition Tutorial
1 Section 5.1 Discrete Probability. 2 LaPlace’s definition of probability Number of successful outcomes divided by the number of possible outcomes This.
CSE (c) S. Tanimoto, 2008 Bayes Nets 1 Probabilistic Reasoning With Bayes’ Rule Outline: Motivation Generalizing Modus Ponens Bayes’ Rule Applying.
I The meaning of chance Axiomatization. E Plurbus Unum.
CHAPTER 6 Statistical Analysis of Experimental Data
QMS 6351 Statistics and Research Methods Probability and Probability distributions Chapter 4, page 161 Chapter 5 (5.1) Chapter 6 (6.2) Prof. Vera Adamchik.
Probability (cont.). Assigning Probabilities A probability is a value between 0 and 1 and is written either as a fraction or as a proportion. For the.
1 Basic Probability Statistics 515 Lecture Importance of Probability Modeling randomness and measuring uncertainty Describing the distributions.
Lecture 6: Descriptive Statistics: Probability, Distribution, Univariate Data.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Probability and Statistics Review Thursday Sep 11.
Lecture II-2: Probability Review
Lecture 10 – Introduction to Probability Topics Events, sample space, random variables Examples Probability distribution function Conditional probabilities.
COMP14112: Artificial Intelligence Fundamentals L ecture 3 - Foundations of Probabilistic Reasoning Lecturer: Xiao-Jun Zeng
Chapter 1 Basics of Probability.
Independence and Dependence 1 Krishna.V.Palem Kenneth and Audrey Kennedy Professor of Computing Department of Computer Science, Rice University.
PBG 650 Advanced Plant Breeding
Bayes for Beginners Presenters: Shuman ji & Nick Todd.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Independence and Bernoulli.
Dr. Gary Blau, Sean HanMonday, Aug 13, 2007 Statistical Design of Experiments SECTION I Probability Theory Review.
Probability, contd. Learning Objectives By the end of this lecture, you should be able to: – Describe the difference between discrete random variables.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Lecture 5a: Bayes’ Rule Class web site: DEA in Bioinformatics: Statistics Module Box 1Box 2Box 3.
Section 7.1. Section Summary Finite Probability Probabilities of Complements and Unions of Events Probabilistic Reasoning.
 Basic Concepts in Probability  Basic Probability Rules  Connecting Probability to Sampling.
Previous Lecture: Data types and Representations in Molecular Biology.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin A Survey of Probability Concepts Chapter 5.
Independence and Bernoulli Trials. Sharif University of Technology 2 Independence  A, B independent implies: are also independent. Proof for independence.
Lecture 2 Review Probabilities Probability Distributions Normal probability distributions Sampling distributions and estimation.
Topic 2: Intro to probability CEE 11 Spring 2002 Dr. Amelia Regan These notes draw liberally from the class text, Probability and Statistics for Engineering.
Uncertainty Management in Rule-based Expert Systems
Monty Hall problem. Joint probability distribution  In the study of probability, given two random variables X and Y, the joint distribution of X and.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Making sense of randomness
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Please turn off cell phones, pagers, etc. The lecture will begin shortly. There will be a very easy quiz at the end of today’s lecture.
Reasoning with Probs How does evidence lead to conclusions in situations of uncertainty? Bayes Theorem Data fusion, use of techniques that combine data.
Basics on Probability Jingrui He 09/11/2007. Coin Flips  You flip a coin Head with probability 0.5  You flip 100 coins How many heads would you expect.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Review of Statistics I: Probability and Probability Distributions.
Measuring chance Probabilities FETP India. Competency to be gained from this lecture Apply probabilities to field epidemiology.
Introduction to Discrete Probability Epp, section 6.x CS 202.
Probability Theory Modelling random phenomena. Permutations the number of ways that you can order n objects is: n! = n(n-1)(n-2)(n-3)…(3)(2)(1) Definition:
Independence and Dependence 1 Krishna.V.Palem Kenneth and Audrey Kennedy Professor of Computing Department of Computer Science, Rice University.
Computer vision: models, learning and inference Chapter 2 Introduction to probability.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
6.2 Transforming and Combining Random Variables Objectives SWBAT: DESCRIBE the effects of transforming a random variable by adding or subtracting a constant.
HL2 Math - Santowski Lesson 93 – Bayes’ Theorem. Bayes’ Theorem  Main theorem: Suppose we know We would like to use this information to find if possible.
Chap 4-1 Chapter 4 Using Probability and Probability Distributions.
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
Ray Karol 2/26/2013. Let’s Make a Deal Monte Hall Problem Suppose you’re on a game show, and you’re given a choice of three doors: Behind one door is.
Chapter5 Statistical and probabilistic concepts, Implementation to Insurance Subjects of the Unit 1.Counting 2.Probability concepts 3.Random Variables.
Probabilities What is the probability that among 23 people (this class) there will be a shared birthday?
Where are we in CS 440? Now leaving: sequential, deterministic reasoning Entering: probabilistic reasoning and machine learning.
Great Theoretical Ideas In Computer Science
Lecture 2: Probability.
Presentation transcript:

Probability, Bayes’ Theorem and the Monty Hall Problem

Probability Distributions A random variable is a variable whose value is uncertain. For example, the height of a randomly selected person in this class is a random variable – I won’t know its value until the person is selected. Note that we are not completely uncertain about most random variables. For example, we know that height will probably be in the 5’-6’ range. In addition, 5’6” is more likely than 5’0” or 6’0” (for women). The function that describes the probability of each possible value of the random variable is called a probability distribution. PSYC 6130, PROF. J. ELDER

Probability Distributions Probability distributions are closely related to frequency distributions. PSYC 6130, PROF. J. ELDER

Probability Distributions Dividing each frequency by the total number of scores and multiplying by 100 yields a percentage distribution. PSYC 6130, PROF. J. ELDER

Probability Distributions Dividing each frequency by the total number of scores yields a probability distribution. PSYC 6130, PROF. J. ELDER

Probability Distributions For a discrete distribution, the probabilities over all possible values of the random variable must sum to 1. PSYC 6130, PROF. J. ELDER

Probability Distributions For a discrete distribution, we can talk about the probability of a particular score occurring, e.g., p(Province = Ontario) = 0.36. We can also talk about the probability of any one of a subset of scores occurring, e.g., p(Province = Ontario or Quebec) = 0.50. In general, we refer to these occurrences as events. PSYC 6130, PROF. J. ELDER

Probability Distributions For a continuous distribution, the probabilities over all possible values of the random variable must integrate to 1 (i.e., the area under the curve must be 1). Note that the height of a continuous distribution can exceed 1! S h a d e r = . 6 8 3 9 5 4 7 PSYC 6130, PROF. J. ELDER

Continuous Distributions For continuous distributions, it does not make sense to talk about the probability of an exact score. e.g., what is the probability that your height is exactly 65.485948467… inches? Normal Approximation to probability distribution for height of Canadian females (parameters from General Social Survey, 1991) 55 60 65 70 75 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 Height (in) Probability p ? PSYC 6130, PROF. J. ELDER

Continuous Distributions It does make sense to talk about the probability of observing a score that falls within a certain range e.g., what is the probability that you are between 5’3” and 5’7”? e.g., what is the probability that you are less than 5’10”? Valid events Normal Approximation to probability distribution for height of Canadian females (parameters from General Social Survey, 1991) 55 60 65 70 75 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 Height (in) Probability p PSYC 6130, PROF. J. ELDER

Probability of Combined Events PSYC 6130, PROF. J. ELDER

Probability of Combined Events PSYC 6130, PROF. J. ELDER

Exhaustive Events Two or more events are said to be exhaustive if at least one of them must occur. For example, if A is the event that the respondent sleeps less than 6 hours per night and B is the event that the respondent sleeps at least 6 hours per night, then A and B are exhaustive. (Although A is probably the more exhausted!!) PSYC 6130, PROF. J. ELDER

Independence PSYC 6130, PROF. J. ELDER

An Example: The Monty Hall Problem PSYC 6130, PROF. J. ELDER

Problem History When problem first appeared in Parade, approximately 10,000 readers, including 1,000 PhDs, wrote claiming the solution was wrong. In a study of 228 subjects, only 13% chose to switch. PSYC 6130, PROF. J. ELDER

Intuition Before Monty opens any doors, there is a 1/3 probability that the car lies behind the door you selected (Door 1), and a 2/3 probability it lies behind one of the other two doors. Thus with 2/3 probability, Monty will be forced to open a specific door (e.g., the car lies behind Door 2, so Monty must open Door 3). This concentrates all of the 2/3 probability in the remaining door (e.g., Door 2). PSYC 6130, PROF. J. ELDER

PSYC 6130, PROF. J. ELDER

Analysis Car hidden behind Door 1 Car hidden behind Door 2 Player initially picks Door 1 Host opens either Door 2 or 3 Host must open Door 3 Host must open Door 2 Switching loses with probability 1/6 Switching loses with probability 1/6 Switching wins with probability 1/3 Switching wins with probability 1/3 Switching loses with probability 1/3 Switching wins with probability 2/3 PSYC 6130, PROF. J. ELDER

Notes It is important that Monty must open a door that reveals a goat Monty cannot open the door you selected These rules mean that your choice may constrain what Monty does. If you initially selected a door concealing a goat, then there is only one door Monty can open. One can rigorously account for the Monty Hall problem using a Bayesian analysis PSYC 6130, PROF. J. ELDER

End of Lecture 2 Sept 17, 2008

Conditional Probability To understand Bayesian inference, we first need to understand the concept of conditional probability. What is the probability I will roll a 12 with a pair of (fair) dice? What if I first roll one die and get a 6? What now is the probability that when I roll the second die they will sum to 12? “Probability of C given A” PSYC 6130, PROF. J. ELDER

Conditional Probability The conditional probability of A given B is the joint probability of A and B, divided by the marginal probability of B. Thus if A and B are statistically independent, However, if A and B are statistically dependent, then PSYC 6130, PROF. J. ELDER

Bayes’ Theorem Bayes’ Theorem is simply a consequence of the definition of conditional probabilities: Bayes’ Equation PSYC 6130, PROF. J. ELDER

Bayes’ Theorem Bayes’ theorem is most commonly used to estimate the state of a hidden, causal variable H based on the measured state of an observable variable D: Likelihood Prior Posterior Evidence PSYC 6130, PROF. J. ELDER

Bayesian Inference Whereas the posterior p(H|D) is often difficult to estimate directly, reasonable models of the likelihood p(D|H) can often be formed. This is typically because H is causal on D. Thus Bayes’ theorem provides a means for estimating the posterior probability of the causal variable H based on observations D. PSYC 6130, PROF. J. ELDER

Marginalizing To calculate the evidence p(D) in Bayes’ equation, we typically have to marginalize over all possible states of the causal variable H. PSYC 6130, PROF. J. ELDER

The Full Monty Let’s get back to The Monty Hall Problem. Let’s assume you initially select Door 1. Suppose that Monty then opens Door 2 to reveal a goat. We want to calculate the posterior probability that a car lies behind Door 1 after Monty has provided these new data. PSYC 6130, PROF. J. ELDER

The Full Monty PSYC 6130, PROF. J. ELDER

The Full Monty PSYC 6130, PROF. J. ELDER

But we’re not on Let’s Make a Deal! Why is the Monty Hall Problem Interesting? It reveals limitations in human cognitive processing of uncertainty It provides a good illustration of many concepts of probability It get us to think more carefully about how we deal with and express uncertainty as scientists. What else is Bayes’ theorem good for? PSYC 6130, PROF. J. ELDER

Clinical Example Christiansen et al (2000) studied the mammogram results of 2,227 women at health centers of Harvard Pilgrim Health Care, a large HMO in the Boston metropolitan area. The women received a total of 9,747 mammograms over 10 years. Their ages ranged from 40 to 80. Ninety-three different radiologists read the mammograms, and overall they diagnosed 634 mammograms as suspicious that turned out to be false positives. This is a false positive rate of 6.5%. The false negative rate has been estimated at 10%. PSYC 6130, PROF. J. ELDER

Clinical Example There are about 58,500,000 women between the ages of 40 and 80 in the US The incidence of breast cancer in the US is about 184,200 per year, i.e., roughly 1 in 318. PSYC 6130, PROF. J. ELDER

Clinical Example PSYC 6130, PROF. J. ELDER