Review of Probability.

Slides:



Advertisements
Similar presentations
Review of Probability.
Advertisements

Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
Chapter 4: Reasoning Under Uncertainty
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
Probability Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
Chapter 4 Probability.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Probability: Review TexPoint fonts used in EMF.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
2-1 Sample Spaces and Events Conducting an experiment, in day-to-day repetitions of the measurement the results can differ slightly because of small.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Probability and Statistics Review Thursday Sep 11.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
Does Naïve Bayes always work?
Sampling Distributions  A statistic is random in value … it changes from sample to sample.  The probability distribution of a statistic is called a sampling.
Chapter 4 Probability Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Probability and Probability Distributions
Recitation 1 Probability Review
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 2nd Mathematical Preparations (1): Probability and statistics Kazuyuki Tanaka Graduate.
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
Chapter 1 Probability and Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Chapter 8 Probability Section R Review. 2 Barnett/Ziegler/Byleen Finite Mathematics 12e Review for Chapter 8 Important Terms, Symbols, Concepts  8.1.
Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Chapter 4 Probability.
Probability theory Petter Mostad Sample space The set of possible outcomes you consider for the problem you look at You subdivide into different.
2-1 Sample Spaces and Events Random Experiments Figure 2-1 Continuous iteration between model and physical system.
2-1 Sample Spaces and Events Random Experiments Figure 2-1 Continuous iteration between model and physical system.
Uncertainty Uncertain Knowledge Probability Review Bayes’ Theorem Summary.
Chapter 4 Probability ©. Sample Space sample space.S The possible outcomes of a random experiment are called the basic outcomes, and the set of all basic.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 5.2: Recap on Probability Theory Jürgen Sturm Technische Universität.
Chapter 13 February 19, Acting Under Uncertainty Rational Decision – Depends on the relative importance of the goals and the likelihood of.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Uncertainty in Expert Systems
Probability Refresher. Events Events as possible outcomes of an experiment Events define the sample space (discrete or continuous) – Single throw of a.
Reasoning Under Uncertainty. 2 Objectives Learn the meaning of uncertainty and explore some theories designed to deal with it Find out what types of errors.
Probability theory Tron Anders Moger September 5th 2007.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 4 Introduction to Probability Experiments, Counting Rules, and Assigning Probabilities.
V7 Foundations of Probability Theory „Probability“ : degree of confidence that an event of an uncertain nature will occur. „Events“ : we will assume that.
Probabilistic Robotics Introduction.  Robotics is the science of perceiving and manipulating the physical world through computer-controlled devices.
Probabilistic Robotics
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Chapter 4 Probability.
Chap 4-1 Chapter 4 Using Probability and Probability Distributions.
Anifuddin Azis UNCERTAINTY. 2 Introduction The world is not a well-defined place. There is uncertainty in the facts we know: What’s the temperature? Imprecise.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Statistics for Managers 5th Edition
Pattern Recognition Probability Review
Lecture 1.31 Criteria for optimal reception of radio signals.
Bayesian Reasoning Chapter 13 Thomas Bayes,
Chapter 3 Probability.
Does Naïve Bayes always work?
Chapter 4 Probability.
Graduate School of Information Sciences, Tohoku University
Quick Review Probability Theory
Quick Review Probability Theory
Appendix A: Probability Theory
Graduate School of Information Sciences, Tohoku University
Review of Probability and Estimators Arun Das, Jason Rebello
State Estimation Probability, Bayes Filtering
Probability Review 11/22/2018.
CSE-490DF Robotics Capstone
Professor Marie desJardins,
Statistical NLP: Lecture 4
Chapter 2 Notes Math 309 Probability.
Class #21 – Monday, November 10
Bayesian Reasoning Chapter 13 Thomas Bayes,
Bayesian Reasoning Chapter 13 Thomas Bayes,
Probability Review 2/24/2019.
Presentation transcript:

Review of Probability

Axioms of Probability Theory Pr(A) denotes probability that proposition A is true. (A is also called event, or random variable).

A Closer Look at Axiom 3 B

Using the Axioms to prove new properties We proved this

Probability of Events Sample space and events Sample space S: (e.g., all people in an area) Events E1  S: (e.g., all people having cough) E2  S: (e.g., all people having cold) Prior (marginal) probabilities of events P(E) = |E| / |S| (frequency interpretation) P(E) = 0.1 (subjective probability) 0 <= P(E) <= 1 for all events Two special events:  and S: P() = 0 and P(S) = 1.0 Boolean operators between events (to form compound events) Conjunctive (intersection): E1 ^ E2 ( E1  E2) Disjunctive (union): E1 v E2 ( E1  E2) Negation (complement): ~E (E = S – E) C

Probabilities of compound events P(~E) = 1 – P(E) because P(~E) + P(E) =1 P(E1 v E2) = P(E1) + P(E2) – P(E1 ^ E2) But how to compute the joint probability P(E1 ^ E2)? Conditional probability (of E1, given E2) How likely E1 occurs in the subspace of E2 E ~E E2 E1 E1 ^ E2 Using Venn diagrams and decision trees is very useful in proofs and reasonings

The main thing to remember for Bayes E1 ^ E2

Independence, Mutual Exclusion and Exhaustive sets of events Independence assumption Two events E1 and E2 are said to be independent of each other if (given E2 does not change the likelihood of E1) It can simplify the computation Mutually exclusive (ME) and exhaustive (EXH) set of events ME: EXH:

Mutual Exclusive set of events Exhaustive sets of events

Mutual Exclusive and Exhaustive set of events No overlap AND

Random Variables

Discrete Random Variables X denotes a random variable. X can take on a finite number of values in set {x1, x2, …, xn}. P(X=xi), or P(xi), is the probability that the random variable X takes on value xi. P( ) is called probability mass function. E.g. . These are four possibilities of value of X. Sum of these values must be 1.0

Discrete Random Variables: visualization Finite set of possible outcomes X binary:

Continuous Random Variable Probability distribution (density function) over continuous values 5 7

Continuous Random Variables X takes on values in the continuum. p(X=x), or p(x), is a probability density function (PDF). E.g. p(x) x

Probability Distribution Probability distribution P(X|x) X is a random variable Discrete Continuous x is background state of information

Joint and Conditional Probabilities

Joint and Conditional Probabilities Joint Probabilities Probability that both X=x and Y=y Conditional Probabilities Probability that X=x given we know that Y=y

Joint and Conditional Probability P(X=x and Y=y) = P(x,y) If X and Y are independent then P(x,y) = P(x) P(y) P(x | y) is the probability of x given y P(x | y) = P(x,y) / P(y) P(x,y) = P(x | y) P(y) If X and Y are independent then P(x | y) = P(x) divided

Law of Total Probability Discrete case Continuous case

Rules of Probability: Marginalization Product Rule Marginalization X binary:

Questions and Problems Axioms of probability theory Use of Veitch Diagrams to understand basics of Probability theory. Conditional Probability. Independence, Mutual Exclusion and Exhaustive sets of events Derivation of Bayes Theorem. What are random variables? What are discrete random variables? What are continuous random variables. Joint and Conditional Probabilities Law of Total Probability Marginalization Rules. What are Gaussian Normal Distribution? What is mean value? What is variance? What are Gaussian Networks?

Questions and Problems Give example of applying Bayesian Reasoning in real life problem

Gaussian, Mean and Variance N(m, s)

Gaussian (normal) distributions N(m, s) different mean different variance

Each variable is a linear function of its parents, Gaussian networks Each variable is a linear function of its parents, with Gaussian noise Joint probability density functions: X Y X Y

Reverend Thomas Bayes (1702-1761) Clergyman and mathematician who first used probability inductively. These researches established a mathematical basis for probability inference

Bayes Rule

B 40 People who have cancer 100 People who smoke All people = 1000 10/40 = probability that you smoke if you have cancer = P(smoke/cancer) 10/100 = probability that you have cancer if you smoke 40 People who have cancer 1000-100 = 900 people who do not smoke 100 People who smoke 1000-40 = 960 people who do not have cancer 10 People who smoke and have cancer B E = smoke, H = cancer Prob(Cancer/Smoke) = P (smoke/Cancer) * P (Cancer) / P(smoke) All people = 1000 P(smoke) = 100/1000 P(cancer) = 40/1000 P(smoke/Cancer) = 10/40 = 25% Prob(Cancer/Smoke) = 10/40 * 40/1000/ 100 = 10/1000 / 100 = 10/10,000 =/1000 = 0.1%

B 40 People who have cancer 100 People who smoke All people = 1000 10/40 = probability that you smoke if you have cancer = P(smoke/cancer) 40 People who have cancer 100 People who smoke 10/100 = probability that you have cancer if you smoke 10 People who smoke and have cancer 1000-100 = 900 people who do not smoke 1000-40 = 960 people who do not have cancer B E = smoke, H = cancer Prob(Cancer/Smoke) = P (smoke/Cancer) * P (Cancer) / P(smoke) All people = 1000 P(smoke) = 100/1000 P(cancer) = 40/1000 P(smoke/Cancer) = 10/40 = 25% Prob(Cancer/Smoke) = 10/40 * 40/1000/ 100 = 10/1000 / 100 = 10/10,000 = 1/1000 = 0.1% Prob(Cancer/Not smoke) = 30/40 * 40/100 / 900 = 30/100*900 = 30 / 90,000 = 1/3,000 = 0.03 % E = smoke, H = cancer Prob(Cancer/Not Smoke) = P (Not smoke/Cancer) * P (Cancer) / P(Not smoke)

Bayes’ Theorem with relative likelihood In the setting of diagnostic/evidential reasoning Know prior probability of hypothesis conditional probability Want to compute the posterior probability Bayes’ theorem (formula 1): If the purpose is to find which of the n hypotheses is more plausible given , then we can ignore the denominator and rank them, use relative likelihood

Relative likelihood can be computed from and , if we assume all hypotheses are ME and EXH Then we have another version of Bayes’ theorem: where , the sum of relative likelihood of all n hypotheses, is a normalization factor Mutually exclusive (ME) and exhaustive (EXH)

Naïve Bayesian Approach Knowledge base: Case input: Find the hypothesis with the highest posterior probability By Bayes’ theorem Assume all pieces of evidence are conditionally independent, given any hypothesis

Do not worry, many examples will follow absolute posterior probability The relative likelihood The absolute posterior probability Evidence accumulation (when new evidence is discovered) substitute Do not worry, many examples will follow

Bayesian Networks and Markov Models Bayesian AI Bayesian Filters Bayesian networks Decision networks Reasoning about changes over time Dynamic Bayesian Networks Markov models