Review of Probability.

Slides:



Advertisements
Similar presentations
IP, IST, José Bioucas, Probability The mathematical language to quantify uncertainty  Observation mechanism:  Priors:  Parameters Role in inverse.
Advertisements

Lecture 4A: Probability Theory Review Advanced Artificial Intelligence.
Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
Review of Probability Jake Blanchard Spring 2010 Uncertainty Analysis for Engineers1.
Set theory and Bayes’ Rule Sets and set operations Axioms of probability Probabilities of sets Bayes’ rule Chapter 2 of Haldar and Mahadevan’s Probability,
1. Frequency Distribution & Relative Frequency Distribution 2. Histogram of Probability Distribution 3. Probability of an Event in Histogram 4. Random.
Statistics Lecture 18. Will begin Chapter 5 today.
Probability Theory Part 2: Random Variables. Random Variables  The Notion of a Random Variable The outcome is not always a number Assign a numerical.
De Morgan’s rule E 1 = pipe 1 breaks E 2 = pipe 2 breaks 1 2 Water Supply E = failure in water supply = E 1 ∪ E 2 no failure in water supply = Ē = E 1.
Chapter 2 Chapter The sample space of an experiment, denoted S , is the set of all possible outcomes of that experiment. An event is any collection.
Adapted from Walch Education The conditional probability of B given A is the probability that event B occurs, given that event A has already occurred.
Basics of Sampling Theory P = { x 1, x 2, ……, x N } where P = population x 1, x 2, ……, x N are real numbers Assuming x is a random variable; Mean/Average.
Standard error of estimate & Confidence interval.
AP STATISTICS.   Theoretical: true mathematical probability  Empirical: the relative frequency with which an event occurs in a given experiment  Subjective:
Probability and Statistics Dr. Saeid Moloudzadeh Axioms of Probability/ Basic Theorems 1 Contents Descriptive Statistics Axioms of Probability.
Chapter 4 Probability Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Probability Theory and Random Processes
Poisson Random Variable Provides model for data that represent the number of occurrences of a specified event in a given unit of time X represents the.
Unit 1 OUTCOMES AND LIKELIHOODS. Unit Essential Question: How do you determine, interpret, and apply principles of probability?
Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Chapter 4 Probability.
4.1 Probability Distributions. Do you remember? Relative Frequency Histogram.
Probability & Statistics I IE 254 Exam I - Reminder  Reminder: Test 1 - June 21 (see syllabus) Chapters 1, 2, Appendix BI  HW Chapter 1 due Monday at.
Chapter 10 Probability. Experiments, Outcomes, and Sample Space Outcomes: Possible results from experiments in a random phenomenon Sample Space: Collection.
Probability theory Petter Mostad Sample space The set of possible outcomes you consider for the problem you look at You subdivide into different.
K. Shum Lecture 6 Various definitions of Probability.
2-1 Sample Spaces and Events Random Experiments Figure 2-1 Continuous iteration between model and physical system.
2-1 Sample Spaces and Events Random Experiments Figure 2-1 Continuous iteration between model and physical system.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 5.2: Recap on Probability Theory Jürgen Sturm Technische Universität.
Optimal Bayes Classification
 How do you know how long your design is going to last?  Is there any way we can predict how long it will work?  Why do Reliability Engineers get paid.
Dr. Ahmed Abdelwahab Introduction for EE420. Probability Theory Probability theory is rooted in phenomena that can be modeled by an experiment with an.
Lecture V Probability theory. Lecture questions Classical definition of probability Frequency probability Discrete variable and probability distribution.
Monty Hall problem. Joint probability distribution  In the study of probability, given two random variables X and Y, the joint distribution of X and.
Methodology Solving problems with known distributions 1.
Chapter 3 Foundation of Mathematical Analysis § 3.1 Statistics and Probability § 3.2 Random Variables and Magnitude Distribution § 3.3 Probability Density.
Probability: Terminology  Sample Space  Set of all possible outcomes of a random experiment.  Random Experiment  Any activity resulting in uncertain.
Week 4 機率論. Basic Probability Concepts Random Trials Sample Spaces.
8-3: Probability and Probability Distributions English Casbarro Unit 8.
V7 Foundations of Probability Theory „Probability“ : degree of confidence that an event of an uncertain nature will occur. „Events“ : we will assume that.
Chapter 4 Probability, Randomness, and Uncertainty.
2003/02/19 Chapter 2 1頁1頁 Chapter 2 : Basic Probability Theory Set Theory Axioms of Probability Conditional Probability Sequential Random Experiments Outlines.
© 2010 Pearson Prentice Hall. All rights reserved. CHAPTER 11 Counting Methods and Probability Theory.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Chapter IV Statistic and simple probability. A. Meaning of probability 1.Random event 2.Simple event 3.Relative frequent and probability of an event Relative.
BUSA Probability. Probability – the bedrock of randomness Definitions Random experiment – observing the close of the NYSE and the Nasdaq Sample.
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 2nd Probability and its fundamental properties Kazuyuki Tanaka Graduate School of Information.
Chapter 14 Week 5, Monday. Introductory Example Consider a fair coin: Question: If I flip this coin, what is the probability of observing heads? Answer:
Onur DOĞAN.  The Classical Interpretation of Probability  The Frequency Interpretation of Probability  The Subjective Interpretation of Probability.
Chapter 2 Probability  Sample Spaces and Events.2 - Axioms, Interpretations, and Properties of Probability.3 - Counting Techniques.4 - Conditional.
MATH Section 7.2.
Basics of Multivariate Probability
Review of Probability.
Graduate School of Information Sciences, Tohoku University
Axioms, Interpretations and Properties
Statistics and Quantitative Analysis U4320
This presentation uses a free template provided by FPPT.com Quantitative Methods Dr. Aravind Banakar –
This presentation uses a free template provided by FPPT.com Quantitative Methods
This presentation uses a free template provided by FPPT.com Quantitative Methods
Basic Probability Concepts
Finding the Complement of Event E AGENDA:
Probability Review 11/22/2018.
PROBABILITY AND STATISTICS
Probability & Statistics Probability Theory Mathematical Probability Models Event Relationships Distributions of Random Variables Continuous Random.
CSE-490DF Robotics Capstone
Chapter 2 Probability Sample Spaces and Events
Probability Measures: Axioms and Properties
Chapter 4 Section 1 Probability Theory.
Probability Review 2/24/2019.
Section 11.7 Probability.
Presentation transcript:

Review of Probability

Axioms of Probability Theory Pr(A) denotes probability that proposition A is true. (A is also called event, or random variable).

A Closer Look at Axiom 3 B

Using the Axioms to prove new properties We proved this

Probability of Events Sample space and events Sample space S: (e.g., all people in an area) Events E1  S: (e.g., all people having cough) E2  S: (e.g., all people having cold) Prior (marginal) probabilities of events P(E) = |E| / |S| (frequency interpretation) P(E) = 0.1 (subjective probability) 0 <= P(E) <= 1 for all events Two special events:  and S: P() = 0 and P(S) = 1.0 Boolean operators between events (to form compound events) Conjunctive (intersection): E1 ^ E2 ( E1  E2) Disjunctive (union): E1 v E2 ( E1  E2) Negation (complement): ~E (E = S – E) C

Probabilities of compound events P(~E) = 1 – P(E) because P(~E) + P(E) =1 P(E1 v E2) = P(E1) + P(E2) – P(E1 ^ E2) But how to compute the joint probability P(E1 ^ E2)? Conditional probability (of E1, given E2) How likely E1 occurs in the subspace of E2 E ~E E2 E1 E1 ^ E2 Using Venn diagrams and decision trees is very useful in proofs and reasonings

Independence, Mutual Exclusion and Exhaustive sets of events Independence assumption Two events E1 and E2 are said to be independent of each other if (given E2 does not change the likelihood of E1) It can simplify the computation Mutually exclusive (ME) and exhaustive (EXH) set of events ME: EXH:

Random Variables

Discrete Random Variables X denotes a random variable. X can take on a finite number of values in set {x1, x2, …, xn}. P(X=xi), or P(xi), is the probability that the random variable X takes on value xi. P( ) is called probability mass function. E.g. . These are four possibilities of value of X. Sum of these values must be 1.0

Discrete Random Variables Finite set of possible outcomes X binary:

Continuous Random Variable Probability distribution (density function) over continuous values 5 7

Continuous Random Variables X takes on values in the continuum. p(X=x), or p(x), is a probability density function (PDF). E.g. p(x) x

Probability Distribution Probability distribution P(X|x) X is a random variable Discrete Continuous x is background state of information

Joint and Conditional Probabilities Probability that both X=x and Y=y Conditional Probability that X=x given we know that Y=y

Joint and Conditional Probabilities Probability that both X=x and Y=y Conditional Probability that X=x given we know that Y=y

Joint and Conditional Probability P(X=x and Y=y) = P(x,y) If X and Y are independent then P(x,y) = P(x) P(y) P(x | y) is the probability of x given y P(x | y) = P(x,y) / P(y) P(x,y) = P(x | y) P(y) If X and Y are independent then P(x | y) = P(x) divided

Law of Total Probability Discrete case Continuous case

Rules of Probability: Marginalization Product Rule Marginalization X binary:

Gaussian, Mean and Variance N(m, s)

Gaussian (normal) distributions N(m, s) N(m, s) different mean different variance

Each variable is a linear function of its parents, Gaussian networks Each variable is a linear function of its parents, with Gaussian noise Joint probability density functions: X Y X Y

Reverend Thomas Bayes (1702-1761) Clergyman and mathematician who first used probability inductively. These researches established a mathematical basis for probability inference

Bayes Rule

B 40 People who have cancer 100 People who smoke All people = 1000 10/40 = probability that you smoke if you have cancer = P(smoke/cancer) 10/100 = probability that you have cancer if you smoke 40 People who have cancer 1000-100 = 900 people who do not smoke 100 People who smoke 1000-40 = 960 people who do not have cancer 10 People who smoke and have cancer B E = smoke, H = cancer Prob(Cancer/Smoke) = P (smoke/Cancer) * P (Cancer) / P(smoke) All people = 1000 P(smoke) = 100/1000 P(cancer) = 40/1000 P(smoke/Cancer) = 10/40 = 25% Prob(Cancer/Smoke) = 10/40 * 40/1000/ 100 = 10/1000 / 100 = 10/10,000 =/1000 = 0.1%

B 40 People who have cancer 100 People who smoke All people = 1000 10/40 = probability that you smoke if you have cancer = P(smoke/cancer) 40 People who have cancer 100 People who smoke 10/100 = probability that you have cancer if you smoke 10 People who smoke and have cancer 1000-100 = 900 people who do not smoke 1000-40 = 960 people who do not have cancer B E = smoke, H = cancer Prob(Cancer/Smoke) = P (smoke/Cancer) * P (Cancer) / P(smoke) All people = 1000 P(smoke) = 100/1000 P(cancer) = 40/1000 P(smoke/Cancer) = 10/40 = 25% Prob(Cancer/Smoke) = 10/40 * 40/1000/ 100 = 10/1000 / 100 = 10/10,000 = 1/1000 = 0.1% Prob(Cancer/Not smoke) = 30/40 * 40/100 / 900 = 30/100*900 = 30 / 90,000 = 1/3,000 = 0.03 % E = smoke, H = cancer Prob(Cancer/Not Smoke) = P (Not smoke/Cancer) * P (Cancer) / P(Not smoke)

Bayes’ Theorem with relative likelihood In the setting of diagnostic/evidential reasoning Know prior probability of hypothesis conditional probability Want to compute the posterior probability Bayes’ theorem (formula 1): If the purpose is to find which of the n hypotheses is more plausible given , then we can ignore the denominator and rank them, use relative likelihood

Relative likelihood can be computed from and , if we assume all hypotheses are ME and EXH Then we have another version of Bayes’ theorem: where , the sum of relative likelihood of all n hypotheses, is a normalization factor

Naïve Bayesian Approach Knowledge base: Case input: Find the hypothesis with the highest posterior probability By Bayes’ theorem Assume all pieces of evidence are conditionally independent, given any hypothesis

absolute posterior probability The relative likelihood The absolute posterior probability Evidence accumulation (when new evidence is discovered)

Bayesian Networks and Markov Models – applications in robotics Bayesian AI Bayesian Filters Kalman Filters Particle Filters Bayesian networks Decision networks Reasoning about changes over time Dynamic Bayesian Networks Markov models