Module #16: Probability Theory

Slides:



Advertisements
Similar presentations
CS433: Modeling and Simulation
Advertisements

Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Basic Business Statistics.
Economics 105: Statistics Any questions? Go over GH2 Student Information Sheet.
1 Probability Theory Rosen 5 th ed., Chapter 5. 2 Random Experiment - Terminology A random (or stochastic) experiment is an experiment whose result is.
1 Business Statistics - QBM117 Assigning probabilities to events.
1 MF-852 Financial Econometrics Lecture 4 Probability Distributions and Intro. to Hypothesis Tests Roy J. Epstein Fall 2003.
Probability theory Much inspired by the presentation of Kren and Samuelsson.
Chapter 9 Introducing Probability - A bridge from Descriptive Statistics to Inferential Statistics.
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
Chapter 1 Probability and Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Theory of Probability Statistics for Business and Economics.
Basic Concepts of Discrete Probability (Theory of Sets: Continuation) 1.
LECTURE 15 THURSDAY, 15 OCTOBER STA 291 Fall
LECTURE 14 TUESDAY, 13 OCTOBER STA 291 Fall
Lecture V Probability theory. Lecture questions Classical definition of probability Frequency probability Discrete variable and probability distribution.
Basic Principles (continuation) 1. A Quantitative Measure of Information As we already have realized, when a statistical experiment has n eqiuprobable.
2. Introduction to Probability. What is a Probability?
Probability Definition : The probability of a given event is an expression of likelihood of occurrence of an event.A probability isa number which ranges.
Sixth lecture Concepts of Probabilities. Random Experiment Can be repeated (theoretically) an infinite number of times Has a well-defined set of possible.
Random Variables Ch. 6. Flip a fair coin 4 times. List all the possible outcomes. Let X be the number of heads. A probability model describes the possible.
V7 Foundations of Probability Theory „Probability“ : degree of confidence that an event of an uncertain nature will occur. „Events“ : we will assume that.
Lecture 7 Dustin Lueker. 2STA 291 Fall 2009 Lecture 7.
Textbook Basics of an Expert System: – “Expert systems: Design and Development,” by: John Durkin, 1994, Chapters 1-4. Uncertainty (Probability, Certainty.
Module #16: Probability Theory
Lecture 6 Dustin Lueker.  Standardized measure of variation ◦ Idea  A standard deviation of 10 may indicate great variability or small variability,
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
Chapter 8: Probability: The Mathematics of Chance Probability Models and Rules 1 Probability Theory  The mathematical description of randomness.  Companies.
Module #1 - Logic 1 Based on Rosen, Discrete Mathematics & Its Applications. Prepared by (c) , Michael P. Frank and Modified By Mingwu Chen Probability.
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Business Statistics,
13.1 CompSci 102© Michael Frank Today’s topics ProbabilityProbability –Definitions –Events –Conditional probability Reading: Sections Reading: Sections.
Sample Space and Events Section 2.1 An experiment: is any action, process or phenomenon whose outcome is subject to uncertainty. An outcome: is a result.
14.1 CompSci 102© Michael Frank Today’s topics ProbabilityProbability –Expected value Reading: Sections 5.3Reading: Sections 5.3 UpcomingUpcoming –Probabilistic.
AP Statistics From Randomness to Probability Chapter 14.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate accuracy.
4. Overview of Probability Network Performance and Quality of Service.
Essential Ideas for The Nature of Probability
Statistics 200 Objectives:
Review of Probability.
Probability – part 1 Prof. Carla Gomes Rosen 1 1.
B504/I538: Introduction to Cryptography
Chapter 6 6.1/6.2 Probability Probability is the branch of mathematics that describes the pattern of chance outcomes.
CS104:Discrete Structures
Reasoning Under Uncertainty in Expert System
5.1 Probability of Simple Events
Discrete Probability Chapter 7 With Question/Answer Animations
From Randomness to Probability
Chapter 4. Probability
From Randomness to Probability
Now it’s time to look at…
Chapter 4 – Part 3.
STA 291 Spring 2008 Lecture 7 Dustin Lueker.
Probability Models Section 6.2.
Chapter 5 Some Important Discrete Probability Distributions
STA 291 Spring 2008 Lecture 6 Dustin Lueker.
Honors Statistics From Randomness to Probability
Module #16: Probability Theory
Discrete Mathematics and its Applications
Probability, Statistics
Discrete Mathematics and its Applications
Discrete Mathematics and its Applications
Mr. Reider AP Stat November 18, 2010
Chapter 8: Probability: The Mathematics of Chance Lesson Plan
28th September 2005 Dr Bogdan L. Vrusias
6.1 Sample space, events, probability
From Randomness to Probability
Module #16: Probability Theory
Chapter 8: Probability: The Mathematics of Chance Lesson Plan
A random experiment gives rise to possible outcomes, but any particular outcome is uncertain – “random”. For example, tossing a coin… we know H or T will.
Sets, Combinatorics, Probability, and Number Theory
Presentation transcript:

Module #16: Probability Theory Rosen 5th ed., ch. 5 23 slides, ~1 lecture 9/19/2018 (c)2001-2002, Michael P. Frank

Why Probability? In the real world, we often don’t know whether a given proposition is true or false. Probability theory gives us a way to reason about propositions whose truth is uncertain. Useful in weighing evidence, diagnosing problems, and analyzing situations whose exact details are unknown. 9/19/2018 (c)2001-2002, Michael P. Frank

Random Variables A random variable V is a variable whose value is unknown, or that depends on the situation. E.g., the number of students in class today Whether it will rain tonight (Boolean variable) Let the domain of V be dom[V]={v1,…,vn} The proposition V=vi may be uncertain, and be assigned a probability. 9/19/2018 (c)2001-2002, Michael P. Frank

Amount of Information The amount of information I[V] in a random variable V is the logarithm of the size of the domain of V, I[V]=log |dom[V]|. The base of the logarithm determines the information unit; base 2 gives a unit of 1 bit. Example: An 8-bit register has 28 = 256 possible values. Log 256 = 8 bits. 9/19/2018 (c)2001-2002, Michael P. Frank

Experiments A (stochastic) experiment is a process by which a given random variable gets assigned a specific value. The sample space S of the experiment is the domain of the random variable. The outcome of the experiment is the specific value of the random variable that is selected. 9/19/2018 (c)2001-2002, Michael P. Frank

Events An event E is a set of possible outcomes That is, E  S = dom[V]. We say that event E occurs when VE. Note that VE is the (uncertain) proposition that the actual outcome will be one of the outcomes in the set E. 9/19/2018 (c)2001-2002, Michael P. Frank

Probability The probability p = Pr[E]  [0,1] of an event E is a real number representing our degree of certainty that E will occur. If Pr[E] = 1, then E is absolutely certain to occur, thus VE is true. If Pr[E] = 0, then E is absolutely certain not to occur, thus VE is false. If Pr[E] = ½, then we are completely uncertain about whether E will occur; that is, VE and VE are considered equally likely. What about other cases? 9/19/2018 (c)2001-2002, Michael P. Frank

Four Definitions of Probability Several alternative definitions of probability are commonly encountered: Frequentist, Bayesian, Laplacian, Axiomatic They have different strengths & weaknesses. Fortunately, they coincide and work well with each other in most cases. 9/19/2018 (c)2001-2002, Michael P. Frank

Probability: Frequentist Definition The probability of an event E is the limit, as n→∞, of the fraction of times that VE in n repetitions of the same experiment. Problems: Only well-defined for experiments that are infinitely repeatable (at least in principle). Can never be measured exactly in finite time! Advantage: Objective, mathematical def’n. 9/19/2018 (c)2001-2002, Michael P. Frank

Probability: Bayesian Definition Suppose a rational entity R is offered a choice between two rewards: Winning $1 if event E occurs. Receiving p dollars (where p[0,1]) unconditionally. If R is indifferent between these two rewards, then we say R’s probability for E is p. Problem: Subjective definition, depends on the reasoner R, and his knowledge & rationality. 9/19/2018 (c)2001-2002, Michael P. Frank

Probability: Laplacian Definition First, assume that all outcomes in the sample space are equally likely This term still needs to be defined. Then, the probability of event E, Pr[E] = |E|/|S|. Very simple! Problems: Still needs a definition for equally likely, and depends on existence of a finite sample space with all equally likely outcomes. 9/19/2018 (c)2001-2002, Michael P. Frank

Probability: Axiomatic Definition Let p be any function p:S→[0,1], such that: 0 ≤ p(s) ≤ 1 for all outcomes sS. ∑ p(s) = 1. Such a p is called a probability distribution. Then, the probability of any event ES is just: Advantage: Totally mathematically well-defined. Problem: Leaves operational def’n unspecified. 9/19/2018 (c)2001-2002, Michael P. Frank

Probability of Complementary Events Let E be an event in a sample space S. Then, E represents the complementary event that VE. Pr[E] = 1 − Pr[E] 9/19/2018 (c)2001-2002, Michael P. Frank

Probability of Unions of Events Let E1,E2  S = dom[V]. Then: Pr[E1 E2] = Pr[E1] + Pr[E2] − Pr[E1E2] By the inclusion-exclusion principle. 9/19/2018 (c)2001-2002, Michael P. Frank

Mutually Exclusive Events Two events E1, E2 are called mutually exclusive if they are disjoint: E1E2 =  Note that two mutually exclusive events cannot both occur in the same instance of a given experiment. For mutually exclusive events, Pr[E1  E2] = Pr[E1] + Pr[E2]. 9/19/2018 (c)2001-2002, Michael P. Frank

Exhaustive Sets of Events A set E = {E1, E2, …} of events in the sample space S is exhaustive if . An exhaustive set of events that are all mutually exclusive with each other has the property that 9/19/2018 (c)2001-2002, Michael P. Frank

Independent Events Two events E,F are independent if Pr[EF] = Pr[E]·Pr[F]. Relates to product rule for number of ways of doing two independent tasks Example: Flip a coin, and roll a die. Pr[ quarter is heads  die is 1 ] = Pr[quarter is heads] × Pr[die is 1] 9/19/2018 (c)2001-2002, Michael P. Frank

Conditional Probability Let E,F be events such that Pr[F]>0. Then, the conditional probability of E given F, written Pr[E|F], is defined as Pr[EF]/Pr[F]. This is the probability that E would turn out to be true, given just the information that F is true. If E and F are independent, Pr[E|F] = Pr[E]. 9/19/2018 (c)2001-2002, Michael P. Frank

Bayes’s Theorem Allows one to compute the probability that a hypothesis H is correct, given data D: Easy to prove from def’n of conditional prob. Extremely useful in artificial intelligence apps: Data mining, automated diagnosis, pattern recognition, statistical modeling, evaluating scientific hypotheses. 9/19/2018 (c)2001-2002, Michael P. Frank

Expectation Values For a random variable V having a numeric domain, its expectation value or expected value or weighted average value or arithmetic mean value Ex[V] is defined as ∑vdom[V] v·p(v). The term “expected value” is widely used, but misleading since the expected value might be totally unexpected or impossible! 9/19/2018 (c)2001-2002, Michael P. Frank

Derived Random Variables Let S be a sample space over values of a random variable V (representing possible outcomes). Then, any function f over S can also be considered to be a random variable (whose value is derived from the value of V). If the range R = range[f] of f is numeric, then Ex[f] can still be defined, as 9/19/2018 (c)2001-2002, Michael P. Frank

Linearity of Expectation Let X1, X2 be any two random variables derived from the same sample space. Then: Ex[X1+X2] = Ex[X1] + Ex[X2] Ex[aX1 + b] = aEx[X1] + b 9/19/2018 (c)2001-2002, Michael P. Frank

Variance The variance Var[X] = σ2(X) of a random variable X is the expected value of the square of the difference between the value of X and its expectation value Ex[X]: The standard deviation or root-mean-square (RMS) difference of X, σ(X) :≡ Var[X]1/2. 9/19/2018 (c)2001-2002, Michael P. Frank

Entropy The entropy H of a probability distribution p over a sample space S over outcomes is a measure of our degree of uncertainty about the outcome. It measures the expected amount of increase in known information from learning the actual outcome. The base of the logarithm gives the unit of entropy; base 2 → 1 bit, base e → 1 nat 1 nat is also known as “Boltzmann’s constant” kB & as the “ideal gas constant” R, first discovered physically 9/19/2018 (c)2001-2002, Michael P. Frank