R. Kass/S07 P416 Lec 3 1 Lecture 3 The Gaussian Probability Distribution Function Plot of Gaussian pdf x p(x)p(x) Introduction l The Gaussian probability.

Slides:



Advertisements
Similar presentations
Evaluating Classifiers
Advertisements

Yaochen Kuo KAINAN University . SLIDES . BY.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Sampling Distributions (§ )

McGraw-Hill Ryerson Copyright © 2011 McGraw-Hill Ryerson Limited. Adapted by Peter Au, George Brown College.
ELEC 303 – Random Signals Lecture 18 – Statistics, Confidence Intervals Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 10, 2009.
1 MF-852 Financial Econometrics Lecture 4 Probability Distributions and Intro. to Hypothesis Tests Roy J. Epstein Fall 2003.
BCOR 1020 Business Statistics Lecture 17 – March 18, 2008.
Statistics. Large Systems Macroscopic systems involve large numbers of particles.  Microscopic determinism  Macroscopic phenomena The basis is in mechanics.
Chapter 6 Continuous Random Variables and Probability Distributions
1 Sociology 601, Class 4: September 10, 2009 Chapter 4: Distributions Probability distributions (4.1) The normal probability distribution (4.2) Sampling.
BCOR 1020 Business Statistics
Probability Distributions Random Variables: Finite and Continuous Distribution Functions Expected value April 3 – 10, 2003.
Introduction to Probability and Statistics Chapter 7 Sampling Distributions.
Chapter 5 Continuous Random Variables and Probability Distributions
Inferences About Process Quality
Chi Square Distribution (c2) and Least Squares Fitting
Continous Probability Distributions
Inferential Statistics
The Central Limit Theorem For simple random samples from any population with finite mean and variance, as n becomes increasingly large, the sampling distribution.
Continuous Probability Distribution  A continuous random variables (RV) has infinitely many possible outcomes  Probability is conveyed for a range of.
Lecture (16) Introduction to Stochastic Hydrology.
1 Chapter 12 Introduction to Statistics A random variable is one in which the exact behavior cannot be predicted, but which may be described in terms of.
Introduction l Example: Suppose we measure the current (I) and resistance (R) of a resistor. u Ohm's law relates V and I: V = IR u If we know the uncertainties.
© Copyright McGraw-Hill CHAPTER 6 The Normal Distribution.
QA in Finance/ Ch 3 Probability in Finance Probability.
Go to Index Analysis of Means Farrokh Alemi, Ph.D. Kashif Haqqi M.D.
Introduction to Data Analysis Probability Distributions.
Continuous Random Variables and Probability Distributions
PROBABILITY & STATISTICAL INFERENCE LECTURE 3 MSc in Computing (Data Analytics)
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Review and Preview This chapter combines the methods of descriptive statistics presented in.
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
Statistics for Data Miners: Part I (continued) S.T. Balke.
Statistical Review We will be working with two types of probability distributions: Discrete distributions –If the random variable of interest can take.
Topics Covered Discrete probability distributions –The Uniform Distribution –The Binomial Distribution –The Poisson Distribution Each is appropriately.
Theory of Probability Statistics for Business and Economics.
Barnett/Ziegler/Byleen Finite Mathematics 11e1 Learning Objectives for Section 11.5 Normal Distributions The student will be able to identify what is meant.
 A probability function is a function which assigns probabilities to the values of a random variable.  Individual probability values may be denoted by.
The Gaussian (Normal) Distribution: More Details & Some Applications.
Physics 270 – Experimental Physics. Standard Deviation of the Mean (Standard Error) When we report the average value of n measurements, the uncertainty.
R Kass/SP07 P416 Lecture 4 1 Propagation of Errors ( Chapter 3, Taylor ) Introduction Example: Suppose we measure the current (I) and resistance (R) of.
Lecture 15: Statistics and Their Distributions, Central Limit Theorem
Normal Distribution Introduction. Probability Density Functions.
 A probability function is a function which assigns probabilities to the values of a random variable.  Individual probability values may be denoted by.
880.P20 Winter 2006 Richard Kass Binomial Probability Distribution For the binomial distribution P is the probability of m successes out of N trials. Here.
Lecture 2 Review Probabilities Probability Distributions Normal probability distributions Sampling distributions and estimation.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
1 Since everything is a reflection of our minds, everything can be changed by our minds.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
EAS31116/B9036: Statistics in Earth & Atmospheric Sciences Lecture 3: Probability Distributions (cont’d) Instructor: Prof. Johnny Luo
1 Topic 5 - Joint distributions and the CLT Joint distributions –Calculation of probabilities, mean and variance –Expectations of functions based on joint.
Physics 270. o Experiment with one die o m – frequency of a given result (say rolling a 4) o m/n – relative frequency of this result o All throws are.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
Expectation. Let X denote a discrete random variable with probability function p(x) (probability density function f(x) if X is continuous) then the expected.
CPE 619 Comparing Systems Using Sample Data Aleksandar Milenković The LaCASA Laboratory Electrical and Computer Engineering Department The University of.
Chapter 7: Sampling Distributions Section 7.1 How Likely Are the Possible Values of a Statistic? The Sampling Distribution.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
Chap 5-1 Discrete and Continuous Probability Distributions.
THE NORMAL DISTRIBUTION
R. Kass/W04 P416 Lec 3 1 Lecture 3 The Gaussian Probability Distribution Function Plot of Gaussian pdf x p(x)p(x) Introduction l The Gaussian probability.
Theoretical distributions: the Normal distribution.
Sampling Distributions and Estimation
The Gaussian Probability Distribution Function
Econ 3790: Business and Economics Statistics
Statistics Lecture 12.
Lecture 2 Binomial and Poisson Probability Distributions
Sampling Distributions (§ )
Presentation transcript:

R. Kass/S07 P416 Lec 3 1 Lecture 3 The Gaussian Probability Distribution Function Plot of Gaussian pdf x p(x)p(x) Introduction l The Gaussian probability distribution is perhaps the most used distribution in all of science. Sometimes it is called the “bell shaped curve” or normal distribution. l Unlike the binomial and Poisson distribution, the Gaussian is a continuous distribution:  = mean of distribution (also at the same place as mode and median)  2 = variance of distribution y is a continuous variable (-∞  y  ∞  l Probability (P) of y being in the range [a, b] is given by an integral: u The integral for arbitrary a and b cannot be evaluated analytically. The value of the integral has to be looked up in a table (e.g. Appendixes A and B of Taylor). The integrals with limits [- ,  ] can be evaluated, see Barlow P. 37. Karl Friedrich Gauss

R. Kass/S07 P416 Lec 3 2 l The total area under the curve is normalized to one by the σ√(2π) factor. l We often talk about a measurement being a certain number of standard deviations (  ) away from the mean (  ) of the Gaussian. We can associate a probability for a measurement to be |  - n  | away from the mean just by calculating the area outside of this region. n  Prob. of exceeding  ±n  It is very unlikely (< 0.3%) that a measurement taken at random from a Gaussian pdf will be more than  from the true mean of the distribution. 95% of area within 2  Only 5% of area outside 2  Gaussian pdf with  =0 and  =1

R. Kass/S07 P416 Lec 3 3 For a binomial distribution: mean number of heads =  = Np = 5000 standard deviation  = [Np(1 - p)] 1/2 = 50 The probability to be within  ±1  for this binomial distribution is: For a Gaussian distribution: Both distributions give about the ~ same probability! See Taylor10.4 Relationship between Gaussian and Binomial & Poisson distribution The Gaussian distribution can be derived from the binomial (or Poisson) assuming: p is finite & N is very large we have a continuous variable rather than a discrete variable Consider tossing a coin 10,000 times. p(head) = 0.5 and N = 10,000 Compare ±1σ area of Poisson and Gaussian: Mean PoissionGaussian % diff

R. Kass/S07 P416 Lec 3 4 Why is the Gaussian pdf so applicable?  Central Limit Theorem A crude statement of the Central Limit Theorem: Things that are the result of the addition of lots of small effects tend to become Gaussian. A more exact statement: Let Y 1, Y 2,...Y n be an infinite sequence of independent random variables each with the same probability distribution. Suppose that the mean (  ) and variance (  2 ) of this distribution are both finite. For any numbers a and b: The C.L.T. tells us that under a wide range of circumstances the probability distribution that describes the sum of random variables tends towards a Gaussian distribution as the number of terms in the sum  ∞. Actually, the Y’s can be from different pdf’s! How close to  does n have to be?? Alternatively we can write the CLT in a different form:

R. Kass/S07 P416 Lec 3 5 —  m is sometimes called “the error in the mean” (more on that later):  For CLT to be valid: The  and  of the pdf must be finite. No one term in sum should dominate the sum. l A random variable is not the same as a random number. “A random variable is any rule that associates a number with each outcome in S” (Devore, in “probability and Statistics for Engineering and the Sciences”). Here S is the set of possible outcomes. l Recall if y is described by a Gaussian pdf with mean (  ) of zero and  =1 then the probability that a<y<b is given by: The CLT is true even if the Y’s are from different pdf’s as long as the means and variances are defined for each pdf ! See Appendix of Barlow for a proof of the Central Limit Theorem.

R. Kass/S07 P416 Lec 3 6 l Example: Generate a Gaussian distribution using uniform random numbers. u Random number generator gives numbers distributed uniformly in the interval [0,1] n  = 1/2 and  2 = 1/12 u Procedure: a) Take 12 numbers (r 1, r 2,……r 12 ) from your computer’s random number generator (ran(iseed)) b) Add them together c) Subtract 6  Get a number that looks as if it is from a Gaussian pdf! Thus the sum of 12 uniform random numbers minus 6 is distributed as if it came from a Gaussian pdf with  = 0 and  = 1. E A) 5000 random numbersB) 5000 pairs (r 1 + r 2 ) of random numbers C) 5000 triplets (r 1 + r 2 + r 3 ) of random numbers D) plets (r 1 + r 2 +…r 12 ) of random numbers. E) plets (r 1 + r 2 +…r ) of random numbers. Gaussian  = 0 and  = 1 12 is close to 

R. Kass/S07 P416 Lec 3 7  Example:A watch makes an error of at most ±1/2 minute per day. After one year, what’s the probability that the watch is accurate to within ±25 minutes? u Assume that the daily errors are uniform in [-1/2, 1/2]. n For each day, the average error is zero and the standard deviation 1/√12 minutes. n The error over the course of a year is just the addition of the daily error. n Since the daily errors come from a uniform distribution with a well defined mean and variance the Central Limit Theorem is applicable: The upper limit corresponds to +25 minutes: The lower limit corresponds to –25 minutes. The probability to be within  25 minutes is: The probability to be off by more than 25 minutes is just: 1-P  There is < 3 in a million chance that the watch will be off by more than 25 minutes in a year! This integral is ≈ 1 to about 3 part in 10 6 !

R. Kass/S07 P416 Lec 3 8 l Example: The daily income of a “card shark” has a uniform distribution in the interval [-$40,$50]. What is the probability that s/he wins more than $500 in 60 days? u Lets use the CLT to estimate this probability: u The probability distribution of daily income is uniform, p(y) = 1. p(y) needs to be normalized in computing the average daily winning (  ) and its standard deviation (  ). u The lower limit of the winnings is $500: u The upper limit is the maximum that the shark could win (50$/day for 60 days): 16% chance to win $500 in 60 days