Chapter 5 Joint Probability Distributions and Random Samples

Slides:



Advertisements
Similar presentations
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Advertisements

Statistics review of basic probability and statistics.
Review of Basic Probability and Statistics
Chapter 4 Discrete Random Variables and Probability Distributions
1 MF-852 Financial Econometrics Lecture 4 Probability Distributions and Intro. to Hypothesis Tests Roy J. Epstein Fall 2003.
Review.
Continuous Random Variables and Probability Distributions
Normal and Sampling Distributions A normal distribution is uniquely determined by its mean, , and variance,  2 The random variable Z = (X-  /  is.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
One Sample  M ean μ, Variance σ 2, Proportion π Two Samples  M eans, Variances, Proportions μ1 vs. μ2 σ12 vs. σ22 π1 vs. π Multiple.
Binomial and Related Distributions 學生 : 黃柏舜 學號 : 授課老師 : 蔡章仁.
Section 15.8 The Binomial Distribution. A binomial distribution is a discrete distribution defined by two parameters: The number of trials, n The probability.
Binomial Distributions Calculating the Probability of Success.
Mid-Term Review Final Review Statistical for Business (1)(2)
ENGR 610 Applied Statistics Fall Week 3 Marshall University CITE Jack Smith.
Bernoulli Trials Two Possible Outcomes –Success, with probability p –Failure, with probability q = 1  p Trials are independent.
CHAPTER Discrete Models  G eneral distributions  C lassical: Binomial, Poisson, etc Continuous Models  G eneral distributions 
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
1 Since everything is a reflection of our minds, everything can be changed by our minds.
Chapter 1 Overview and Descriptive Statistics 1111.1 - Populations, Samples and Processes 1111.2 - Pictorial and Tabular Methods in Descriptive.
Chapter 3 Discrete Random Variables and Probability Distributions  Random Variables.2 - Probability Distributions for Discrete Random Variables.3.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Probability Distributions, Discrete Random Variables
Continuous Random Variables and Probability Distributions
Chapter 5 Sampling Distributions. Introduction Distribution of a Sample Statistic: The probability distribution of a sample statistic obtained from a.
Chapter 5 Joint Probability Distributions and Random Samples  Jointly Distributed Random Variables.2 - Expected Values, Covariance, and Correlation.3.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Chapter 5 Joint Probability Distributions and Random Samples  Jointly Distributed Random Variables.2 - Expected Values, Covariance, and Correlation.3.
Sums of Random Variables and Long-Term Averages Sums of R.V. ‘s S n = X 1 + X X n of course.
Binomial Distribution
Sampling and Sampling Distributions
Chapter 5 Joint Probability Distributions and Random Samples
Chapter 5 Confidence Interval
Chapter 3 Discrete Random Variables and Probability Distributions
Supplemental Lecture Notes
Discrete Probability Distributions
ASV Chapters 1 - Sample Spaces and Probabilities
STAT 311 Chapter 1 - Overview and Descriptive Statistics
STAT 312 Chapter 7 - Statistical Intervals Based on a Single Sample
STAT 311 REVIEW (Quick & Dirty)
Hypergeometric Distribution
Discrete random variable X Examples: shoe size, dosage (mg), # cells,…
Samples and Populations
Sample Mean Distributions
Chapter 5 Sampling Distributions
Chapter 3 Discrete Random Variables and Probability Distributions
Parameter, Statistic and Random Samples
Combining Random Variables
Chapter 5 Sampling Distributions
Chapter 4 Continuous Random Variables and Probability Distributions
Determining the distribution of Sample statistics
Chapter 3 Discrete Random Variables and Probability Distributions
ASV Chapters 1 - Sample Spaces and Probabilities
Sampling Distributions and the Central Limit Theorem
The Binomial and Geometric Distributions
Discrete random variable X Examples: shoe size, dosage (mg), # cells,…
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
Chapter 5 Sampling Distributions
Lecture Slides Elementary Statistics Twelfth Edition
Chapter 3 Discrete Random Variables and Probability Distributions
Random Variable X, with pmf p(x) or pdf f(x)
Binomial Distribution
Introduction to Statistics
CHAPTER 6 Statistical Inference & Hypothesis Testing
Chapter 3 Discrete Random Variables and Probability Distributions
Chapter 7: The Distribution of Sample Means
ASV Chapters 1 - Sample Spaces and Probabilities
Bernoulli Trials Two Possible Outcomes Trials are independent.
Mathematical Foundations of BME Reza Shadmehr
1/2555 สมศักดิ์ ศิวดำรงพงศ์
Presentation transcript:

Chapter 5 Joint Probability Distributions and Random Samples 5.1 - Jointly Distributed Random Variables 5.2 - Expected Values, Covariance, and Correlation 5.3 - Statistics and Their Distributions 5.4 - The Distribution of the Sample Mean 5.5 - The Distribution of a Linear Combination

Population Distribution of X Suppose X ~ N(μ, σ), then… X = Age of women in U.S. at first birth X Density Each of these individual ages x is a particular value of the random variable X. Most are in the neighborhood of μ, but there are occasional outliers in the tails of the distribution.     x4 x5 x1 x2  x3 … etc…. σ = 1.5 x x x x x μ = 25.4

Population Distribution of X Suppose X ~ N(μ, σ), then… Sample, n = 400 Sample, n = 400 X = Age of women in U.S. at first birth Sample, n = 400 Sample, n = 400 X Density Sample, n = 400 How are these values distributed? … etc…. Each of these sample mean values is a “point estimate” of the population mean μ… σ = 1.5 μ = 25.4

Population Distribution of X Sampling Distribution of for any sample size n. Suppose X ~ N(μ, σ), then… Suppose X ~ N(μ, σ), then… X = Age of women in U.S. at first birth Density μ = X Density μ = σ = 1.5 “standard error” The vast majority of sample means are extremely close to μ, i.e., extremely small variability. How are these values distributed? … etc…. Each of these sample mean values is a “point estimate” of the population mean μ… μ = 25.4

Population Distribution of X Sampling Distribution of Suppose X ~ N(μ, σ), then… Suppose X ~ N(μ, σ), then…   for large sample size n. for any sample size n. X = Age of women in U.S. at first birth Density μ = X Density μ = σ = 2.4 Suppose X ~ N(μ, σ), then… “standard error” The vast majority of sample means are extremely close to μ, i.e., extremely small variability. … etc…. Each of these sample mean values is a “point estimate” of the population mean μ… μ = 25.4

Population Distribution of X Sampling Distribution of X ~ Anything with finite μ and σ Suppose X  N(μ, σ), then…  for large sample size n. for any sample size n. X = Age of women in U.S. at first birth Density μ = X Density μ = σ = 2.4 Suppose X ~ N(μ, σ), then… “standard error” The vast majority of sample means are extremely close to μ, i.e., extremely small variability. … etc…. Each of these sample mean values is a “point estimate” of the population mean μ… μ = 25.4

Density   Density  “standard error”

Probability that a single house selected at random costs less than $300K = ? = Cumulative area under density curve for X up to 300. = Z-score Density   Density  “standard error” Example: X = Cost of new house ($K)  300

Probability that a single house selected at random costs less than $300K = ? 0.6554 = Z-score Density   Density  “standard error” Example: X = Cost of new house ($K)  300

Probability that a single house selected at random costs less than $300K = ? 0.6554 = Z-score Probability that the sample mean of n = 36 houses selected at random is less than $300K = ? = Cumulative area under density curve for up to 300. Density   Density  “standard error” Example: X = Cost of new house ($K) $12.5K  300 300

Probability that a single house selected at random costs less than $300K = ? 0.6554 = Z-score Probability that the sample mean of n = 36 houses selected at random is less than $300K = ? 0.9918 = Z-score Density   Density  “standard error” Example: X = Cost of new house ($K) $12.5K  300 300

     large Density Density “standard error” mild skew approximately  large Density  Density “standard error” mild skew   

~ CENTRAL LIMIT THEOREM ~ approximately continuous or discrete,  as n  , large   Density  Density “standard error”  

~ CENTRAL LIMIT THEOREM ~ continuous or discrete, approximately  as n  , large   Density  Example: X = Cost of new house ($K) Density “standard error”  

Probability that a single house selected at random costs less than $300K = ? = Cumulative area under density curve for X up to 300. Probability that the sample mean of n = 36 houses selected at random is less than $300K = ? 0.9918 = Z-score Density  “standard error” Example: X = Cost of new house ($K) Density $12.5K  300 300 

x p(x) 0.5 10 0.3 20 0.2

.25 5 .30 =.15+.15 10 .29 =.10+.09+.10 15 .12 =.06+.06 20 .04

possibly log-normal… More on CLT… but remember Cauchy and 1/x2, both of which had nonexistent … CLT may not work! More on CLT… heavily skewed tail each based on 1000 samples

Population Distribution of X Density Population Distribution of X X ~ Dist(μ, σ) Random Variable More on CLT… X = Age of women in U.S. at first birth  If this first individual has been randomly chosen, and the value of X measured, then the result is a fixed number x1, with no random variability… and likewise for x2, x3, etc. DATA! BUT…

Population Distribution of X Density Population Distribution of X X ~ Dist(μ, σ) Random Variable More… X = Age of women in U.S. at first birth  If this first individual has been randomly chosen, and the value of X measured, then the result is a fixed number x1, with no random variability… and likewise for x2, x3, etc. DATA! However, if this is not the case, then this first “value” of X is unknown, thus can be considered as a random variable X1 itself… and likewise for X2, X3, etc. BUT… The collection {X1, X2, X3, …, Xn} of “independent, identically-distributed” (i.i.d.) random variables is said to be a random sample.

Population Distribution of X Density Population Distribution of X X ~ Dist(μ, σ) Random Variable Sample, size n More… X = Age of women in U.S. at first birth CENTRAL LIMIT THEOREM Sampling Distribution of etc…… Density Claim: for any n Proof:

Population Distribution of X Density Population Distribution of X X ~ Dist(μ, σ) Random Variable More… X = Age of women in U.S. at first birth CENTRAL LIMIT THEOREM Sampling Distribution of etc…… Density Claim: for any n Proof:

Recall… More on CLT… Normal Approximation to the Binomial Distribution continuous discrete Recall… More on CLT… Normal Approximation to the Binomial Distribution Suppose a certain outcome exists in a population, with constant probability . We will randomly select a random sample of n individuals, so that the binary “Success vs. Failure” outcome of any individual is independent of the binary outcome of any other individual, i.e., n Bernoulli trials (e.g., coin tosses). Then X is said to follow a Binomial distribution, written X ~ Bin(n, ), with “probability function” p(x) = , x = 0, 1, 2, …, n. P(Success) =  P(Failure) = 1 –  Discrete random variable X = # Successes (0, 1, 2,…, n) in a random sample of size n

Normal Approximation to the Binomial Distribution CLT continuous discrete Normal Approximation to the Binomial Distribution Suppose a certain outcome exists in a population, with constant probability . We will randomly select a random sample of n individuals, so that the binary “Success vs. Failure” outcome of any individual is independent of the binary outcome of any other individual, i.e., n Bernoulli trials (e.g., coin tosses). Then X is said to follow a Binomial distribution, written X ~ Bin(n, ), with “probability function” p(x) = , x = 0, 1, 2, …, n. P(Success) =  P(Failure) = 1 –  CLT See Prob 5.3/7 Discrete random variable X = # Successes (0, 1, 2,…, n) in a random sample of size n

Normal Approximation to the Binomial Distribution CLT continuous discrete ?? Normal Approximation to the Binomial Distribution Suppose a certain outcome exists in a population, with constant probability . We will randomly select a random sample of n individuals, so that the binary “Success vs. Failure” outcome of any individual is independent of the binary outcome of any other individual, i.e., n Bernoulli trials (e.g., coin tosses). Then X is said to follow a Binomial distribution, written X ~ Bin(n, ), with “probability function” p(x) = , x = 0, 1, 2, …, n. P(Success) =  P(Failure) = 1 –  CLT Discrete random variable X = # Successes (0, 1, 2,…, n) in a random sample of size n