Chapter 6. Point Estimation Weiqi Luo ( 骆伟祺 ) School of Software Sun Yat-Sen University : Office : # A313

Slides:



Advertisements
Similar presentations
Estimation of Means and Proportions
Advertisements

Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 6 Point Estimation.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 7. Statistical Estimation and Sampling Distributions
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Statistical Estimation and Sampling Distributions
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Chap 8-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 8 Estimation: Single Population Statistics for Business and Economics.
SOLVED EXAMPLES.
Copyright © Cengage Learning. All rights reserved.
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
Evaluating Hypotheses
Chapter 8 Estimation: Single Population
Chapter 7 Estimation: Single Population
1 STATISTICAL INFERENCE PART I EXPONENTIAL FAMILY & POINT ESTIMATION.
Copyright © Cengage Learning. All rights reserved. 6 Point Estimation.
Inferences About Process Quality
Maximum likelihood (ML)
Business Statistics: Communicating with Numbers
Chapter 7 Estimation: Single Population
Estimation Basic Concepts & Estimation of Proportions
Lecture 14 Sections 7.1 – 7.2 Objectives:
QBM117 Business Statistics Estimating the population mean , when the population variance  2, is known.
1 Introduction to Estimation Chapter Concepts of Estimation The objective of estimation is to determine the value of a population parameter on the.
STATISTICAL INFERENCE PART I POINT ESTIMATION
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Chapter 7. Statistical Intervals Based on a Single Sample Weiqi Luo ( 骆伟祺 ) School of Software Sun Yat-Sen University : Office.
Random Sampling, Point Estimation and Maximum Likelihood.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Chapter 7 Point Estimation
Chapter 5.6 From DeGroot & Schervish. Uniform Distribution.
1 Lecture 16: Point Estimation Concepts and Methods Devore, Ch
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
CLASS: B.Sc.II PAPER-I ELEMENTRY INFERENCE. TESTING OF HYPOTHESIS.
Chapter 7. Statistical Intervals Based on a Single Sample Weiqi Luo ( 骆伟祺 ) School of Software Sun Yat-Sen University : Office.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
CHAPTER 9 Inference: Estimation The essential nature of inferential statistics, as verses descriptive statistics is one of knowledge. In descriptive statistics,
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Copyright © Cengage Learning. All rights reserved. 5 Joint Probability Distributions and Random Samples.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Chapter 5 Special Distributions Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Statistics for Business and Economics 8 th Edition Chapter 7 Estimation: Single Population Copyright © 2013 Pearson Education, Inc. Publishing as Prentice.
Conditional Expectation
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
LEARNING OBJECTIVES After careful study of this chapter you should be able to do the following: 1.Explain the general concepts of estimating the parameters.
Copyright © Cengage Learning. All rights reserved. 16 Quality Control Methods.
Copyright © Cengage Learning. All rights reserved. 6 Point Estimation.
Virtual University of Pakistan
Copyright © Cengage Learning. All rights reserved.
Points and Interval Estimates
STATISTICS POINT ESTIMATION
STATISTICAL INFERENCE
12. Principles of Parameter Estimation
6 Point Estimation.
Some General Concepts of Point Estimation
Random Sampling Population Random sample: Statistics Point estimate
CONCEPTS OF ESTIMATION
6 Point Estimation.
12. Principles of Parameter Estimation
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Applied Statistics and Probability for Engineers
Presentation transcript:

Chapter 6. Point Estimation Weiqi Luo ( 骆伟祺 ) School of Software Sun Yat-Sen University : Office : # A313

School of Software  6.1. Some General Concepts of Point Estimation  6.2. Methods of Point Estimation 2 Chapter 6: Point Estimation

School of Software  In ordert to get some population characteristics, statistical inference needs obtain sample data from the population under study, and achieve the conclusions can then be based on the computed values of various sample quantities (statistics).  Typically, we will use the Greek letter θ for the parameter of interest. The objective of point estimation is to select a single number, based on sample data (statistic ), that represents a sensible value for θ. 6.1 Some General Concepts of Point Estimation 3

School of Software  Point Estimation A point estimate of a parameter θ is a single number that can be regarded as a sensible value for θ. A point estimate is obtained by selecting a suitable statistic and computing its value from the given sample data. The selected statistic is called the point estimator of θ. 6.1 Some General Concepts of Point Estimation 4 Population Parameters θ A Sample A quantity A function (statistic ) Q #1: How to get the candiate estimators based on the population? Q #2: How to measure the candidate estimators? Here, the type of population under study is usually known, while the paprameters are unkown. Estimating

School of Software  Example 6.1 The manufacturer has used this bumper in a sequence of 25 controlled crashes against a wall, each at 10 mph, using one of its compact car models. Let X = the number of crashes that result in no visible damage to the automobile. What is a sensible estimate of the parameter p = the proportion of all such crashes that result in no damage If X is observed to be x = 15, the most reasonable estimator and estimate are 6.1 Some General Concepts of Point Estimation 5 estimator estimate =

School of Software  Example 6.2 Reconsider the accompanying 20 observations on dielectric breakdown voltage for pieces of epoxy resin first introduced in Example 4.29 (pp. 193) The pattern in the normal probability plot given there is quite straight, so we now assume that the distribution of breakdown voltage is normal with mean value μ. Because normal distribution are symmetric, μ is also the median lifetime of the distribution. The given observation are then assumed to be the result of a random sample X 1, X 2, …, X 20 from this normal distribution. 6.1 Some General Concepts of Point Estimation

School of Software  Example 6.2 (Cont’) Consider the following estimators and resulting estimates for μ 6.1 Some General Concepts of Point Estimation 7 a. Estimator=, estimate= b. Estimator=, estimate= c. Estimator [ min(X i ) + max(X j )]/2 = the average of the two extreme lifetimes, estimate=[ min(x i )+max(x i )]/2 = ( )/2 = d. Estimator =, the 10% trimmed mean (discard the smallest and largest 10% of the sample and then average)

School of Software  Example 6.3 In the near future there will be increasing interest in developing low-cost Mg- based alloys for various casting processes. It is therefore important to have practical ways of determining various mechanical properties of such alloys. Assume that the observations of a random sample X 1, X 2, …, X 8 from the population distribution of elastic modulus under such circumstances. We want to estimate the population variance σ Some General Concepts of Point Estimation 8 Method #1: sample variance Method #2: Divided by n rather than n-1

School of Software  Estimation Error Analysis Note that is a function of the sample X i ’s, so it is a random variable. Therefore, a n accurate estimator would be one resulting in small estimation errors, so that estimated values will be near the true value θ (unkown). A good estimator should have the two properties: 1. unbiasedness (i.e. the average error should be zero) 2. minimum variance (i.e. the variance of error should be samll) 6.1 Some General Concepts of Point Estimation 9 + error of estimation

School of Software  Unbiased Estimator A point estimator is said to be an unbiased estimator of θ if for every possible value of θ. If is not unbiased, the difference is called the bias of 6.1 Some General Concepts of Point Estimation 10 pdf of Bias of θ 1 θ Note: “centered” here means the expected vaule, not the median, of the the distribution of is equal to θ

School of Software  Proposition When X is a binomial rv with parameters n and p, the sample proportion =X/n is an unbiased estimator of p. Refer to Example 6.1, the sample proportion X/n was used as an estimator of p, where X, the number of sample successes, had a binomial distribution with parameters n and p, thus Some General Concepts of Point Estimation

School of Software  Example 6.4 Suppose that X, the reaction time to a certain stimulus, has a uniform distribution on the interval from 0 to an unknown upper limit θ. It is desired to estimate θ on the basis of a random sample X 1, X 2, …, X n of reaction times. Since θ is the largest possible time in the entire population of reaction times, consider as a first estimator the largest sample reaction time: Since (refer to Ex. 32 in pp. 279 ) Another estimator 6.1 Some General Concepts of Point Estimation 12 biased estimator, why? unbiased estimator

School of Software  Proposition Let X 1, X 2, …, X n be a random sample from a distribution with mean μ and variance σ 2. Then the estimator is an unbiased estimator of σ 2, namely Refer to pp. 259 for the proof. However, 6.1 Some General Concepts of Point Estimation 13

School of Software  Proposition If X 1, X 2,…X n is a random sample from a distribution with mean μ, then is an unbiased estimator of μ. If in addition the distribution is continuous and symmetric, then and any trimmed mean are also unbiased estimator of μ 6.1 Some General Concepts of Point Estimation 14 Refer to the estimators in Example 6.2

School of Software  Estimators with Minimum Variance Some General Concepts of Point Estimation pdf of, another unbiased estimator pdf of, an unbiased estimator θ Obivously, the estimator is better than the in this example

School of Software  Example 6.5 (Ex. 6.4 Cont’) When X 1, X 2, … X n is a random sample from a uniform distribution on [0, θ], the estimator is unbiased for θ It is also shown that is the MVUE of θ Some General Concepts of Point Estimation

School of Software  Theorem Let X 1, X 2, …, X n be a random sample from a normal distribution with parameters μ and δ. Then the estimator is the MVUE for μ. 6.1 Some General Concepts of Point Estimation 17 How about those un-normal distributions?

School of Software  Estimator Selection  When choosing among several different estimators of θ, select one that is unbiased.  Among all estimators of θ that are unbiased, choose the one that has minimum variance. The resulting is called the minimum variance unbiased estimator (MVUE) of θ Some General Concepts of Point Estimation pdf of, the MVUE pdf of, a biased estimator θ In some cases, a biased estimator is perferable to the MVUE

School of Software  Example 6.6 Suppose we wish to estimate the thermal conductivity μ of a certain material. We will obtain a random sample X 1, X 2, …, X n of n thermal conductivity measurements. Let’s assume that the population distribution is a member of one of the following three families: 6.1 Some General Concepts of Point Estimation 19 Gaussian Distribution Cauchy Distribution Uniform Distribution

School of Software Some General Concepts of Point Estimation 1. If the random sample comes from a normal distribution, then is the best of the four estimators, since it is the MVUE. 2. If the random sample comes from a Cauchy distribution, then and (the average of the two extreme observations) are terrible estimators for μ, whereas is quite good; is bad because it is very sensitive to outlying observations, and the heavy tails of the Cauchy distribution make a few such observation likely to appear in any sample. 3. If the underlying distribution is uniform, the best estimator is ; this estimator is greatly influenced by outlying observations, but the lack of tails makes such observations impossible. 4. The trimmed mean is best in none of these three situations, but works reasonably well in all three. That is, does not suffer too much in any of the three situations. A Robust estimator

School of Software  The Standard Error The standard error of an estimator is its standard deviation. If the standard error itself involves unknown parameters whose values can be estimated, substitution of these estimates into yields the estimated standard error (estimated standard deviation) of the estimator. The estimated standard error can be denoted either by or by Some General Concepts of Point Estimation

School of Software  Example 6.8 Assuming that breakdown voltage is normally distributed, is the best estimator of μ. If the value of σ is known to be 1.5, the standard error of is If, as is usually the case, the value of σ is unknown, the estimate is substituted into to obtain the estimated standard error Some General Concepts of Point Estimation

School of Software  Homework Ex. 1, Ex. 8, Ex. 9, Ex Some General Concepts of Point Estimation

School of Software  Two “constructive” methods for obtaining point estimators  Method of Moments  Maximum Likehood Estimation 6.2 Methods of Point Estimation 24

School of Software  Moments Let X 1, X 2,…, X n be a random sample from a pmf or pdf f(x). For k = 1, 2, 3, …, the kth population moment, or kth moment of the distribution f(x), is. The kth sample moment is. 6.2 Methods of Point Estimation 25

School of Software  Moment Estimator Let X 1, X 2, …, X n be a random sample from a distribution with pmf or pdf f(x;θ 1,…,θ m ), where θ 1,…,θ m are parameters whose values are unknown. Then the moment estimators are obtained by equating the first m sample moments to the corresponding first m population moments and solving for θ 1,…,θ m. 6.2 Methods of Point Estimation 26 K-th sample moment K-th population moment n is large With unkonwn θ i With the given sample x i

School of Software 6.2 Methods of Point Estimation 27 The first m population moments The solution of equations Use the first m sample moment to represent the population moments μ l General Algorithm :

School of Software  Example 6.11 Let X 1, X 2, …, X n represent a random sample of service times of n customers at a certain facility, where the underlying distribution is assumed exponential with parameter λ. How to estimate λ by using the method of moments? Step #1: The 1 st population moment E(X) = 1/λ then we have λ = 1/ E(X) Step #2: Use the 1 st sample moment to represent 1 st poulation moment E(X), and get the estimator 6.2 Methods of Point Estimation 28

School of Software  Example 6.12 Let X 1, …, X n be a random sample from a gamma distribution with parameters α and β. Its pdf is There are two parameter need to be estimated, thus, consider the first two monents 6.2 Methods of Point Estimation 29

School of Software  Example 6.12 (Cont’) 6.2 Methods of Point Estimation 30 Step #1: Step #2:

School of Software  Example 6.13 Let X 1, …, X n be a random sample from a generalized negative binomial distribution with parameters r and p. Its pmf is Determine the moment estimators of parameters r and p. Note: There are two parameters needs to estimate, thus the first two moments are considered. 6.2 Methods of Point Estimation 31

School of Software  Example 6.13 (Cont’) 6.2 Methods of Point Estimation 32 Step #2: Step #1:

School of Software  Maximum Likelihood Estimation (Basic Idea) 6.2 Methods of Point Estimation 33 Box 1Box 2 Experiment: We firstly randomly choose a box, And then randomly choose a ball. Q: If we get a white ball, which box has the Maximum Likelihood being chosen?

School of Software  Maximum Likelihood Estimation (Basic Idea) 6.2 Methods of Point Estimation 34 Q: What is the probability p of hitting the target? … The best one among the four options

School of Software  Example 6.14 A sample of ten new bike helmets manufactured by a certain company is obtained. Upon testing, it is found that the first, third, and tenth helmets are flawed, whereas the others are not. Let p = P(flawed helmet) and define X 1, …, X 10 by X i = 1 if the ith helmet is flawed and zero otherwise. Then the observed x i ’s are 1,0,1,0,0,0,0,0,0, Methods of Point Estimation 35 The Joint pmf of the sample is For what value of p is the observed sample most likely to have occurred? Or, equivalently, what value of the parameter p should be taken so that the joint pmf of the sample is maximized?

School of Software  Example 6.14 (Cont’) 6.2 Methods of Point Estimation 36 Equating the derivative of the logarithm of the pmf to zero gives the maximizing value (why?) where x is the observed number of successes (flawed helmets). The estimate of p is now. It is called the maximum likelihood estimate because for fixed x 1,…, x 10, it is the parameter value that maximizes the likelihood of the observed sample.

School of Software  Maximum Likelihood Estimation Let X 1, X 2, …, X n have joint pmf or pdf where the parameters θ 1, …, θ m have unknown values. When x 1, …, x n are the observed sample values and f is regarded as a function of θ 1, …, θ m, it is called the likelihood function. The maximum likelihood estimates(mle’s) are those values of the θ i ’s that maximize the likelihood function, so that When the X i ’s are substituted in place of the x i ’s, the maximum likelihood estimators result. 6.2 Methods of Point Estimation 37 for all θ 1, …, θ m

School of Software  Example 6.15 Suppose X 1, X 2, …, X n is a random sample from an exponential distribution with the unknown parameter λ. Determine the maximum likelihood estimator of λ. 6.2 Methods of Point Estimation 38 The joint pdf is (independence) The ln(likelihood) is Equating to zero the derivative w.r.t. λ: The estimator

School of Software  Example 6.16 Let X 1, X 2, …, X n is a random sample from a normal distribution N(μ,δ 2 ). Determine the maximum likelihood estimator of μ and δ 2. The joint pdf is 6.2 Methods of Point Estimation 39 Equating to 0 the partial derivatives w.r.t. μ and σ 2, finally we have Here the mle of δ 2 is not the unbiased estimator.

School of Software  Three steps 1.Write the joint pmf/pdf (i.e. Likelihood function) 2.Get the ln(likelihood) (if necessary) 3.Take the partial derivative of ln(f) with respect to θ i, equal them to 0, and solve the resulting m equations. 6.2 Methods of Point Estimation 40

School of Software  Estimating Function of Parameters The Invariance Principle Let be the mle’s of the parameters θ 1, …, θ m. Then the mle of any function h(θ 1,…,θ m ) of these parameters is the function of the mle’s. 6.2 Methods of Point Estimation 41

School of Software  Example 6.19 (Ex.6.16 Cont’) In the normal case, the mle’s of μ and σ 2 are and To obtain the mle of the function substitute the mle’s into the function 6.2 Methods of Point Estimation 42

School of Software  Large Sample Behavior of the MLE Under very general conditions on the joint distribution of the sample, when the sample size n is large, the maximum likelihood estimator of any parameter θ is approximately unbiased and has variance that is nearly as small as can be achieved by any estimator. Stated another way, the mle is approximately the MVUE of θ.  Maximum likelihood estimators are generally preferable to moment estimators because of the above efficiency properties.  However, the mle’s often require significantly more computation than do moment estimators. Also, they require that the underlying distribution be specified. 6.2 Methods of Point Estimation 43

School of Software  Example 6.21 Suppose my waiting time for a bus is uniformly distributed on [0,θ] and the results x 1, …, x n of a random sample from this distribution have been observed. Since f(x;θ) = 1/θ for 0 ≤ x ≤ θ and 0 otherwise, As long as max(x i ) ≤ θ, the likelihood is 1/θ n, which is positive, but as soon asθ <max(x i ), the likelihood drops to 0. Calculus will not work because the maximum of the likelihood occurs at a point of discontinuity. 6.2 Methods of Point Estimation 44

School of Software  Example 6.21 (Cont’) 6.2 Methods of Point Estimation 45 Likelihood max(x i ) θ the figure shows that. Thus, if my waiting times are 2.3, 3.7, 1.5, 0.4, and 3.2, then the mle is. the maximum of the likelihood

School of Software  Homework Ex. 20, Ex. 21, Ex. 29, Ex Methods of Point Estimation 46