Virtual University of Pakistan Lecture No. 34 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah.

Slides:



Advertisements
Similar presentations
Estimation of Means and Proportions
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Statistical Estimation and Sampling Distributions
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
The General Linear Model. The Simple Linear Model Linear Regression.
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
Maximum likelihood (ML) and likelihood ratio (LR) test
Point estimation, interval estimation
Topic 7 Sampling And Sampling Distributions. The term Population represents everything we want to study, bearing in mind that the population is ever changing.
Copyright © Cengage Learning. All rights reserved. 6 Point Estimation.
Maximum likelihood (ML)
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Chapter 6 The Normal Probability Distribution
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Business Statistics: Communicating with Numbers
1 Introduction to Estimation Chapter Concepts of Estimation The objective of estimation is to determine the value of a population parameter on the.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 10 Introduction to Estimation.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Copyright © 2009 Cengage Learning Chapter 10 Introduction to Estimation ( 추 정 )
PROBABILITY (6MTCOAE205) Chapter 6 Estimation. Confidence Intervals Contents of this chapter: Confidence Intervals for the Population Mean, μ when Population.
Random Sampling, Point Estimation and Maximum Likelihood.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Statistical Methods Introduction to Estimation noha hussein elkhidir16/04/35.
Virtual University of Pakistan Lecture No. 30 Statistics and Probability Miss Saleha Naghmi Habibullah.
Chapter 10 Introduction to Estimation Sir Naseer Shahzada.
1 Lecture 16: Point Estimation Concepts and Methods Devore, Ch
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
CLASS: B.Sc.II PAPER-I ELEMENTRY INFERENCE. TESTING OF HYPOTHESIS.
Virtual University of Pakistan Lecture No. 44 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
Sampling Design and Analysis MTH 494 Lecture-21 Ossam Chohan Assistant Professor CIIT Abbottabad.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
CHAPTER – 1 UNCERTAINTIES IN MEASUREMENTS. 1.3 PARENT AND SAMPLE DISTRIBUTIONS  If we make a measurement x i in of a quantity x, we expect our observation.
Virtual University of Pakistan Lecture No. 26 Statistics and Probability Miss Saleha Naghmi Habibullah.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
CHAPTER 4 ESTIMATES OF MEAN AND ERRORS. 4.1 METHOD OF LEAST SQUARES I n Chapter 2 we defined the mean  of the parent distribution and noted that the.
1 Ka-fu Wong University of Hong Kong A Brief Review of Probability, Statistics, and Regression for Forecasting.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Probability & Statistics Review I 1. Normal Distribution 2. Sampling Distribution 3. Inference - Confidence Interval.
Virtual University of Pakistan Lecture No. 35 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Virtual University of Pakistan
Virtual University of Pakistan
Virtual University of Pakistan
Virtual University of Pakistan
Virtual University of Pakistan
Copyright © Cengage Learning. All rights reserved.
ESTIMATION.
Virtual University of Pakistan
STATISTICS POINT ESTIMATION
12. Principles of Parameter Estimation
Virtual University of Pakistan
Virtual University of Pakistan
CONCEPTS OF ESTIMATION
Geology Geomath Chapter 7 - Statistics tom.h.wilson
Simple Linear Regression
Virtual University of Pakistan
12. Principles of Parameter Estimation
Presentation transcript:

Virtual University of Pakistan Lecture No. 34 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah

IN THE LAST LECTURE, YOU LEARNT  Sampling Distribution of (continued) Sampling Distribution of  Point Estimation  Desirable Qualities of a Good Point Estimator –Unbiasedness –Consistency

TOPICS FOR TODAY  Desirable Qualities of a Good Point Estimator:  Efficiency Methods of Point Estimation:  The Method of Moments  The Method of Least Squares  The Method of Maximum Likelihood Interval Estimation: Confidence Interval for 

The students will recall that, in the last lecture, we presented the basic definition of a point estimator, and considered some desirable qualities of a good point estimator.

Three main qualities of a good point estimator are:

DESIRABLE QUALITIES OF A GOOD POINT ESTIMATOR unbiasedness consistency efficiency

In the last lecture, we discussed in some detail the concepts of unbiasedness and consistency.

Why is consistency considered a desirable quality? In order to obtain an idea regarding the answer to this question, consider the following:

As a sample is only a part of the population, it is obvious that the larger the sample size, the more representative we expect it to be of the population from which it has been drawn.

In agreement with the above argument, we will expect our estimator to be close to the corresponding parameter if the sample size is large. Hence, we will naturally be happy if the probability of our estimator being close to the parameter increases with an increase in the sample size. As such, consistency is a desirable property.

Another important desirable quality of a good point estimator is EFFICIENCY :

EFFICIENCY An unbiased estimator is defined to be efficient if the variance of its sampling distribution is smaller than that of the sampling distribution of any other unbiased estimator of the same parameter.

Sampling Distribution of T 1 Sampling Distribution of T 2 In the following diagram, since Var (T 1 ) < Var (T 2 ), hence T 1 is more efficient than T 2 :

And, if we multiply the above expression by 100, we obtain the relative efficiency in percentage form.

It thus provides a criterion for comparing different unbiased estimators of a parameter.

Both the sample mean and the sample median for a population that has a normal distribution, are unbiased and consistent estimators of  but the variance of the sampling distribution of sample means is smaller than the variance of the sampling distribution of sample medians.

Hence, the sample mean is more efficient than the sample median as an estimator of . The sample mean may therefore be preferred as an estimator.

Next, we consider various methods of point estimation. A point estimator of a parameter can be obtained by several methods. We shall be presenting a brief account of the following three methods:

METHODS OF POINT ESTIMATION  The Method of Moments  The Method of Least Squares  The Method of Maximum Likelihood

These methods give estimates which may differ as the methods are based on different theories of estimation. We begin with the Method of Moments :

THE METHOD OF MOMENTS The method of moments which is due to Karl Pearson ( ), consists of calculating a few moments of the sample values and equating them to the corresponding moments of a population, thus getting as many equations as are needed to solve for the unknown parameters.

The procedure is described below:

Let X 1, X 2, …, X n be a random sample of size n from a population. Then the rth sample moment about zero is and the corresponding rth population moment is

We then match these moments and get as many equations as we need to solve for the unknown parameters.

The following examples illustrate the method:

EXAMPLE-1 Let X be uniformly distributed on the interval (0,  ). Find an estimator of  by the method of moments.

SOLUTION The probability density function of the given uniform distribution is

Since the uniform distribution has only one parameter, (i.e.  ), therefore, in order to find the maximum likelihood estimator of  by the method of moments, we need to consider only one equation.

The first sample moment about zero is And, the first population moment about zero is

Hence, the moment estimator of  is equal to i.e. In other words, the moment estimator of  is just twice the sample mean. Matching these moments, we obtain:

An Interesting Observation: It should be noted that, for the above uniform distribution, the mean is given by (This is so due to the absolute symmetry of the uniform distribution around the value

Now, implies that In other words, if we wish to have the exact value of , all we need to do is to multiply the population mean  by 2.

Generally, it is not possible to determine , and all we can do is to draw a sample from the probability distribution, and compute the sample mean  X. Hence, naturally, the equation will be replaced by the equation (Asprovides an estimate of , hence a ‘hat’ is placed on top of .)

It is interesting to notethat is exactly the same quantity as what we obtained as an estimate of  by the method of moments ! (The result obtained by the method of moments coincides with what we obtain through simple logic.)

EXAMPLE-2 Let X 1, X 2, …, X n be a random sample of size n from a normal population with parameters  and  2. Find these parameters by the method of moments.

SOLUTION Here we need two equations as there are two unknown parameters,  and  2.

The first two sample moments about zero are

The corresponding two moments of a normal distribution are  1 =  and  2 =  2 +  2. (  2 =  2 –  1 2 =  2 –  2 )

To get the desired estimators by the method of moments, we match them. Thus, we have :

A shortcoming of this method is that the moment estimators are, in general, inefficient.

Next, we consider the Method of Least Squares :

The method of Least Squares, which is due to Gauss ( ) and Markov ( ), is based on the theory of linear estimation. It is regarded as one of the important methods of point estimation.

THE METHOD OF LEAST SQUARES An estimator found by minimizing the sum of squared deviations of the sample values from some function that has been hypothesized as a fit for the data, is called the least squares estimator.

The method of least-squares has already been discussed in connection with regression analysis that was presented in Lecture No. 15. The students will recall that, when fitting a straight line y = a+bx to real data, ‘a’ and ‘b’ were determined by minimizing the sum of squared deviations between the fitted line and the data-points.

The y-intercept and the slope of the fitted line i.e. ‘a’ and ‘b’ are least-square estimates (respectively) of the y-intercept and the slope of the TRUE line that would have been obtained by considering the entire population of data- points, and not just a sample.

Next, we consider the Method of Maximum Likelihood :

The method of maximum likelihood is regarded as the MOST important method of estimation, and is the most widely used method. This method was introduced in 1922 by Sir Ronald A. Fisher ( ).

The mathematical technique of finding Maximum Likelihood Estimators is a bit advanced, and involves the concept of the Likelihood Function. In this course, we consider only the overall logic of this method:

RATIONALE OF THE METHOD OF MAXIMUM LIKELIHOOD (ML) “To consider every possible value that the parameter might have, and for each value, compute the probability that the given sample would have occurred if that were the true value of the parameter. That value of the parameter for which the probability of a given sample is greatest, is chosen as an estimate.”

An estimate obtained by this method is called the maximum likelihood estimate (MLE).

It should be noted that the method of maximum likelihood is applicable to both discrete and continuous random variables.

Let us consider a few examples:

EXAMPLES OF MLE’s IN CASE OF DISCRETE DISTRIBUTIONS Example-1: For the Poisson distribution given by the MLE of  is  X (the sample mean).

Example-2: For the geometric distribution given by the MLE of p is Hence, the MLE of p is equal to the reciprocal of the mean.

Example-3: For the Bernoulli distribution given by the MLE of p is (the sample mean).

EXAMPLES OF MLE’s IN CASE OF CONTINUOUS DISTRIBUTIONS Example-1: For the exponential distribution given by the MLE of  is (the reciprocal of the sample mean).

Example-2: For the normal distribution with parameters  and  2, the joint ML estimators of  and  2 are the sample mean and the sample variance S 2 (which is not an unbiased estimator of  2 ).

As indicated many times earlier, the normal distribution is encountered frequently in practice, and, in this regard, it is both interesting and important to note that, in the case of this frequently encountered distribution, the simplest formulae (i.e. the sample mean and the sample variance ) fulfil the criteria of the relatively advanced method of maximum likelihood estimation !

The last example among the five presented above (the one on the normal distribution) points to another important fact --- and that is :

The Maximum Likelihood Estimators are consistent and efficient but not necessarily unbiased.

(As we know, S 2 is not an unbiased estimator of  2.)

Let us apply this concept to a real-life example:

EXAMPLE It is well-known that human weight is an approximately normally distributed variable. Suppose that we are interested in estimating the mean and the variance of the weights of adult males in one particular province of a country.

A random sample of 15 adult males from this particular population yields the following weights (in pounds): Find the maximum likelihood estimates for  1 =  and  2 =  2.

SOLUTION The above data is that of a random sample of size 15 from N( ,  2 ). It has been mathematically proved that the joint maximum likelihood estimators of  and  2 are  X and S 2.

We compute these quantities for this particular sample, and obtain  X = , and S 2 = These are the Maximum Likelihood Estimates of the mean and variance of the population of weights in this particular example.

Having discussed the concept of point estimation in some detail, we now begin the discussion of the concept of interval estimation :

As stated earlier, whenever a single quantity computed from the sample acts as an estimate of a population parameter, we call that quantity a point estimate e.g. the sample mean is a point estimate of the population mean .

The limitation of point estimation is that we have no way of ascertaining how close our point estimate is to the true value (the parameter).

For example, we know that is an unbiased estimator of  i.e. if we had taken all possible samples of a particular size from the population and calculated the mean of each sample, then the mean of the sample means would have been equal to the population mean (  ), but in an actual survey we will be selecting only one sample from the population and will calculate its mean. We will have no way of ascertaining how close this particular is to .

Whereas a point estimate is a single value that acts as an estimate of the population parameter, interval estimation is a procedure of estimating the unknown parameter which specifies a range of values within which the parameter is expected to lie.

A confidence interval is an interval computed from the sample observations x 1, x 2 ….x n, with a statement of how confident we are that the interval does contain the population parameter.

We develop the concept of interval estimation with the help of the example of the Ministry of Transport test to which all cars, irrespective of age, have to be submitted :

EXAMPLE Let us examine the case of an annual Ministry of Transport test to which all cars, irrespective of age, have to be submitted. The test looks for faulty breaks, steering, lights and suspension, and it is discovered after the first year that approximately the same number of cars have 0, 1, 2, 3, or 4 faults.

You will recall that when we drew all possible samples of size 2 from this uniformly distributed population, the sampling distribution of  X was triangular :

1/ /25 3/25 4/25 5/ Sampling Distribution of  X for n = 2

But when we considered what happened to the shape of the sampling distribution with if the sample size is increased, we found that it was somewhat like a normal distribution:

4/ /125 12/125 16/125 20/125 Sampling Distribution of  X for n = 3

And, when we increased the sample size to 4, the sampling distribution resembled a normal distribution even more closely :

Sampling Distribution of  X for n = 4 20/ /625 60/625 80/ /

It is clear from the above discussion that as larger samples are taken, the shape of the sampling distribution of  X undergoes discernible changes. In all three cases the line charts are symmetrical, but as the sample size increases, the overall configuration changed from a triangular distribution to a bell- shaped distribution.

In other words, for large samples, we are dealing with a normal sampling distribution of. In other words:

When sampling from an infinite population such that the sample size n is large,  X is normally distributed with mean  and variance i.e.  X is

Hence, the standardized version of  X i.e. is normally distributed with mean 0 and variance 1 i.e. Z is N(0, 1).

Now, for the standard normal distribution, we have:

Z For the standard normal distribution, we have: The above is equivalent to P(-1.96 < Z < 1.96) = = 0.95

Z

In other words: The above can be re-written as:

or

The above equation yields the 95% confidence interval for  :

The 95% confidence interval for  is

In other words, the 95% C.I. for  is given by

In a real-life situation, the population standard deviation is usually not known and hence it has to be estimated.

It can be mathematically proved that the quantity is an unbiased estimator of  2 (the population variance).

(just as the sample mean is an unbiased estimator of  ).

In this situation, the 95% Confidence Interval for  is given by:

The points are called the lower and upper limits of the 95% confidence interval.

IN TODAY’S LECTURE, YOU LEARNT  Desirable Qualities of a Good Point Estimator:  Efficiency Methods of Point Estimation:  The Method of Moments  The Method of Least Squares  The Method of Maximum Likelihood Interval Estimation: Confidence Interval for 

IN THE NEXT LECTURE, YOU WILL LEARN Confidence Interval for  (continued) Confidence Interval for  1 -  2 Large Sample Confidence Intervals for p and p 1 -p 2 Determination of Sample Size