LEARNING OBJECTIVES After careful study of this chapter you should be able to do the following: 1.Explain the general concepts of estimating the parameters.

Slides:



Advertisements
Similar presentations
Chapter 4 Inference About Process Quality
Advertisements

Estimation of Means and Proportions
Previous Lecture: Distributions. Introduction to Biostatistics and Bioinformatics Estimation I This Lecture By Judy Zhong Assistant Professor Division.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 7. Statistical Estimation and Sampling Distributions
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Statistical Estimation and Sampling Distributions
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Maximum likelihood (ML) and likelihood ratio (LR) test
Chapter 6 Introduction to Sampling Distributions
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Maximum likelihood (ML)
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
Evaluating Hypotheses
Part III: Inference Topic 6 Sampling and Sampling Distributions
Copyright © Cengage Learning. All rights reserved. 6 Point Estimation.
Inferences About Process Quality
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
BCOR 1020 Business Statistics
Maximum likelihood (ML)
Chapter 6: Sampling Distributions
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Virtual COMSATS Inferential Statistics Lecture-6
STA Lecture 161 STA 291 Lecture 16 Normal distributions: ( mean and SD ) use table or web page. The sampling distribution of and are both (approximately)
Estimation of Statistical Parameters
Lecture 14 Sections 7.1 – 7.2 Objectives:
1 Introduction to Estimation Chapter Concepts of Estimation The objective of estimation is to determine the value of a population parameter on the.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
PARAMETRIC STATISTICAL INFERENCE
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
1 Estimation From Sample Data Chapter 08. Chapter 8 - Learning Objectives Explain the difference between a point and an interval estimate. Construct and.
Chapter 7 Point Estimation
1 Lecture 16: Point Estimation Concepts and Methods Devore, Ch
Week 6 October 6-10 Four Mini-Lectures QMM 510 Fall 2014.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Chapter 8: Confidence Intervals based on a Single Sample
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Confidence Interval Estimation For statistical inference in decision making:
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
Review of Statistics.  Estimation of the Population Mean  Hypothesis Testing  Confidence Intervals  Comparing Means from Different Populations  Scatterplots.
SAMPLING DISTRIBUTION OF MEANS & PROPORTIONS. PPSS The situation in a statistical problem is that there is a population of interest, and a quantity or.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Virtual University of Pakistan
Chapter 6: Sampling Distributions
Copyright © Cengage Learning. All rights reserved.
Chapter 5 STATISTICAL INFERENCE: ESTIMATION AND HYPOTHESES TESTING
STATISTICAL INFERENCE
12. Principles of Parameter Estimation
Sampling Distributions and Estimation
Some General Concepts of Point Estimation
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Chapter 6: Sampling Distributions
CONCEPTS OF ESTIMATION
6 Point Estimation.
12. Principles of Parameter Estimation
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Applied Statistics and Probability for Engineers
Some General Concepts of Point Estimation
Presentation transcript:

LEARNING OBJECTIVES After careful study of this chapter you should be able to do the following: 1.Explain the general concepts of estimating the parameters of a population or a probability distribution 2. Explain important properties of point estimators, including bias, variance, and mean square error 3. Know how to construct point estimators using the method of moments and the method of maxi- mum likelihood 4. Know how to compute and explain the precision with which a parameter is estimated 5. Understand the central limit theorem 6. Explain the important role of the normal distribution as a sampling distribution

Introduction Statistical inference is almost always directed toward drawing some type of conclusion about one or more parameters (population characteristics). These methods utilize the information contained in a sample from the population in drawing conclusions. When discussing general concepts and methods of inference, it is convenient to have a generic symbol for the parameter of interest. We will use the Greek letter  for this purpose. The objective of point estimation is to select a single number, based on sample data, that represents a sensible value for  Statistical inference may be divided into two major areas: Parameter estimation Hypothesis testing

Example As an example of a parameter estimation problem, suppose that a structural engineer is analyzing the tensile strength of a component used in an automobile chassis. Since variability in tensile strength is naturally present between the individual components because of differences in raw material batches, manufacturing processes, and measurement procedures (for example), the engineer is interested in estimating the mean tensile strength of the components. In practice, the engineer will use sample data to compute a number that is in some sense a reasonable value (or guess) of the true mean. This number is called a point estimate. We will see that it is possible to establish the precision of the estimate. Hypothesis testing: Now consider a situation in which two different reaction temperatures can be used in a chemical process, say t1 and t2. The engineer conjectures that t1 results in higher yields than does t2. Statistical hypothesis testing is a framework for solving problems of this type. In this case, the hypothesis would be that the mean yield using temperature t1 is greater than the mean yield using temperature t2. Notice that there is no emphasis on estimating yields; instead, the focus is on drawing conclusions about a stated hypothesis.

How Suppose that we want to obtain a point estimate of a population parameter. We know that before the data is collected, the observations are considered to be random variables, say X1, X2, p, Xn. Therefore, any function of the observation, or any statistic, is also a random variable. For example, the sample mean X and the sample variance S2 are statistics and they are also random variables. Since a statistic is a random variable, it has a probability distribution. We call the probability distribution of a statistic a sampling distribution. The notion of a sampling distribution is very important When discussing inference problems, it is convenient to have a general symbol to represent the parameter of interest. We will use the Greek symbol e (theta) to represent the parameter. The objective of point estimation is to select a single number, based on sample data, that is the most plausible value for e. A numerical value of a sample statistic will be used as the point estimate.

Point estimation In statistics, point estimation involves the use of sample data to calculate a single value (known as a statistic) which is to serve as a "best guess" or "best estimate" of an unknown (fixed or random) population parameter. More formally, it is the application of a point estimator to the data. In general, point estimation should be contrasted with interval estimation: such interval estimates are typically either confidence intervals. Introduction

Definition

Point estimators Minimum-variance mean-unbiased estimator (MVUE), Minimum mean squared error (MMSE) Method of moments, generalized method of moments Maximum likelihood (ML) Minimizes the risk (expected loss) of the squared-error loss-function. Best linear unbiased estimator (BLUE) Median-unbiased estimator, minimizes the risk of the absolute-error loss function Point Estimators

Introduction

Some General Concepts of Point Estimation

Unbiased Estimators example 7-1 proof of that

Unbiased Estimators

Example unbiased estimator

Variance of a Point Estimator In a sense, the MVUE is most likely among all unbiased estimators to produce an estimate eˆ that is close to the true value of e.

Standard Error: Reporting a Point Estimate When the numerical value or point estimate of a parameter is reported, it is usually desirable to give some idea of the precision of estimation. The measure of precision usually employed is the standard error of the estimator that has been used.

Standard Error: Reporting a Point Estimate

Mean Square Error of an Estimator

 Sometimes we prefer biased estimator than unbiased estimator because they have smaller square error. Optimum estimator: An estimator that has a mean square error less than or equal to the mean square error of any other estimator. o Optimum estimator rarely exist

Recap How The objective of point estimation is to select a single number, based on sample data, that is the most plausible value, point estimation involves the use of sample data to calculate a single value (known as a statistic) which is to serve as a "best guess" or "best estimate" of an unknown (fixed or random) population parameter. 1.Suppose that we want to obtain a point estimate of a population parameter. 2.We know that before the data is collected, the observations are considered to be random variables, say X1, X2, p, Xn. 3.Therefore, any function of the observation, or any statistic, is also a random variable. 4. For example, the sample mean X and the sample variance S2 are statistics and they are also random variables. 5.Since a statistic is a random variable, it has a probability distribution. 6.We call the probability distribution of a statistic a sampling distribution. 7. The notion of a sampling distribution is very important 8. When discussing inference problems, it is convenient to have a general symbol to represent the parameter of interest. We will use the Greek symbol e (theta) to represent the parameter.

Methods of point estimation The definitions of unbiasness and other properties of estimators do not provide any guidance about how good estimators can be obtained. In this section, we discuss two methods for obtaining point estimators: 1.the method of moments and 2.the method of maximum likelihood. Maximum likelihood estimates are generally preferable to moment estimators because they have better efficiency properties. However, moment estimators are sometimes easier to compute. Both methods can produce unbiased point estimators.

Methods of Point Estimation other text!! The definition of unbiasedness does not in general indicate how unbiased estimators can be derived. We now discuss two “constructive” methods for obtaining 1.point estimators: the method of moments and the 2.method of maximum likelihood. By constructive استنتاجي we mean that the general definition of each type of estimator suggests explicitly how to obtain the estimator in any specific problem.

Method of Moments The general idea behind the method of moments is to equate population moments, which are defined in terms of expected values, to the corresponding sample moments. The basic idea of this method is to equate certain sample characteristics, such as the mean, to the corresponding population expected values. Then solving these equations for unknown parameter values yields the estimators.

Method of Moments Example exponential

Method of Moments example

Method of Moments example Gamma distribution

Method of Maximum Likelihood Most statisticians recommend this method, at least when the sample size is large, since the resulting estimators have certain desirable efficiency properties. One of the best methods of obtaining a point estimator of a parameter is the method of maxi- mum likelihood. This technique was developed in the 1920s by a famous British statistician, Sir R. A. Fisher. As the name implies, the estimator will be the value of the parameter that maximizes the likelihood function أرجحية.

Properties of the Maximum Likelihood Estimator The method of maximum likelihood is often the estimation method that mathematical statisticians prefer, because it is usually easy to use and produces estimators with good statistical properties. We summarize these properties as follows. Properties 1 and 2 essentially state that the maximum likelihood estimator is approximately an MVUE. This is a very desirable result and, coupled with the fact that it is fairly easy to obtain in many situations and has an asymptotic normal distribution (the “asymptotic” means “when n is large”), explains why the maximum likelihood estimation technique is widely used. To use maximum likelihood estimation, remember that the distribution of the population must be either known or assumed.

Large Sample Behavior of the MLE

Method of Maximum Likelihood

Maximum Likelihood Estimation

Maximum Likelihood Estimation Example

Illustrate graphically just how the method of maximum likelihood works. It is easy to illustrate graphically just how the method of maximum likelihood works. Figure 7-3(a) plots the log of the likelihood function for the exponential parameter from Example 7- 8, using the n = 8 observations on failure time given following Example 7-3. We found that the estimate of A was Aˆ = From Example 7-8, we know that this is a maximum likelihood estimate. Figure 7-3(a) shows clearly that the log likelihood function is maximized at a value of A that is approximately equal to Notice that the log likelihood function is relatively flat in the region of the maximum. This implies that the parameter is not estimated very precisely. If the parameter were estimated precisely, the log likelihood function would be very peaked at the maximum value. The sample size here is relatively small, and this has led to the imprecision in estimation. This is illustrated in Fig. 7-3(b) where we have plot- ted the difference in log likelihoods for the maximum value, assuming that the sample sizes were n = 8, 20, and 40 but that the sample average time to failure remained constant at x = Notice how much steeper the log likelihood is for n = 20 in comparison to n = 8, and for n = 40 in comparison to both smaller sample sizes.

Theory & Theorem Definitions of theorem a general proposition not self-evident but proved by a chain of reasoning; a truth established by means of accepted truths. A theorem is a result that can be proven to be true from a set of axioms البديهيات. The term is used especially in mathematics where the axioms are those of mathematical logic and the systems in question. A theory is a set of ideas used to explain why something is true, or a set of rules on which a subject is based on.

Review Sampling Distributions and the Central Limit Theorem Statistical inference is concerned with making decisions about a population based on the information contained in a random sample from that population. For instance, we may be interested in the mean f ill volume of a can of soft drink. The mean f ill volume in the population is required to be 300 milliliters. An engineer takes a random sample of 25 cans and computes the sample average fill volume to be x = 298 milliliters. The engineer will probably decide that the population mean is = 300 milliliters, even though the sample mean was 298 milliliters because he or she knows that the sample mean is a reasonable estimate of and that a sample mean of 298 milliliters is very likely to occur, even if the true population mean is = 300 milliliters. In fact, if the true mean is 300 milliliters, tests of 25 cans made repeatedly, perhaps every five minutes, would produce values of x that vary both above and below =300 milliliters. The sample mean is a statistic; that is, it is a random variable that depends on the results obtained in each particular sample. Since a statistic is a random variable, it has a probability distribution.

Review Sampling Distributions and the Central Limit Theorem Definitions: For example, the probability distribution of X is called the sampling distribution of the mean. The sampling distribution of a statistic depends on the distribution of the population, the size of the sample, and the method of sample selection.

Review Sampling distribution Each observation say X1, X2, X3, X4, ……., Xn is normally and independently distributed random variable with mean μ and variance σ 2

Sampling Distributions and the Central Limit Theorem of the most useful theorems in statistics, called the central limit theorem.

Die & Dice’s

Sampling Distributions and the Central Limit Theorem Example

Sampling Distributions and the Central Limit Theorem Figure 7-2 Probability for Example 7-1

Sampling Distributions and the Central Limit Theorem two independent populations If the two populations are not normally distributed and if both sample sizes n1 and n2 are greater than 30, we may use the central limit theorem and assume that X1 and X2 follow approximately independent normal distributions.

Example two independent populations

Example

The compressive strength of concrete is normally distributed with μ=2500 psi and σ = 50 psi, find the probability that a random sample of n = 5 specimens will have a sample mean diameter that falls in the interval from 2499 to 2510 psi.

Recap Point estimation of parameters

Process or Population How do you estimate these unknown parameters?  = ?  2 = ? p = ?

Using properly drawn sample data to draw conclusions about the population is called statistical inference. Process or Population  = ? Sample is a sample estimate of .

Statistical inference permits the estimation of statistical measures (means, proportions, etc.) with a known degree of confidence. This ability to assess the quality (confidence) of estimates is one of the significant benefits statistics brings to decision making and problem solving. If samples are selected randomly, the uncertainty of the inference is measurable. If samples are not selected randomly, there will be no known relationship between the sample results and the population

This chapter is devoted to the one-sample problem.  That is, a sample consisting of n measurements, x 1, x 2,..., x n, of some population characteristic will be analyzed with the objective of making inferences about the population or process.  = ?  2 = ? p = ? The One-sample Problem

 If we wish to estimate the mean of a population, we might consider the sample mean, the sample median, or perhaps the average of the smallest and largest observations in the sample as point estimators In order to decide which point estimator of a particular parameter is the best one to use, we need to examine their statistical properties and develop some criteria for comparing estimators

Example: Prove that the estimator of σ 2 is S 2 The sample variance S 2 i s unbiased estimator of σ 2