Ch. 19 Unbiased Estimators Ch. 20 Efficiency and Mean Squared Error CIS 2033: Computational Probability and Statistics Prof. Longin Jan Latecki Prepared.

Slides:



Advertisements
Similar presentations
Multiple Regression Analysis
Advertisements

The Simple Regression Model
Ch. 17 Basic Statistical Models CIS 2033: Computational Probability and Statistics Prof. Longin Jan Latecki Prepared by: Nouf Albarakati.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Statistical Estimation and Sampling Distributions
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
McGraw-Hill Ryerson Copyright © 2011 McGraw-Hill Ryerson Limited. Adapted by Peter Au, George Brown College.
Econ 140 Lecture 61 Inference about a Mean Lecture 6.
Point estimation, interval estimation
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Chapter 6 Introduction to Sampling Distributions
Part 2b Parameter Estimation CSE717, FALL 2008 CUBS, Univ at Buffalo.
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
SAMPLING DISTRIBUTIONS. SAMPLING VARIABILITY
Visual Recognition Tutorial
The Simple Regression Model
Ka-fu Wong © 2007 ECON1003: Analysis of Economic Data Lesson6-1 Lesson 6: Sampling Methods and the Central Limit Theorem.
CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics Instructor Longin Jan Latecki C12: The Poisson process.
Basic Biostat8: Intro to Statistical Inference1. In Chapter 8: 8.1 Concepts 8.2 Sampling Behavior of a Mean 8.3 Sampling Behavior of a Count and Proportion.
 The Law of Large Numbers – Read the preface to Chapter 7 on page 388 and be prepared to summarize the Law of Large Numbers.
Review of Probability and Statistics
The Basics  A population is the entire group on which we would like to have information.  A sample is a smaller group, selected somehow from.
CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics, 2007 Instructor Longin Jan Latecki Chapter 7: Expectation and variance.
Chapter 6: Sampling Distributions
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Statistical Inference: Estimation and Hypothesis Testing chapter.
Simulation Output Analysis
Sampling distributions, Point Estimation Week 3: Lectures 3 Sampling Distributions Central limit theorem-sample mean Point estimators-bias,efficiency Random.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 7 Sampling Distributions.
AP Statistics Chapter 9 Notes.
Montecarlo Simulation LAB NOV ECON Montecarlo Simulations Monte Carlo simulation is a method of analysis based on artificially recreating.
1 Sampling Distributions Lecture 9. 2 Background  We want to learn about the feature of a population (parameter)  In many situations, it is impossible.
CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics Instructor Longin Jan Latecki C22: The Method of Least Squares.
CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics Michael Baron. Probability and Statistics for Computer Scientists,
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 9 Samples.
Copyright © 2005 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics Thomas Maurice eighth edition Chapter 4.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
June 16. In Chapter 8: 8.1 Concepts 8.2 Sampling Behavior of a Mean 8.3 Sampling Behavior of a Count and Proportion.
MATH 3033 based on Dekking et al. A Modern Introduction to Probability and Statistics Slides by Sean Hercus Instructor Longin Jan Latecki Ch. 6 Simulations.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Confidence Interval & Unbiased Estimator Review and Foreword.
CIS 2033 A Modern Introduction to Probability and Statistics Understanding Why and How Chapter 17: Basic Statistical Models Slides by Dan Varano Modified.
Sampling and estimation Petter Mostad
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Chapter 5 Sampling Distributions. The Concept of Sampling Distributions Parameter – numerical descriptive measure of a population. It is usually unknown.
Review of Statistical Terms Population Sample Parameter Statistic.
Sampling Theory and Some Important Sampling Distributions.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Sampling and Sampling Distributions. Sampling Distribution Basics Sample statistics (the mean and standard deviation are examples) vary from sample to.
CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics B: Michael Baron. Probability and Statistics for Computer Scientists,
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Virtual University of Pakistan
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Section 9.1: Parameter estimation CIS Computational Probability.
Chapter 6: Sampling Distributions
STATISTICAL INFERENCE
Chapter 19: Unbiased estimators
Basic Estimation Techniques
Chapter 4. Inference about Process Quality
Chapter 6: Sampling Distributions
Ch9 Random Function Models (II)
Behavioral Statistics
Consolidation & Review
Basic Estimation Techniques
Sampling Distributions
C19: Unbiased Estimators
CIS 2033 based on Dekking et al
C19: Unbiased Estimators
CIS 2033 based on Dekking et al
Presentation transcript:

Ch. 19 Unbiased Estimators Ch. 20 Efficiency and Mean Squared Error CIS 2033: Computational Probability and Statistics Prof. Longin Jan Latecki Prepared in part by: Nouf Albarakati

An Estimate  An estimate is a value that only depends on the dataset x 1, x 2,...,x n, i.e., t is some function of the dataset only: t = h(x 1, x 2,...,x n )  That means: value t, computed from our dataset x 1, x 2,...,x n, gives some indication of the “true” value of the parameter of interest θ

An Example  Consider the example of arrivals of packages at a network server  a dataset x 1, x 2,...,x n, where x i represents the number of arrivals in the i th minute  The intensity of the arrivals is modeled by the parameter µ  The percentage of minutes during which no packages arrive (idle) is modeled by the probability of zero arrivals: e −µ

An Example (cont.)  The percentage of idle minutes is modeled by the probability of zero arrivals  Since the parameter µ is the expectation of the model distribution,  the law of large numbers suggests the sample mean as a natural estimate for µ   If µ is estimated by, e −µ also is estimated by

An Estimator  Since dataset x 1, x 2,...,x n is modeled as a realization of a random sample X 1, X 2,...,X n, the estimate t is a realization of a random variable T

The Behavior of an Estimator  Network server example: The dataset is modeled as a realization of a random sample of size n = 30 from a Pois(μ) distribution  Estimating the probability p 0 of zero arrivals?  Two estimators S and T  pretend we know μ, simulate the estimation process in the case of n = 30 observations  choose μ = ln 10, so that p 0 = e −μ = 0.1.  draw 30 values from a Poisson distribution with parameter μ = ln 10 and compute the value of estimators S and T

The Behavior of an Estimator

Sampling Distribution  Let T = h(X 1,X 2,..., X n ) be an estimator based on a random sample X 1,X 2,..., X n. The probability distribution of T is called the sampling distribution of T  The sampling distribution of S can be found as follows  where Y is the number of X i equal to zero. If for each i we label X i = 0 as a success, then Y is equal to the number of successes in n independent trials with p 0 as the probability of success. It follows that Y has a Bin(n, p 0 ) distribution. Hence the sampling distribution of S is that of a Bin(n, p 0 ) distributed random variable divided by n

Unbiasedness  An estimator T is called an unbiased estimator for the parameter θ, if E[T] = θ irrespective of the value of θ  The difference E[T ] − θ is called the bias of T ; if this difference is nonzero, then T is called biased  For S and T estimators example, the estimator T has positive bias, though the bias decreases to zero as the sample size becomes larger, while estimator S is unbiased estimator.

Since E[X i ]=0, Var[X i ] = E[X i 2 ] – (E[X i ]) 2 = E[X i 2 ].

Unbiased estimators for expectation and variance

We show first that is an unbiased estimator for μ. We assume that for every i, E(X i )= μ. Using the linearity of the expectation we find that

We show now that is an unbiased estimator for σ 2.

Recall that our goal was to estimate the probability p 0 = e −μ of zero arrivals (of packages) in a minute. We did have two promising candidates as estimators: In Figure 20.2 we depict histograms of one thousand simulations of the values of S and T computed for random samples of size n = 25 from a Pois(μ) distribution, where μ = 2. Considering the way the values of the (biased!) estimator T are more concentrated around the true value e −μ = e −2 = , we would be inclined to prefer T over S.

This choice is strongly supported by the fact that T is more efficient than S: MSE(T ) is always smaller than MSE(S), as illustrated in Figure 20.3.