Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 6 Point Estimation.
Point Estimation Notes of STAT 6205 by Dr. Fan.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 7. Statistical Estimation and Sampling Distributions
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Statistical Estimation and Sampling Distributions
Sampling: Final and Initial Sample Size Determination
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
SOLVED EXAMPLES.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Today Today: Chapter 9 Assignment: Recommended Questions: 9.1, 9.8, 9.20, 9.23, 9.25.
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
SAMPLING DISTRIBUTIONS. SAMPLING VARIABILITY
2. Point and interval estimation Introduction Properties of estimators Finite sample size Asymptotic properties Construction methods Method of moments.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Statistical Background
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Statistical inference Population - collection of all subjects or objects of interest (not necessarily people) Sample - subset of the population used to.
Maximum likelihood (ML)
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
A statistical model Μ is a set of distributions (or regression functions), e.g., all uni-modal, smooth distributions. Μ is called a parametric model if.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
University of Ottawa - Bio 4118 – Applied Biostatistics © Antoine Morin and Scott Findlay 08/10/ :23 PM 1 Some basic statistical concepts, statistics.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Chapter 7 Point Estimation
1 Lecture 16: Point Estimation Concepts and Methods Devore, Ch
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Physics 270 – Experimental Physics. Let say we are given a functional relationship between several measured variables Q(x, y, …) x ±  x and x ±  y What.
Week 41 Estimation – Posterior mean An alternative estimate to the posterior mode is the posterior mean. It is given by E(θ | s), whenever it exists. This.
Chapter 2 Statistical Background. 2.3 Random Variables and Probability Distributions A variable X is said to be a random variable (rv) if for every real.
Consistency An estimator is a consistent estimator of θ, if , i.e., if
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Confidence Interval & Unbiased Estimator Review and Foreword.
Summarizing Risk Analysis Results To quantify the risk of an output variable, 3 properties must be estimated: A measure of central tendency (e.g. µ ) A.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Week 41 How to find estimators? There are two main methods for finding estimators: 1) Method of moments. 2) The method of Maximum likelihood. Sometimes.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Stats Term Test 4 Solutions. c) d) An alternative solution is to use the probability mass function and.
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Conditional Expectation
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate accuracy.
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
STATISTICS POINT ESTIMATION
Some General Concepts of Point Estimation
Probability Theory and Parameter Estimation I
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Parameter Estimation 主講人:虞台文.
STATISTICAL INFERENCE PART I POINT ESTIMATION
Parametric Methods Berlin Chen, 2005 References:
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Applied Statistics and Probability for Engineers
Presentation transcript:

Parameter Estimation

Statistics Probability specified inferred Steam engine pump “prediction” “estimation”

Parameter Unknown, (fixed) quantity associated with population (example: the true, unknown mean weight of a population) Statistic Summary of sample measurements from the population (example: average weight of a sample) Estimator Statistic associated with a parameter (e.g., average  mean) Denoted by ^ (“hat”) Estimate Specific value from an estimator, for a given sample Vary from sample to sample

Concepts: Bias and Precision Describe the error associated with estimators and estimates

Bias, Precision Bias = systematic error relative to true value Variance (precision) = error around the expected value of estimator

Attributes of Estimators Expected value E(Y) = ΣY*f(y) or E(Y) = ∫Y*f(y) Weighted average of R.V. over all possible values Mean value of die roll: E(Y) = 1*(1/6) + 2*(1/6) *(1/6) = 3.5 Bias: E(Ŷ) – Y Systematic error of the estimator = Difference between expected value of estimator and true value of the parameter Unbiased => E(Ŷ) = Y Variance (precision): var(Ŷ) = E([Ŷ - E(Ŷ)] 2 ) Error (variability) around expected value of estimator = Expected (squared) difference between any single estimate Ŷ and the expected value of the estimator E(Ŷ). It depends on sample size and population variability

Accuracy (balance of bias and precision) Mean Square Error MSE = Variance + Bias² MSE(Ŷ) = Var(Ŷ) + Bias(Ŷ)² MSE(Ŷ) = E[(Ŷ - Y)²] MSE = Expected (squared) distance between any point estimate Ŷ and the true parameter value Y. MSE combines both types of errors: bias and variance.

Bias, Precision Accurate !

Deriving Estimators We assume data is collected under good experimental design (i.e., random sample) Think of data as observations of random variables We have an idea of a reasonable probability distribution that generated the data (‘model’) Approaches for developing estimators of parameters – Least-squares – Method of moments – Maximum likelihood

Maximum Likelihood Estimation Model = probability distribution is used as the likelihood function Roles of parameter and variable are switched, use “data” to estimate parameter: L (θ | x) Given a specific model (stat. distribution), what parameter value is the most likely to have produced the data ? Data: fixed and known Likelihood is evaluated over the space of parameter values θ’s Choose that maximizes

Probability – given a parameter what is the probability of each value of x MLE – given a value of x, which p has the highest likelihood? maximize (f[7|p]) Coin toss n = 10

MLE – given a value of x, which p has the highest likelihood? maximize (f[7|p])

MLE of is found by differentiating log L (θ | x) with respect to θ, setting equal to zero, then solving for θ

Optimization

MLE’s have some nice properties Often match intuition Asymptotically normal Consistent (estimator matches parameter as sample size grows large) For large sample sizes, has smallest variances among unbiased estimates Likelihood also can be used to derive variances and covariances (Fishers information)

Confidence Intervals for MLEs

Normal-based CI’s MLE is within 1.96 standard deviations of with probability P = 0.95, 95 percent confidence interval

Confidence Intervals based on Profile Likelihoods Normal-based CI’s not always valid (small samples) Profile likelihoods provides an another alternative A confidence interval may be obtained by solving for θ 0 chi-square distribution df=1

Bayesian Inference Inference is based on Pr(θ|x): Posterior Distribution Variance in θ represents our uncertainty in the value of θ Principle Based on Bayes’ Theorem Pr(θ | x) ~ Pr(x|θ) Pr(θ) Prior Knowledge + Insights from Data= Basis of Inference Distribution of parameters Pr(θ): Prior Distribution Updated with information from the data X: How likely is it to obtain data X with parameter value θ: Pr(x|θ) (Likelihood) => Pr(θ|x) (variance of θ shrinks)

END

Bayes’ Theorem Use for Bayesian Model updating Use for model Bayesian estimation