ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Managerial Economics in a Global Economy
Lesson 10: Linear Regression and Correlation
Kin 304 Regression Linear Regression Least Sum of Squares
The Simple Regression Model
Chapter 12 Simple Linear Regression
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
SOLVED EXAMPLES.
Objectives (BPS chapter 24)
Simple Linear Regression
A Short Introduction to Curve Fitting and Regression by Brad Morantz
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
Chapter 4 Multiple Regression.
Simulation Modeling and Analysis Session 12 Comparing Alternative System Designs.
Probability & Statistics for Engineers & Scientists, by Walpole, Myers, Myers & Ye ~ Chapter 11 Notes Class notes for ISE 201 San Jose State University.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Simple Linear Regression and Correlation
Introduction to Regression Analysis, Chapter 13,
1 Simple Linear Regression 1. review of least squares procedure 2. inference for least squares lines.
Correlation & Regression
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
CPE 619 Simple Linear Regression Models Aleksandar Milenković The LaCASA Laboratory Electrical and Computer Engineering Department The University of Alabama.
Simple Linear Regression Models
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
7.1 Multiple Regression More than one explanatory/independent variable This makes a slight change to the interpretation of the coefficients This changes.
Lecture 3: Inference in Simple Linear Regression BMTRY 701 Biostatistical Methods II.
CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics Instructor Longin Jan Latecki C22: The Method of Least Squares.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1Spring 02 First Derivatives x y x y x y dy/dx = 0 dy/dx > 0dy/dx < 0.
Regression Regression relationship = trend + scatter
1 Multiple Regression A single numerical response variable, Y. Multiple numerical explanatory variables, X 1, X 2,…, X k.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
Confidence Interval & Unbiased Estimator Review and Foreword.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
1 Ka-fu Wong University of Hong Kong A Brief Review of Probability, Statistics, and Regression for Forecasting.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Inference about the slope parameter and correlation
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves.
The simple linear regression model and parameter estimation
Chapter 4: Basic Estimation Techniques
Chapter 4 Basic Estimation Techniques
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Part 5 - Chapter
Part 5 - Chapter 17.
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
Basic Estimation Techniques
Kin 304 Regression Linear Regression Least Sum of Squares
Ch3: Model Building through Regression
CH 5: Multivariate Methods
BPK 304W Regression Linear Regression Least Sum of Squares
Slope of the regression line:
Basic Estimation Techniques
Part 5 - Chapter 17.
Linear regression Fitting a straight line to observations.
Regression Lecture-5 Additional chapters of mathematics
Simple Linear Regression
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves.
12. Principles of Parameter Estimation
Presentation transcript:

ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves Several methods to do this, we’ll look at only one: MAXIMUM LIKELIHOOD First, define Likelihood: L(y 1, y 2, …., y N ) is the joint probability density evaluated at the observations y i where y 1, y 2, …., y N are sample observations of random variables Y 1, Y 2, …., Y N PDF of random variables Y 1, Y 2, …., Y N

MAXIMUM LIKELIHOOD METHOD Choose the parameter values that maximize Example: Apply method to estimates of  and  2 for a normal population. Let y 1, y 2, …., y N be a random sample of the normal population Find Maximum Likelihood

Simplify by taking the log N (L): Taking derivative with respect to  and  2 Making them equal to zero to get the maximum, the maximum likelihood:

Making them equal to zero to get maximum likelihood estimators of mean and variance: are the Maximum Likelihood estimators of  and  2 is an unbiased estimator of , but is not unbiased for  2 substituting  hat into

can be adjusted to the unbiased estimator: So, for a normally distributed oceanographic data set, we can readily obtain Maximum Likelihood estimates of  and  2 This technique (ML) is really useful for variables that are not normally distributed. Spectral energy values from current velocities or sea level, show  2 rather than normal distribution Following the ML procedure, we find that the mean of the spectral values is and the variance is 2

So, with the ML approach you can calculate the best parameters that fit certain models. For instance, you can apply it to a pulse of current velocity data to obtain the best dissipation value  and fitting coefficient  in the inertial subrange, on the basis of Kolmogorov’s law for turbulence:

As another example, you can apply it to a segment of temperature gradient in a profile to obtain the best Batchelor length scale (or wave number  B ) and dissipation of temperature variance  T, to get dissipation values on the basis of Batchelor spectrum for turbulence: Steinbuck et al., 2009 So in general, to apply the ML method to a sample: - Determine appropriate PDF for sample values - Find joint likelihood function - Take natural logs - Differentiate wrt parameter of interest - Set derivative = 0 to find max - Obtain value of parameter

LINEAR ESTIMATION (REGRESSION) Consider the values y of a random variable Y called dependent variable. The values y are a function of one or more non-random variables x 1, x 2, …, x N called independent variables. The random variable can be modeled (represented) as: The random variable  (not to be confused with dissipation used before) gives the departure from linearity and has a specific PDF with mean of zero. Simple linear regression:

If N independent variables are involved then we have a multiple linear regression: A powerful method to fit the independent variables x 1, x 2, …, x N to the dependent variable y is the method of least squares The simplest case is to fit a straight line to a set of points using the “best” coefficients b 0, b 1 The method of least squares does what we do by eye, i.e., minimize deviations (residuals) between data points and fitted line. x y

Let: where: is the deterministic portion of the data is the residual or error To find b 0, b 1 minimize the sum of the squared errors (SSE) Sum of Squares Total (data variance) Sum of Squares Regression (variance explained by regression)

To minimize the sum of the squared errors (SSE) Two equations, two unknowns; solve for the parameters

Regression line splits the scatter of observations such that the positive residuals cancel out with negative residuals x y Regression line always goes through

Percent explained variance R 2 : Sum of Squares Total (data variance) Sum of Squares Regression (variance explained by regression) Goodness of Fit (Correlation of Determination) Least squares can be used to fit any curve – we’ll see it in harmonic analysis Least squares can be considered a Maximum Likelihood Estimator

x y x’ y’  x cos  -x sin  (x, 0) (0, y) y cos  y sin  Rotation of axes  can be obtained from linear regression of scatter diagram

Rotation of axes

CORRELATION Concept linked to time series analysis Correlation coefficient: determines how well two variables co-vary in time or space. For two random variables, x and y the correlation coefficient can be: C xy is the covariance of x and y, and s x and s y are the stdev

AUTOCORRELATION x are the measurements L represents a lag N is the total number of measurements overbar represents mean over the N measurements r x is the autocorrelation coefficient for x r x oscillates between -1 and 1 r x equals1 at L = 0