Presentation is loading. Please wait.

Presentation is loading. Please wait.

ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves.

Similar presentations


Presentation on theme: "ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves."— Presentation transcript:

1 ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves Several methods to do this, we’ll look at only one: MAXIMUM LIKELIHOOD First, define Likelihood: L(y1, y2, …., yN) where y1, y2, …., yN are sample observations of random variables Y1, Y2, …., YN is the joint probability density evaluated at the observations yi PDF of random variables Y1, Y2, …., YN

2 MAXIMUM LIKELIHOOD METHOD
Choose the parameter values that maximize Example: Apply method to estimates of  and 2 for a normal population. Let y1, y2, …., yN be a random sample of the normal population Find Maximum Likelihood

3 Simplify by taking the logN (L):
Taking derivative with respect to  and 2 Making them equal to zero to get the maximum, the maximum likelihood:

4 are the Maximum Likelihood estimators of  and 2
Making them equal to zero to get maximum likelihood estimators of mean and variance: substituting hat into are the Maximum Likelihood estimators of  and 2 is an unbiased estimator of  , but is not unbiased for 2

5 can be adjusted to the unbiased estimator:
So, for a normally distributed oceanographic data set, we can readily obtain Maximum Likelihood estimates of  and 2 This technique (ML) is really useful for variables that are not normally distributed. Spectral energy values from current velocities or sea level, show 2 rather than normal distribution Following the ML procedure, we find that the mean of the spectral values is  and the variance is 2

6 So, with the ML approach you can calculate the best parameters that fit certain models.
For instance, you can apply it to a pulse of current velocity data to obtain the best dissipation value  and fitting coefficient  in the inertial subrange , on the basis of Kolmogorov’s law for turbulence:

7 Vertical velocity, Matanzas Inlet
3 min 8 min m2/s2/Hz Vertical velocity, Matanzas Inlet Jan 20, 2018

8 So in general, to apply the ML method to a sample:
As another example, you can apply it to a segment of temperature gradient in a profile to obtain the best Batchelor length scale (or wave number B) and dissipation of temperature variance T, to get dissipation values on the basis of Batchelor spectrum for turbulence: So in general, to apply the ML method to a sample: - Determine appropriate PDF for sample values - Find joint likelihood function - Take natural logs - Differentiate wrt parameter of interest - Set derivative = 0 to find max - Obtain value of parameter Steinbuck et al., 2009

9 LINEAR ESTIMATION (REGRESSION)
Consider the values y of a random variable Y called dependent variable. The values y are a function of one or more non-random variables x1, x2, …, xN called independent variables. The random variable can be modeled (represented) as: The random variable  (not to be confused with dissipation used before) gives the departure from linearity and has a specific PDF with mean of zero. Simple linear regression:

10 If N independent variables are involved then we have a multiple linear regression:
A powerful method to fit the independent variables x1 , x2 , … , xN to the dependent variable y is the method of least squares x y The simplest case is to fit a straight line to a set of points using the “best” coefficients b0 , b1 The method of least squares does what we do by eye, i.e., minimize deviations (residuals) between data points and fitted line.

11 Sum of Squares Regression (variance explained by regression)
Let: where: is the deterministic portion of the data is the residual or error To find b0 , b1 minimize the sum of the squared errors (SSE) Sum of Squares Total (data variance) Sum of Squares Regression (variance explained by regression)

12 To minimize the sum of the squared errors (SSE)
Two equations, two unknowns; solve for the parameters

13 x y Regression line always goes through Regression line splits the scatter of observations such that the positive residuals cancel out with negative residuals

14 Percent explained variance R2:
Sum of Squares Total (data variance) Sum of Squares Regression (variance explained by regression) Goodness of Fit (Coefficient of Determination) Least squares can be used to fit any curve – we’ll see it in harmonic analysis Least squares can be considered a Maximum Likelihood Estimator

15 q can be obtained from linear regression of scatter diagram (0, y)
Rotation of axes q can be obtained from linear regression of scatter diagram y sinq (0, y) x’ y cosq x cosq -x sinq q x (x, 0)

16 Rotation of axes

17 CORRELATION Concept linked to time series analysis
Correlation coefficient: determines how well two variables co-vary in time or space. For two random variables, x and y the correlation coefficient can be: Cxy is the covariance of x and y, and sx and sy are the stdev

18 AUTOCORRELATION x are the measurements L represents a lag
N is the total number of measurements overbar represents mean over the N measurements rx is the autocorrelation coefficient for x rx oscillates between -1 and 1 rx equals1 at L = 0

19 Examples of Covariance, Correlation, Cross-Correlation & Autocorrelation
load(‘matanzas_data.mat') time=matadat(:,1)+matadat(:,2)/24; % define time u=matadat(:,3); % identify u component i1=find(matadat(:,6) == matadat(1,6)); % identify indices of data at the surface (bin 1) i2=find(matadat(:,6) == matadat(2,6)); % identify indices of data at bin 2 u1=u(i1); u2=u(i2); tim=time(i2); n=length(tim); u1i=interp1(time(i1),u1,tim); % interpolate u1 to the same time of u2 u1=u1i; c1=sum((u1-mean(u1)).^2)/(n-1); %covariance of u1 with itself (autocovariance) c3=sum((u1-mean(u1)).*(u2-mean(u2)))/(n-1); % covariance between u1 and u2 c2=sum((u2-mean(u2)).^2)/(n-1); %covariance of u2 with itself (autocovariance) c=cov(u1,u2); % covariance matrix r1=c1/var(u1); r2=c2/var(u2); r3=c3/(std(u1)*std(u2)); % correlation coefficients r=corr(u1,u2); % autocorrelation [xc,lags]=xcorr(u1,u1,'coeff'); figure; plot(lags/4,xc) hold on ; ef=1-exp(-1); plot([-10000,10000],[ef,ef],'r'); hold off % cross-correlation [xc,lags]=xcorr(u1,u2,'coeff'); figure; plot(lags/4,xc)


Download ppt "ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves."

Similar presentations


Ads by Google