Download presentation
Presentation is loading. Please wait.
Published byArabella Wright Modified over 8 years ago
1
580.691 Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models
2
Review: maximum likelihood estimate of parameters and noise The “true” underlying process What we measured Our model of the process Our ML estimate, given X : Log-likelihood function to maximize:
3
Variance of scalar and vector random variables cov of vector random variables produce symmetric positive definite matrices
4
We say that vector x has a multivariate Gaussian distribution with mean vector and variance-covariance matrix , iff each element has Normal distribution. Multivariate Normal distribution d: dimension of x Variance-covariance matrix
5
Bias of the parameter estimates for a given X Suppose the outputs y were actually produced by the process: The “true” underlying process What we measured Our model of the process Given a constant X, the underlying process would give us different y every time that we run it. If on each run we find an ML w and , how would w and vary with respect to w* ? If there were no noise: ML estimate: In the absence of noise, ML estimate would recover w* exactly.
6
Bias of the parameter estimates for a given X How does the ML estimate behave in the presence of noise in y? The “true” underlying process What we measured Our model of the process nx1 vector ML estimate: Because is normally distributed: In other words:
7
Bias and variance of an estimator Parameters of a distribution can be estimated (e.g., via ML). Here we assess the “goodness” of that estimate by quantifying its bias and variance. Given some data samples: Bias of the estimator is the expectation of the deviation from the true value of the parameter. Variance of the estimator is the anticipated uncertainty in the estimate due to the particular selection of the samples. Note that bias of an estimator is similar to structural errors. Variance of the estimator is similar to approximation errors.
8
… so is unbiased. But what about its variance? Bias of the ML parameter estimates for a given X
9
Variance of the parameter estimates for a given X For a given X, the ML (or least square) estimate of our parameter has this normal distribution: Matrix of constants vector of random variables Assume: mxm
10
Variance of the parameter estimates for a given X More formally:
11
Example m: dimension of w determinant probability density
12
Example m: dimension of w determinant -1.5-0.500.511.5 -2 0 2 4 6 8 10 12
13
Bias of the ML estimate of Preliminaries: a. compute expected value of the residuals Preliminaries: b. compute variance of the residuals nxn matrix n:number of data points
14
Bias of the ML estimate of Preliminaries: b. compute variance of the residuals
15
Bias of the ML estimate of Preliminaries: C. useful properties of trace operator
16
Bias of the ML estimate of Remember that residuals are mean zero.
17
Bias of the ML estimate of Number of parameters in w Number of data points nxn nxm So the ML estimate of is biased, it tends to underestimate the actual variance of the noise. The bias becomes smaller as number of data points increase with respect to the number of unknown parameters.
18
When noise in each data sample has an independent variance (Midterm 2005)
19
When noise in each data sample has an independent variance This is the weighted least squares solution where each data point is weighted by the inverse of the noise.
20
Example: sensory noise may be proportional to the “signal”.
21
Summary n=Number of data points m=Number of parameters in w
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.