Estimator Properties and Linear Least Squares

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

The Maximum Likelihood Method
Tests of Static Asset Pricing Models
SOME GENERAL PROBLEMS.
Lecture 13 L1 , L∞ Norm Problems and Linear Programming
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Nonlinear Regression Ecole Nationale Vétérinaire de Toulouse Didier Concordet ECVPT Workshop April 2011 Can be downloaded at
The General Linear Model. The Simple Linear Model Linear Regression.
Visual Recognition Tutorial
Maximum likelihood (ML) and likelihood ratio (LR) test
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
Maximum likelihood (ML)
Parametric Inference.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Visual Recognition Tutorial
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
Rao-Cramer-Frechet (RCF) bound of minimum variance (w/o proof) Variance of an estimator of single parameter is limited as: is called “efficient” when the.
Maximum likelihood (ML)
Today Wrap up of probability Vectors, Matrices. Calculus
Slide Chapter 7 Systems and Matrices 7.1 Solving Systems of Two Equations.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 7- 1.
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
1 Standard error Estimated standard error,s,. 2 Example 1 While measuring the thermal conductivity of Armco iron, using a temperature of 100F and a power.
Lecture 16 - Approximation Methods CVEN 302 July 15, 2002.
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
2.1 – Linear and Quadratic Equations Linear Equations.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Estimation Econometría. ADE.. Estimation We assume we have a sample of size T of: – The dependent variable (y) – The explanatory variables (x 1,x 2, x.
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
Lecture 2 Linear Inverse Problems and Introduction to Least Squares.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
MathematicalMarketing Slide 3c.1 Mathematical Tools Chapter 3: Part c – Parameter Estimation We will be discussing  Nonlinear Parameter Estimation  Maximum.
Labwork 3.
Stat 223 Introduction to the Theory of Statistics
The Maximum Likelihood Method
Differential Calculus of 3D Orientations
Visual Recognition Tutorial
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
The Inverse of a Square Matrix
ICS 280 Learning in Graphical Models
CH 5: Multivariate Methods
The Maximum Likelihood Method
Review of Probability and Estimators Arun Das, Jason Rebello
The Maximum Likelihood Method
Chapter 2 Minimum Variance Unbiased estimation
Linear regression Fitting a straight line to observations.
10701 / Machine Learning Today: - Cross validation,
Summarizing Data by Statistics
Homework | Determine the Maximum Likelihood Estimator for the mean and variance of a Gaussian Distribution Jason Rebello | Waterloo Autonomous Vehicles.
Lecture 3 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
مدلسازي تجربي – تخمين پارامتر
Solving an estimation problem
Stat 223 Introduction to the Theory of Statistics
Computing and Statistical Data Analysis / Stat 7
Parametric Methods Berlin Chen, 2005 References:
Chapter 9 Chapter 9 – Point estimation
12. Principles of Parameter Estimation
Topic 11: Matrix Approach to Linear Regression
Presentation transcript:

Estimator Properties and Linear Least Squares Arun Das 06/05/2017 Arun Das | Waterloo Autonomous Vehicles Lab

Definitions| Important concepts Expected Value: Variance Cauchy–Schwarz inequality Arun Das | Waterloo Autonomous Vehicles Lab

Properties of Estimators| Bias The estimator Bias is defined as: Called unbiased if: Let’s do an example with the sample mean Arun Das | Waterloo Autonomous Vehicles Lab

Estimation| How Good Can the Estimate Be? Suppose we have an unbiased estimator Let’s manipulate the above equation a bit… Arun Das | Waterloo Autonomous Vehicles Lab

Definitions| Important concepts Expected Value: Variance Cauchy–Schwarz inequality Log identity Arun Das | Waterloo Autonomous Vehicles Lab

Cramer Rao Lower Bound| We have just derived the Cramer Rao Lower Bound Bounds the variance of the estimator as a function Of the derivative of the log likelihood function An estimator that meets this bound is called Efficient Arun Das | Waterloo Autonomous Vehicles Lab

Where is called the Fisher Information Matrix Cramer Rao Lower Bound| Can also derive multivariate case: Where is called the Fisher Information Matrix Arun Das | Waterloo Autonomous Vehicles Lab

Cramer Rao Lower Bound| Graphical Interpretation Arun Das | Waterloo Autonomous Vehicles Lab

Cramer Rao Lower Bound| Graphical Interpretation More sharpness Less Variance High Information Less sharpness High Variance Lower Information Arun Das | Waterloo Autonomous Vehicles Lab

CRLB| Graphical Interpretation Estimator Variance has inverse relationship to curvature of log likelihood! Arun Das | Waterloo Autonomous Vehicles Lab

Linear Least Squares| Problem Setup Suppose we have an overdetermined system of equations Measurement Model State Measurements Arun Das | Waterloo Autonomous Vehicles Lab

Linear Least Squares| Problem Setup No exact solution, so we want to define a cost function that allows us to best fit the model to the data Arun Das | Waterloo Autonomous Vehicles Lab

Linear Least Squares| Problem Setup No exact solution, so we want to define a cost function that allows us to best fit the model to the data Let’s pick squared deviation of the measurements from the modelled measurements Arun Das | Waterloo Autonomous Vehicles Lab

Expanding out the quadratic term: Linear Least Squares| Solution Expanding out the quadratic term: Note that: and is scalar. Thus: Arun Das | Waterloo Autonomous Vehicles Lab

Linear Least Squares| Solution Take derivative and set equal to zero: Result is least-squares normal equations Arun Das | Waterloo Autonomous Vehicles Lab

Linear Least Squares| Connection to Maximum Likelihood Recall maximum likelihood: We can model the measurements as follows: Which gives us the following likelihood function we want to maximize Arun Das | Waterloo Autonomous Vehicles Lab

Linear Least Squares| Connection to Maximum Likelihood Which gives us the following likelihood function we want to maximize Maximizing the above is equivalent to minimizing: Arun Das | Waterloo Autonomous Vehicles Lab

Linear Least Squares| Connection to Maximum Likelihood Which gives us the following likelihood function we want to maximize Maximizing the above is equivalent to minimizing: Which is also our quadratic (weighted) least squares cost! So, for Gaussian Likelihood, the least squares estimate is also the maximum likelihood estimate! Arun Das | Waterloo Autonomous Vehicles Lab

Linear Least Squares| Linear Least Squares and CRLB Possible to show that the Least Squares estimate is efficient, with information matrix: Arun Das | Waterloo Autonomous Vehicles Lab

Show that these two forms of the CRLB are equivalent Homework| Show that these two forms of the CRLB are equivalent Arun Das | Waterloo Autonomous Vehicles Lab