Least SquaresELE 774 - Adaptive Signal Processing 1 Method of Least Squares.

Slides:



Advertisements
Similar presentations
General Linear Model With correlated error terms  =  2 V ≠  2 I.
Advertisements

Chapter 28 – Part II Matrix Operations. Gaussian elimination Gaussian elimination LU factorization LU factorization Gaussian elimination with partial.
AGC DSP AGC DSP Professor A G Constantinides©1 Modern Spectral Estimation Modern Spectral Estimation is based on a priori assumptions on the manner, the.
OPTIMUM FILTERING.
ELE Adaptive Signal Processing
Solving Linear Systems (Numerical Recipes, Chap 2)
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Dan Witzner Hansen  Groups?  Improvements – what is missing?
Chapter 5 Orthogonality
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
Math for CSLecture 41 Linear Least Squares Problem Over-determined systems Minimization problem: Least squares norm Normal Equations Singular Value Decomposition.
Matrices and Systems of Equations
Ch 7.3: Systems of Linear Equations, Linear Independence, Eigenvalues
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Chapter 2 Matrices Definition of a matrix.
1 Neural Nets Applications Vectors and Matrices. 2/27 Outline 1. Definition of Vectors 2. Operations on Vectors 3. Linear Dependence of Vectors 4. Definition.
3D Geometry for Computer Graphics
Ordinary least squares regression (OLS)
Linear and generalised linear models
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Linear and generalised linear models
Basics of regression analysis
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Boot Camp in Linear Algebra Joel Barajas Karla L Caballero University of California Silicon Valley Center October 8th, 2008.
Linear Prediction Problem: Forward Prediction Backward Prediction
RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.
Chapter 5ELE Adaptive Signal Processing 1 Least Mean-Square Adaptive Filtering.
1 Statistical Analysis Professor Lynne Stokes Department of Statistical Science Lecture 5QF Introduction to Vector and Matrix Operations Needed for the.
Today Wrap up of probability Vectors, Matrices. Calculus
5  Systems of Linear Equations: ✦ An Introduction ✦ Unique Solutions ✦ Underdetermined and Overdetermined Systems  Matrices  Multiplication of Matrices.
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
SVD(Singular Value Decomposition) and Its Applications
Professor Dr. S. K. Bhattacharjee Department of Statistics
Boyce/DiPrima 9th ed, Ch 7.3: Systems of Linear Equations, Linear Independence, Eigenvalues Elementary Differential Equations and Boundary Value Problems,
ECON 1150 Matrix Operations Special Matrices
BMI II SS06 – Class 3 “Linear Algebra” Slide 1 Biomedical Imaging II Class 3 – Mathematical Preliminaries: Elementary Linear Algebra 2/13/06.
CHAPTER FIVE Orthogonality Why orthogonal? Least square problem Accuracy of Numerical computation.
CPSC 491 Xin Liu Nov 17, Introduction Xin Liu PhD student of Dr. Rokne Contact Slides downloadable at pages.cpsc.ucalgary.ca/~liuxin.
Week 2ELE Adaptive Signal Processing 1 STOCHASTIC PROCESSES AND MODELS.
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
Matrices CHAPTER 8.1 ~ 8.8. Ch _2 Contents  8.1 Matrix Algebra 8.1 Matrix Algebra  8.2 Systems of Linear Algebra Equations 8.2 Systems of Linear.
Method of Least Squares. Least Squares Method of Least Squares:  Deterministic approach The inputs u(1), u(2),..., u(N) are applied to the system The.
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Computing Eigen Information for Small Matrices The eigen equation can be rearranged as follows: Ax = x  Ax = I n x  Ax - I n x = 0  (A - I n )x = 0.
Vector Norms and the related Matrix Norms. Properties of a Vector Norm: Euclidean Vector Norm: Riemannian metric:
Solving Linear Systems of Equations
LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.
Eigenvalues The eigenvalue problem is to determine the nontrivial solutions of the equation Ax= x where A is an n-by-n matrix, x is a length n column.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Discrete-time Random Signals
Recursive Least-Squares (RLS) Adaptive Filters
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Instructor: Mircea Nicolescu Lecture 8 CS 485 / 685 Computer Vision.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
Boot Camp in Linear Algebra TIM 209 Prof. Ram Akella.
Matrices, Vectors, Determinants.
MATRICES A rectangular arrangement of elements is called matrix. Types of matrices: Null matrix: A matrix whose all elements are zero is called a null.
ELG5377 Adaptive Signal Processing Lecture 13: Method of Least Squares.
EE611 Deterministic Systems Vector Spaces and Basis Changes Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Part B. Linear Algebra, Vector Calculus
ELG5377 Adaptive Signal Processing
5 Systems of Linear Equations and Matrices
Systems of First Order Linear Equations
Chapter 10: Solving Linear Systems of Equations
Modern Spectral Estimation
Boyce/DiPrima 10th ed, Ch 7.3: Systems of Linear Equations, Linear Independence, Eigenvalues Elementary Differential Equations and Boundary Value Problems,
Singular Value Decomposition SVD
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 – 14, Tuesday 8th November 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR)
Presentation transcript:

Least SquaresELE Adaptive Signal Processing 1 Method of Least Squares

ELE Adaptive Signal Processing2 Least Squares Method of Least Squares:  Deterministic approach The inputs u(1), u(2),..., u(N) are applied to the system The outputs y(1), y(2),..., y(N) are observed  Find a model which fits the input-output relation to a (linear?) curve, f(n,u(n))  ‘best’ fit by minimising the sum of the squres of the difference f - y

ELE Adaptive Signal Processing3 Least Squares The curve fitting problem can be formulated as Error: Sum-of-error-squares: Minimum (least-squares of error) is achieved when the gradient is zero model observations variable

ELE Adaptive Signal Processing4 Least Squares Problem Statement For the inputs to the system, u(i) The observed desired response is, d(i) Relation is assumed to be linear Unobservable measurement error  Zero mean  White  Then deterministic

ELE Adaptive Signal Processing5 Least Squares Problem Statement Design a transversal filter which finds the least squares solution Then, sum of error squares is

ELE Adaptive Signal Processing6 Least Squares Data Windowing We will express the input in matrix form Depending on the limits i 1 and i 2 this matrix changes Covariance Method i 1 =M, i 2 =N Prewindowing Method i 1 =1, i 2 =N Postwindowing Method i 1 =M, i 2 =N+M1 Autocorr. Method i 1 =1, i 2 =N+M1

ELE Adaptive Signal Processing7 Least Squares Error signal Least squares (minimum of sum of squares) is achieved when i.e., when The minimum-error time series e min (i) is orthogonal to the time series of the input u(i-k) applied to tap k of a transversal filter of length M for k=0,1,...,M-1 when the filter is operating in its least-squares condition. Principle of Orthogonality !Time averaging! (For Wiener filtering) (this was ensemble average)

ELE Adaptive Signal Processing8 Least Squares Corollary of Principle of Orthogonality LS estimate of the desired response is Multiply principle of orthogonality by w k * and take summation over k Then When a transversal filter operates in its least-squares condition, the least-squares estimate of the desired response -produced at the output of the filter- and the minimum estimation error time series are orthogonal to each other over time i.

ELE Adaptive Signal Processing9 Least Squares Energy of Minimum Error Due to the principle of orthogonality, second and third terms are orthogonal, hence where, when e o (i)= 0 for all i, impossible, when the problem is underdetermined fewer data points than parameters infinitely many solutions (no unique soln.)!

ELE Adaptive Signal Processing10 Least Squares Normal Equations Hence, Expanded system of the normal equations for linear least-squares filters. Minimum error: Principle of Orthogonality →  (t,k), 0≤(t,k) ≤M-1 time-average autocorrelation function of the input z(-k), 0 ≤k ≤M-1 time-average cross-correlation bw the desired response and the input

ELE Adaptive Signal Processing11 Least Squares Normal Equations (Matrix Formulation) Matrix form of the normal equations for linear least-squares filters: Linear least-squares counterpart of the Wiener-Hopf eqn.s. Here  and z are time averages, whereas in Wiener-Hopf eqn.s they were ensemble averages. (if  -1 exists!)

ELE Adaptive Signal Processing12 Least Squares Minimum Sum of Error Squares Energy contained in the time series is Or, Then the minimum sum of error squares is

ELE Adaptive Signal Processing13 Least Squares Properties of the Time-Average Correlation Matrix  Property I: The correlation matrix  is Hermitian symmetric, Property II: The correlation matrix  is nonnegative definite, Property III: The correlation matrix  is nonsingular iff det(  ) is nonzero Property IV: The eigenvalues of the correlation matrix  are real and non-negative.

ELE Adaptive Signal Processing14 Least Squares Properties of the Time-Average Correlation Matrix  Property V: The correlation matrix  is the product of two rectangular Toeplitz matrices that are Hermitian transpose of each other.

ELE Adaptive Signal Processing15 Least Squares Normal Equations (Reformulation) But we know that which yields Substituting into the minimum sum of error squares expression gives then ! Pseudo-inverse !

ELE Adaptive Signal Processing16 Least Squares Projection The LS estimate of d is given by The matrix is a projection operator  onto the linear space spanned by the columns of data matrix A  i.e. the space U i. The orthogonal complement projector is

ELE Adaptive Signal Processing17 Least Squares Projection - Example M=2 tap filter, N=4 → N-M+1=3 Let Then And orthogonal

ELE Adaptive Signal Processing18 Least Squares Projection - Example

ELE Adaptive Signal Processing19 Least Squares Uniqueness of the LS Solution LS always has a solution, is that solution unique? The least-squares estimate is unique if and only if the nullity (the dimension of the null space) of the data matrix A equals zero. A KxM, (K=N-M+1) Solution is unique when A is of full column rank, K≥M  All columns of A are linearly independent  Overdetermined system (more eqns. than variables (taps))  (A H A) -1 nonsingular → exists and unique  Infinitely many solutions when A has linearly dependent columns, K<M  (A H A) -1 is singular

ELE Adaptive Signal Processing20 Least Squares Properties of the LS Estimates Property I: The least-squares estimate is unbiased, provided that the measurement error process e o (i) has zero mean. Property II: When the measurement error process e o (i) is white with zero mean and variance  2, the covariance matrix of the least- squares estimate equals  2  -1. Property III: When the measurement error process e o (i) is white with zero mean, the least squares estimate is the best linear unbiased estimate. Property IV: When the measurement error process e o (i) is white and Gaussian with zero mean, the least-squares estimate achieves the Cramer-Rao lower bound for unbiased estimates.

ELE Adaptive Signal Processing21 Least Squares Computation of the LS Estimates The rank (W) of an KxN (K≥N or K<N) matrix A gives  The number of linearly independent columns/rows  The number of non-zero eigenvalues/singular values The matrix is said to be full rank (full column or row rank) if  Otherwise, it is said to be rank-deficient Rank is an important parameter for matrix inversion  If K=N (square matrix) and the matrix is full rank (W=K=N) (non- singular) inverse of the matrix can be calculated, A -1 =adj(A)/det(A)  If the matrix is not square (K≠N), and/or it is rank-deficient (singular), A -1 does not exist, instead we can use the pseudo-inverse (a projection of the inverse), A +

ELE Adaptive Signal Processing22 Least Squares SVD We can calculate the pseudo-inverse using SVD. Any KxN matrix (K≥N or K<N) can be decomposed using the Singular Value Decomposition (SVD) as follows:

ELE Adaptive Signal Processing23 Least Squares SVD The system of eqn.s,  is overdetermined if K>N, more eqn.s than unknowns, Unique solution (if A is full-rank) Non-unique, infinitely many solutions (if A is rank-deficient)  is underdetermined if K<N, more unknowns than eqn.s, Non-unique, infinitely many solutions  In either case the solution(s) is(are) where

ELE Adaptive Signal Processing24 Least Squares Computation of the LS Estimates Find the solution of (A: KxM) If K>M and rank(A)=M, ( ) the unique solution is Otherwise, infinitely many solutions, but pseudo-inverse gives the minimum-norm solution to the least squares problem.  Shortest length possible in the Euclidean norm sense.

ELE Adaptive Signal Processing25 Least Squares Minimum-Norm Solution We know that Then  min is achieved when where  min is determined by c 2 (desired response, uncontrollable)  min is independent of b 2 !

ELE Adaptive Signal Processing26 Least Squares Minimum-Norm Solution Then the optimum filter coefficients become Norm of filter coeff.s is (V H V=I) which is minimum when then Even when, the vector is unique in the sense that  it is the only tap-weight vector that simultaneously satisfy Minimum sum-of-error-squares (LS solution) The smallest Euclidean norm possible.  Hence, is called the minimum-norm LS solution. ≥0