RLSELE 774 - Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.

Slides:



Advertisements
Similar presentations
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Advertisements

Use of Kalman filters in time and frequency analysis John Davis 1st May 2011.
AGC DSP AGC DSP Professor A G Constantinides©1 Modern Spectral Estimation Modern Spectral Estimation is based on a priori assumptions on the manner, the.
OPTIMUM FILTERING.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The FIR Adaptive Filter The LMS Adaptive Filter Stability and Convergence.
ELE Adaptive Signal Processing
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Newton’s Method Application to LMS Recursive Least Squares Exponentially-Weighted.
Lecture 11: Recursive Parameter Estimation
280 SYSTEM IDENTIFICATION The System Identification Problem is to estimate a model of a system based on input-output data. Basic Configuration continuous.
SYSTEMS Identification
Prediction and model selection
Retrieval Theory Mar 23, 2008 Vijay Natraj. The Inverse Modeling Problem Optimize values of an ensemble of variables (state vector x ) using observations:
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
EE513 Audio Signals and Systems Wiener Inverse Filter Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Adaptive Signal Processing
Normalised Least Mean-Square Adaptive Filtering
Linear Prediction Problem: Forward Prediction Backward Prediction
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Adaptive Noise Cancellation ANC W/O External Reference Adaptive Line Enhancement.
Chapter 5ELE Adaptive Signal Processing 1 Least Mean-Square Adaptive Filtering.
Constant process Separate signal & noise Smooth the data: Backward smoother: At any give T, replace the observation yt by a combination of observations.
Yuan Chen Advisor: Professor Paul Cuff. Introduction Goal: Remove reverberation of far-end input from near –end input by forming an estimation of the.
Equalization in a wideband TDMA system
Probability of Error Feature vectors typically have dimensions greater than 50. Classification accuracy depends upon the dimensionality and the amount.
Algorithm Taxonomy Thus far we have focused on:
STAT 497 LECTURE NOTES 2.
Week 2ELE Adaptive Signal Processing 1 STOCHASTIC PROCESSES AND MODELS.
Real time DSP Professors: Eng. Julian Bruno Eng. Mariano Llamedo Soria.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Least SquaresELE Adaptive Signal Processing 1 Method of Least Squares.
Method of Least Squares. Least Squares Method of Least Squares:  Deterministic approach The inputs u(1), u(2),..., u(N) are applied to the system The.
Curve-Fitting Regression
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Vector Norms and the related Matrix Norms. Properties of a Vector Norm: Euclidean Vector Norm: Riemannian metric:
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION ASEN 5070 LECTURE 11 9/16,18/09.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
CY3A2 System identification
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
EE513 Audio Signals and Systems
LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION The Minimum Variance Estimate ASEN 5070 LECTURE.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Derivation Computational Simplifications Stability Lattice Structures.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Chapter 1 Random Process
An Introduction To The Kalman Filter By, Santhosh Kumar.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Professors: Eng. Diego Barral Eng. Mariano Llamedo Soria Julian Bruno
3.7 Adaptive filtering Joonas Vanninen Antonio Palomino Alarcos.
Overview of Adaptive Filters Quote of the Day When you look at yourself from a universal standpoint, something inside always reminds or informs you that.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Autoregressive (AR) Spectral Estimation
Discrete-time Random Signals
Recursive Least-Squares (RLS) Adaptive Filters
Neural Networks 2nd Edition Simon Haykin 柯博昌 Chap 3. Single-Layer Perceptrons.
METHOD OF STEEPEST DESCENT ELE Adaptive Signal Processing1 Week 5.
ELG5377 Adaptive Signal Processing Lecture 15: Recursive Least Squares (RLS) Algorithm.
Impulse Response Measurement and Equalization Digital Signal Processing LPP Erasmus Program Aveiro 2012 Digital Signal Processing LPP Erasmus Program Aveiro.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
State-Space Recursive Least Squares with Adaptive Memory College of Electrical & Mechanical Engineering National University of Sciences & Technology (NUST)
Geology 6600/7600 Signal Analysis 26 Oct 2015 © A.R. Lowry 2015 Last time: Wiener Filtering Digital Wiener Filtering seeks to design a filter h for a linear.
DSP-CIS Part-III : Optimal & Adaptive Filters Chapter-9 : Kalman Filters Marc Moonen Dept. E.E./ESAT-STADIUS, KU Leuven
ELG5377 Adaptive Signal Processing Lecture 13: Method of Least Squares.
STATISTICAL ORBIT DETERMINATION Kalman (sequential) filter
ELG5377 Adaptive Signal Processing
Ch3: Model Building through Regression
Kalman Filtering: Control with Limited/Noisy Measurements
Modern Spectral Estimation
METHOD OF STEEPEST DESCENT
Recursively Adapted Radial Basis Function Networks and its Relationship to Resource Allocating Networks and Online Kernel Learning Weifeng Liu, Puskal.
Presentation transcript:

RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters

ELE Adaptive Signal Processing2 RLS Definition With the arrival of new data samples estimates are updated recursively. Introduce a weighting factor to the sum-of-error-squares definition Weighting factor Forgetting factor : real, positive, <1, →1 =1 → ordinary LS 1/(1- ): memory of the algorithm (ordinary LS has infinite memory) w(n) is kept fixed during the observation interval 1≤i ≤n for which the cost function  (n) is defined. two time-indices n: outer, i: inner

ELE Adaptive Signal Processing3 RLS Definition

ELE Adaptive Signal Processing4 RLS Regularisation LS cost function can be ill-posed  There is insufficient information in the input data to reconstruct the input-output mapping uniquely  Uncertainty in the mapping due to measurement noise. To overcome the problem, take ‘prior information’ into account Prewindowing is assumed!  (not the covariance method) Regularisation term Smooths and stabilises the solution  : regularisation parameter

ELE Adaptive Signal Processing5 RLS Normal Equations From method of least-squares we know that then the time-average autocorrelation matrix of the input u(n) becomes Similarly, the time-average cross-correlation vector between the tap inputs and the desired response is (unaffected from regularisation) Hence, the optimum (in the LS sense) filter coefficients should satisfy autocorrelation matrix is always non-singular due to this term. (  -1 always exists!)

ELE Adaptive Signal Processing6 RLS Recursive Computation Isolate the last term for i=n: Similarly We need to calculate  -1 to find w → direct calculation can be costly!  Use Matrix Inversion Lemma (MIL)

ELE Adaptive Signal Processing7 RLS Recursive Least-Squares Algorithm Let Then, using MIL Now, letting We obtain inverse correlation matrix gain vector Riccati equation

ELE Adaptive Signal Processing8 RLS Recursive Least-Squares Algorithm Rearranging How can w be calculated recursively? Let After substituting the recursion for P(n) into the first term we obtain But P(n)u(n)=k(n), hence

ELE Adaptive Signal Processing9 RLS Recursive Least-Squares Algorithm The term is called the a priori estimation error, Whereas the term is called the a posteriori estimation error. (Why?) Summary; the update eqn.  -1 is calculated recursively and with scalar division Initialisation: (n=0)  If no a priori information exists  gain vector a priori error regularisation parameter

ELE Adaptive Signal Processing10 RLS Recursive Least-Squares Algorithm

ELE Adaptive Signal Processing11 RLS Recursive Least-Squares Algorithm

ELE Adaptive Signal Processing12 RLS Recursion for the Sum-of-Weighted-Error-Squares From LS, we know that where Then Hence

ELE Adaptive Signal Processing13 RLS Convergence Analysis Assume stationary environment and =1 To avoid transitions, consider times n>M Assumption I: The desired response d(n) and the tap-input vector u(n) are related by the linear regression model where w o is the regression parameter vector and e o (n) is the measurement noise. The noise e o (n) is white with zero mean and variance  o 2 which makes it independent of the regressor u(n).

ELE Adaptive Signal Processing14 RLS Convergence Analysis Assumption II: The input vector u(n) is drawn from a stochastic process, which is ergodic in the autocorrelation function.  R: ensemble average,  : time average autocorrelation matrices Assumption III: The fluctuations in the weight-error vector  (n) are slow compared with those of the input signal vector u(n).  Justification:  (n) is an accumulation of the a priori error → hence, the input →Smoothing (low-pass filtering) effect.  Consequence:

ELE Adaptive Signal Processing15 RLS Convergence in Mean Value Then, Substituting into w(n) and taking the expectation, we get Applying Assumptions I and II, above expression simplifies to  biased estimate due to the initialization, but bias →0 as n→∞. =1

ELE Adaptive Signal Processing16 RLS Mean-Square Deviation Weight-error correlation matrix and invoking Assumption I and simplifying we obtain Then But, mean-square-deviation is

ELE Adaptive Signal Processing17 RLS Mean-Square Deviation Observations: Mean-Square Deviation D (n)  is proportional to the sum of reciprocal of eigenvalues of R The sensitivity of the RLS algorithm to eigenvalue spread is determined by the reciprocal of the smallest eigenvalue. ill-conditioned LS problems may lead to poor convergence behaviour.  decays almost linearly with the number of iterations w(n) converges to the Wiener solution w o as n grows. ^

ELE Adaptive Signal Processing18 RLS Ensemble-Average Learning Curve There are two error terms  A priori error,  A posteriori error, Learning curve considering  (n) yields the same general shape as that for the LMS algorithm.  Both RLS and LMS learning curves can be compared with this choice. The learning curve for RLS (a posteriori error) is We know that

ELE Adaptive Signal Processing19 RLS Ensemble-Average Learning Curve Substitution yields 1 st term (Assumption I) 2 nd term (Assumption III) 3 & 4 th terms (Assumption I)

ELE Adaptive Signal Processing20 RLS Ensemble-Average Learning Curve Combining all terms Observations  The ensemble-average learning curve of the RLS algorithm converges in about 2M iterations Typically an order of magnitude faster than LMS  As the number of iterations n→∞ the MSE J’(n) approaches the final value σ o 2 which is the variance of the measur. error e o (n). in theory RLS produces zero excess MSE!.  Convergence of the RLS algorithm in the mean square is independent of the eigenvalues of the ensemble-average correlation matrix R of the input vector u(n).

ELE Adaptive Signal Processing21 RLS Ensemble-Average Learning Curve