Normalised Least Mean-Square Adaptive Filtering

Slides:



Advertisements
Similar presentations
AGC DSP AGC DSP Professor A G Constantinides©1 Modern Spectral Estimation Modern Spectral Estimation is based on a priori assumptions on the manner, the.
Advertisements

ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The FIR Adaptive Filter The LMS Adaptive Filter Stability and Convergence.
ELE Adaptive Signal Processing
AGC DSP AGC DSP Professor A G Constantinides©1 A Prediction Problem Problem: Given a sample set of a stationary processes to predict the value of the process.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Newton’s Method Application to LMS Recursive Least Squares Exponentially-Weighted.
Visual Recognition Tutorial
1/44 1. ZAHRA NAGHSH JULY 2009 BEAM-FORMING 2/44 2.
280 SYSTEM IDENTIFICATION The System Identification Problem is to estimate a model of a system based on input-output data. Basic Configuration continuous.
SYSTEMS Identification
Prediction and model selection
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
1 Adjoint Method in Network Analysis Dr. Janusz A. Starzyk.
Frequency-Domain Adaptive Filters Wu, Yihong EE 491D
Maximum likelihood (ML)
EE513 Audio Signals and Systems Wiener Inverse Filter Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Adaptive Signal Processing
Linear Prediction Problem: Forward Prediction Backward Prediction
RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.
Chapter 5ELE Adaptive Signal Processing 1 Least Mean-Square Adaptive Filtering.
Constant process Separate signal & noise Smooth the data: Backward smoother: At any give T, replace the observation yt by a combination of observations.
Equalization in a wideband TDMA system
Algorithm Taxonomy Thus far we have focused on:
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Introduction to Adaptive Digital Filters Algorithms
SVM by Sequential Minimal Optimization (SMO)
1 CS 551/651: Structure of Spoken Language Lecture 8: Mathematical Descriptions of the Speech Signal John-Paul Hosom Fall 2008.
By Asst.Prof.Dr.Thamer M.Jamel Department of Electrical Engineering University of Technology Baghdad – Iraq.
Simplex method (algebraic interpretation)
1 BIEN425 – Lecture 8 By the end of the lecture, you should be able to: –Compute cross- /auto-correlation using matrix multiplication –Compute cross- /auto-correlation.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Least SquaresELE Adaptive Signal Processing 1 Method of Least Squares.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
Unit-V DSP APPLICATIONS. UNIT V -SYLLABUS DSP APPLICATIONS Multirate signal processing: Decimation Interpolation Sampling rate conversion by a rational.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION ASEN 5070 LECTURE 11 9/16,18/09.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
EE513 Audio Signals and Systems
LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.
CHAPTER 5 SIGNAL SPACE ANALYSIS
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Derivation Computational Simplifications Stability Lattice Structures.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Dept. E.E./ESAT-STADIUS, KU Leuven
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Professors: Eng. Diego Barral Eng. Mariano Llamedo Soria Julian Bruno
DSP-CIS Part-IV : Filter Banks & Subband Systems Chapter-12 : Frequency Domain Filtering Marc Moonen Dept. E.E./ESAT-STADIUS, KU Leuven
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Autoregressive (AR) Spectral Estimation
Baseband Receiver Receiver Design: Demodulation Matched Filter Correlator Receiver Detection Max. Likelihood Detector Probability of Error.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Neural Networks 2nd Edition Simon Haykin 柯博昌 Chap 3. Single-Layer Perceptrons.
METHOD OF STEEPEST DESCENT ELE Adaptive Signal Processing1 Week 5.
Impulse Response Measurement and Equalization Digital Signal Processing LPP Erasmus Program Aveiro 2012 Digital Signal Processing LPP Erasmus Program Aveiro.
Geology 6600/7600 Signal Analysis 26 Oct 2015 © A.R. Lowry 2015 Last time: Wiener Filtering Digital Wiener Filtering seeks to design a filter h for a linear.
DSP-CIS Part-III : Optimal & Adaptive Filters Chapter-9 : Kalman Filters Marc Moonen Dept. E.E./ESAT-STADIUS, KU Leuven
1 Chapter 8 The Discrete Fourier Transform (cont.)
DIGITAL SIGNAL PROCESSING ELECTRONICS
Adaptive Filters Common filter design methods assume that the characteristics of the signal remain constant in time. However, when the signal characteristics.
Equalization in a wideband TDMA system
Modern Spectral Estimation
Equalization in a wideband TDMA system
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
By Viput Subharngkasen
METHOD OF STEEPEST DESCENT
Finite Wordlength Effects
Support Vector Machines
Simplex method (algebraic interpretation)
Presentation transcript:

Normalised Least Mean-Square Adaptive Filtering (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing LMS Filtering The update equation for the LMS algorithm is which is derived from SD as an approximation where the step size  is originally considered for a deterministic gradient. LMS suffers from gradient noise due to its random nature. Above update is problematic due to this noise Gradient noise amplification when ||u(n)|| is large. Step size Error signal Filter input Step size (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Normalised LMS u(n) is random → instantaneous samples can assume any value for the norm ||u(n)|| which can be very large. Solution: input samples can be forced to have constant norm Normalisation Update equation for the normalised LMS algorithm. Note the similarity bw. NLMS and LMS update eqn.s NLMS can be considered same as LMS except time-varying step size. (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Normalised LMS Block diagram very similar to that of LMS The difference is in the Weight-Control Mechanism block. (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Normalised LMS We have seen that LMS algorithm optimises the H∞ criterion instead of MSE. Similarly, NLMS optimises another problem: From one iteration to the next, the weight vector of an adaptive filter should be changed in a minimal manner, subject to a constraint imposed on the updated filter’s output. Mathematically, which can be optimised by the method Lagrange multipliers (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Normalised LMS 1. Take the first derivative of J(n) wrt and set to zero to find 2. Substitute this result into the constraint to solve for the multiplier 3. Combining these results and adding a step-size parameter to control the progress gives 4. Hence the update eqn. for NLMS becomes (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Normalised LMS Observations: We may view an NLMS filter as an LMS filter with a time-varying step-size parameter Rate of convergence of NLMS is faster than LMS ||u(n)|| can be very large, however, likewise it can also be very small Causes problem since it appears in the denominator Solution: include a small correction term to avoid stability problems. (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Normalised LMS (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Stability of NLMS What should be the value of step size  for convergence? Assume that the desired response is governed by Substituting the weight-error vector into the NLMS update equation we get which provides the update for the mean-square deviation Undisturbed error signal (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Stability of NLMS Find the range for  so that Right hand side is a quadratic function of , is satisfied when Differentiate wrt and equate to 0 to find opt This step size yields maximum drop in the MSD! For clarity of notation assume real-valued signals (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Stability of NLMS Assumption I: The fluctuations in the input signal energy ||u(n)||2 from one iteration to the next are small enough so that Then Assumption II: Undisturbed error signal u(n) is uncorrelated with the disturbance noise (n) e(n): observable, u(n): unobservable (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Stability of NLMS Assumption III: The spectral content of the input signal u(n) is essentially flat over a frequency band larger than that occupied by each element of the weight-error vector (n) , hence Then (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Block LMS In conventional LMS, filter coefficients are updated for each sample What happens if we update the filter in every L samples? (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Block LMS Express the sample time n in terms of the block index k Stack the L consecutive samples of the input signal vector u(n) into a matrix corresponding to the k-th block where the whole k-th block will be processed by the filter i=0 i=1 i=L-1 convolution (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Block LMS Filter length: M=6 Block size: L=4 The output of the filter is And the error signal is Error is generated for every sample in a block! (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Block LMS Example: M=L=3 (k-1), k, (k+1)th block (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Block LMS Algorithm Conventional LMS algorithm is For a block of length L, w(k) is fixed, however, for every sample in a block we obtain separate error signals e(n). How can we link these two? Sum the product , i.e. where (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Block LMS Algorithm Then the estimate of the gradient becomes And the block LMS update eqn. İs where Block LMS step-size Block size LMS step-size (Fast) Block LMS ELE 774 - Adaptive Signal Processing

Convergence of Block LMS Conventional LMS Block LMS Main difference is the sample averaging in Block LMS yields better estimate of the gradient vector. Convergence rate is similiar to conventional LMS, not faster Block LMS requires more samples for the ‘better’ gradient estimate Same analysis done for conventional LMS can also be applied here. Small-step size analysis (Fast) Block LMS ELE 774 - Adaptive Signal Processing

Convergence of Block LMS Average time constant if B<1/max Misadjustment same as LMS same as LMS (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Choice of Block Size? Block LMS introduces processing delay Results are obtained at every L samples, L can be >>1 What should be length of a block? L=M: optimal choice from the viewpoint of computational complexity. L<M: reduces processing delay, although not optimal, better computational efficiency wrt conventional LMS L>M: redundant operations in the adaptation, estimation of the gradient uses more information than the filter itself. (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Fast-Block LMS Correlation is equivalent to convolution when one of the sequences is order reversed. Linear convolution can be effectively computed using FFT Overlap-Add, Overlap-Save methods A natural extension of Block LMS is to use FFT Let block size be equal to the filter length, L=M, Use Overlap-Save method with an FFT size of N=2M. (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Fast-Block LMS Using these two in the convolution, for the k-th block The Mx1 desired response vector is and the Mx1 error vector is (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Fast-Block LMS Correlation is convolution with one of the seq.s order reversed: Then the update equation becomes (in the frequency domain) Computational Complexity: Conventional LMS: requires 2M multiplications per sample 2M2 multiplications per block (of length M) Fast-Block LMS: 1 FFT = N log2(N) real multiplications (N=2M) 5 (I)FFTs, UH(k)E(k): 4N multiplications → Total: 10Mlog2M+28M mult.s For M=1024, Fast Block LMS is 16 times faster than conventional LMS (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Fast-Block LMS Same step size for all frequency bins (of FFT). Rate of convergence can be improved by assigning separate step-size parameters to every bin : constant, Pi: estimate of the average power in the i-th freq. Bin Assumes wss. environment. If not wss., use the recursion i-th input of the Fast-LMS algorithm (freq. dom.) for the k-th block (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing Fast-Block LMS Run the iterations for all blocks to obtain Then the step size parameter  can be replace by the matrix  where Update the Fast-LMS algorithm as follows: 1. 2. replace  by  in the update eqn. (Fast) Block LMS ELE 774 - Adaptive Signal Processing

ELE 774 - Adaptive Signal Processing ’ (Fast) Block LMS ELE 774 - Adaptive Signal Processing