Algorithm Taxonomy Thus far we have focused on:

Slides:



Advertisements
Similar presentations
DSP C5000 Chapter 16 Adaptive Filter Implementation Copyright © 2003 Texas Instruments. All rights reserved.
Advertisements

ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
CHAPTER 3 CHAPTER 3 R ECURSIVE E STIMATION FOR L INEAR M ODELS Organization of chapter in ISSO –Linear models Relationship between least-squares and mean-square.
Adaptive Filters S.B.Rabet In the Name of GOD Class Presentation For The Course : Custom Implementation of DSP Systems University of Tehran 2010 Pages.
Adaptive IIR Filter Terry Lee EE 491D May 13, 2005.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The FIR Adaptive Filter The LMS Adaptive Filter Stability and Convergence.
ECE 8443 – Pattern Recognition ECE 3163 – Signals and Systems Objectives: Review Resources: Wiki: State Variables YMZ: State Variable Technique Wiki: Controllability.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Newton’s Method Application to LMS Recursive Least Squares Exponentially-Weighted.
The loss function, the normal equation,
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Artificial Intelligence Lecture 2 Dr. Bo Yuan, Professor Department of Computer Science and Engineering Shanghai Jiaotong University
280 SYSTEM IDENTIFICATION The System Identification Problem is to estimate a model of a system based on input-output data. Basic Configuration continuous.
Adaptive FIR Filter Algorithms D.K. Wise ECEN4002/5002 DSP Laboratory Spring 2003.
1 Chapter 9 Differential Equations: Classical Methods A differential equation (DE) may be defined as an equation involving one or more derivatives of an.
Adaptive Signal Processing
Normalised Least Mean-Square Adaptive Filtering
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Adaptive Noise Cancellation ANC W/O External Reference Adaptive Line Enhancement.
RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.
Chapter 5ELE Adaptive Signal Processing 1 Least Mean-Square Adaptive Filtering.

Equalization in a wideband TDMA system
Introduction to Adaptive Digital Filters Algorithms
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Introduction SNR Gain Patterns Beam Steering Shading Resources: Wiki:
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
By Asst.Prof.Dr.Thamer M.Jamel Department of Electrical Engineering University of Technology Baghdad – Iraq.
CHAPTER 4 S TOCHASTIC A PPROXIMATION FOR R OOT F INDING IN N ONLINEAR M ODELS Organization of chapter in ISSO –Introduction and potpourri of examples Sample.
4/5/00 p. 1 Postacademic Course on Telecommunications Module-3 Transmission Marc Moonen Lecture-6 Adaptive Equalization K.U.Leuven/ESAT-SISTA Module-3.
Neural Networks - Berrin Yanıkoğlu1 Applications and Examples From Mitchell Chp. 4.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Discriminant Functions
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Acoustic Noise Cancellation
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Signal and Noise Models SNIR Maximization Least-Squares Minimization MMSE.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Definitions Random Signal Analysis (Review) Discrete Random Signals Random.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Motivation Thus far we have dealt primarily with the input/output characteristics of linear systems. State variable, or state space, representations describe.
LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.
Introduction to Digital Signals
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Derivation Computational Simplifications Stability Lattice Structures.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
NCAF Manchester July 2000 Graham Hesketh Information Engineering Group Rolls-Royce Strategic Research Centre.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
ECE 8443 – Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional Likelihood Mutual Information Estimation (CMLE) Maximum MI Estimation.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Professors: Eng. Diego Barral Eng. Mariano Llamedo Soria Julian Bruno
3.7 Adaptive filtering Joonas Vanninen Antonio Palomino Alarcos.
Lecture 10b Adaptive Filters. 2 Learning Objectives  Introduction to adaptive filtering.  LMS update algorithm.  Implementation of an adaptive filter.
1. Adaptive System Identification Configuration[2] The adaptive system identification is primarily responsible for determining a discrete estimation of.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Autoregressive (AR) Spectral Estimation
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Neural Networks 2nd Edition Simon Haykin 柯博昌 Chap 3. Single-Layer Perceptrons.
METHOD OF STEEPEST DESCENT ELE Adaptive Signal Processing1 Week 5.
September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:
ECE 8443 – Pattern Recognition ECE 3163 – Signals and Systems Objectives: Frequency Response Response of a Sinusoid DT MA Filter Filter Design DT WMA Filter.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
Neural Networks Winter-Spring 2014
Chapter 10. Numerical Solutions of Nonlinear Systems of Equations
METHOD OF STEEPEST DESCENT
Presentation transcript:

LECTURE 08: LMS VARIANTS Objectives: Algorithm Taxonomy Normalized LMS Variable Adaptation Leaky LMS Sign Algorithms Smoothing Block Algorithms Volterra Filter Resources: DJ: Family of LMS Algorithms MATLAB: Leaky LMS MATLAB: Block LMS NCTU: Block LMS • URL: .../publications/courses/ece_8423/lectures/current/lecture_08.ppt • MP3: .../publications/courses/ece_8423/lectures/current/lecture_08.mp3

Algorithm Taxonomy Thus far we have focused on: Adaptive algorithm: LMS Criterion: Least-squares Iterative: Steepest descent Implementation: Tapped delay line and this doesn’t even include statistical methods for feature and model adaptation. There are many alternative optimization criteria (including ML methods). For example: Least Mean Fourth (LMF) Algorithm: N=2

Normalized LMS The stability, convergence and steady-state properties of the LMS algorithm are directly influenced by the length of the filter and the power of the input signal. We can normalize the adaptation constant with respect to these: We can estimate the power using a time average: The update equations become: We can add a small positive constant to avoid division by zero: For the homogeneous problem ( ), it can be shown that: For zero-mean Gaussian inputs, it can be shown to converge in the mean.

Variable Adaptation Rate Another variation of the LMS algorithm is to use an adaptation constant that varies as a function of time: This problem has been studied extensively in the context of neural networks. Adaptation can be implemented using a constant proportional to the “velocity” or “acceleration” of the error signal. One extension of this approach is the Variable Step algorithm: This is a generalization in that each filter coefficient has its own adaptation constant. Also, the adaptation constant is a function of time. The step sizes are determined heuristically, such as examining the sign of the instantaneous gradient:

The Leaky LMS Algorithm Consider an LMS update modified to include a constant “leakage” factor: Under what conditions would it be advantageous to employ such a factor? What are the drawbacks of this approach? Substituting our expression for e(n) in terms of d(n) and rearranging terms: Assuming independence of the filter and the data vectors: It can be shown: There is a bias from the LMS solution, R-1g, that demonstrates leakage is a compromise between bias and coefficient protection.

Sign Algorithms To decrease the computational complexity, it is also possible to update the filter coefficients using just the sign of the error: Pilot LMS, or Signed Error, or Sign Algorithm: Clipped LMS or Signed Regressor: Zero-Forcing LMS or Sign-Sign: The algorithms are useful in applications where very high-speed data is being processed, such as communications equipment. We can also derive these algorithms using a modified cost function: The Sign-Sign algorithm is used in the CCITT ADPCM standard.

Smoothing of LMS Gradient Estimates The gradient descent approach using an instantaneous sampling of the error signal is a noisy estimate. It is a tradeoff between computational simplicity and performance. We can define an Averaged LMS algorithm: A more general form of this uses a low-pass filter: The filter can be FIR or IIR. A popular variant of this approach is known as the Momentum LMS filter: A nonlinear approach, known as the Median LMS (MLMS) Algorithm: There are obviously many variants of these types of algorithms involving heuristics and higher-order approximations to the derivative.

The Block LMS (BLMS) Algorithm There is no need to update coefficients every sample. We can update coefficients every N samples and use a block update for the coefficients: The BLMS algorithm requires (NL+L) multiplications per block compared to (NL+N) per block of N points for the LMS. It is easy to show that the convergence is slowed by a factor of N for the block algorithm, but on the other hand it uses excessive smoothing. The block algorithm produces a different result than the per-sample updates (the two algorithms do not produce the same result), but the results are sufficiently close. In many real applications, block algorithms are preferred because they fit better with hardware constraints and significantly reduce computations.

The LMS Volterra Filter A second-order digital Volterra filter is defined as: where f(1)(j) are the usual linear filter coefficients, and f(2)(j1,j2) are the L(L+1)/2 quadratic coefficients. We can define the update equations in a manner completely analogous to the standard LMS filter: Not surprisingly, one form of the solution is:

The LMS Volterra Filter (Cont.) The correlation matrix, R, is defined as: The extended cross-correlation vector is given by: An adaptive second-order Volterra filter can be derived: Once the basic extended equations have been established, the filter functions much like our standard LMS adaptive filter.

Summary There are many types of adaptive algorithms. The LMS adaptive filter was have previously discussed is just one implementation. The gradient descent algorithm can be approximated in many ways. Coefficients can be updated per sample or on a block-by-block basis. Coefficients can be smoothed using a block-based expectation. Filters can exploit higher-order statistical information in the signal (e.g., the Volterra filter).