CHAPTER 3 CHAPTER 3 R ECURSIVE E STIMATION FOR L INEAR M ODELS Organization of chapter in ISSO –Linear models Relationship between least-squares and mean-square.

Slides:



Advertisements
Similar presentations
Chapter 16 Adaptive Filters
Advertisements

Slides from: Doug Gray, David Poole
Linear Regression.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
CHAPTER 13 M ODELING C ONSIDERATIONS AND S TATISTICAL I NFORMATION “All models are wrong; some are useful.”  George E. P. Box Organization of chapter.
Pattern Recognition and Machine Learning
Adaptive Filters S.B.Rabet In the Name of GOD Class Presentation For The Course : Custom Implementation of DSP Systems University of Tehran 2010 Pages.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The FIR Adaptive Filter The LMS Adaptive Filter Stability and Convergence.
Chapter 2: Lasso for linear models
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Newton’s Method Application to LMS Recursive Least Squares Exponentially-Weighted.
The loss function, the normal equation,
CHAPTER 16 MARKOV CHAIN MONTE CARLO
The Simple Linear Regression Model: Specification and Estimation
Performance Optimization
Least-Mean-Square Algorithm CS/CMPE 537 – Neural Networks.
Stochastic Approximation Neta Shoham. References This Presentation is totally based on the book Introduction to Stochastic Search and Optimization (2003)
SYSTEMS Identification
Goals of Adaptive Signal Processing Design algorithms that learn from training data Algorithms must have good properties: attain good solutions, simple.
Prediction and model selection
EE491D Special Topics in Communications Adaptive Signal Processing Spring 2005 Prof. Anthony Kuh POST 205E Dept. of Elec. Eng. University of Hawaii Phone:
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
Adaptive Signal Processing
Normalised Least Mean-Square Adaptive Filtering
RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.
Chapter 5ELE Adaptive Signal Processing 1 Least Mean-Square Adaptive Filtering.

Collaborative Filtering Matrix Factorization Approach
Equalization in a wideband TDMA system
Least-Squares Regression
Algorithm Taxonomy Thus far we have focused on:
Introduction to Adaptive Digital Filters Algorithms
CHAPTER 17 O PTIMAL D ESIGN FOR E XPERIMENTAL I NPUTS Organization of chapter in ISSO* –Background Motivation Finite sample and asymptotic (continuous)
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
By Asst.Prof.Dr.Thamer M.Jamel Department of Electrical Engineering University of Technology Baghdad – Iraq.
CHAPTER 4 S TOCHASTIC A PPROXIMATION FOR R OOT F INDING IN N ONLINEAR M ODELS Organization of chapter in ISSO –Introduction and potpourri of examples Sample.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
CHAPTER 5 S TOCHASTIC G RADIENT F ORM OF S TOCHASTIC A PROXIMATION Organization of chapter in ISSO –Stochastic gradient Core algorithm Basic principles.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION ASEN 5070 LECTURE 11 9/16,18/09.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Derivation Computational Simplifications Stability Lattice Structures.
CHAPTER 6 STOCHASTIC APPROXIMATION AND THE FINITE-DIFFERENCE METHOD
CHAPTER 17 O PTIMAL D ESIGN FOR E XPERIMENTAL I NPUTS Organization of chapter in ISSO –Background Motivation Finite sample and asymptotic (continuous)
ADALINE (ADAptive LInear NEuron) Network and
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
DSP-CIS Chapter-13: Least Mean Squares (LMS) Algorithm Marc Moonen Dept. E.E./ESAT-STADIUS, KU Leuven
Adaptive Control Loops for Advanced LIGO
Overview of Adaptive Filters Quote of the Day When you look at yourself from a universal standpoint, something inside always reminds or informs you that.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Object Tracking - Slide 1 Object Tracking Computer Vision Course Presentation by Wei-Chao Chen April 05, 2000.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Neural Networks 2nd Edition Simon Haykin 柯博昌 Chap 3. Single-Layer Perceptrons.
METHOD OF STEEPEST DESCENT ELE Adaptive Signal Processing1 Week 5.
CY3A2 System identification Input signals Signals need to be realisable, and excite the typical modes of the system. Ideally the signal should be persistent.
1 Consensus-Based Distributed Least-Mean Square Algorithm Using Wireless Ad Hoc Networks Gonzalo Mateos, Ioannis Schizas and Georgios B. Giannakis ECE.
Geology 6600/7600 Signal Analysis 26 Oct 2015 © A.R. Lowry 2015 Last time: Wiener Filtering Digital Wiener Filtering seeks to design a filter h for a linear.
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
DSP-CIS Part-III : Optimal & Adaptive Filters Chapter-9 : Kalman Filters Marc Moonen Dept. E.E./ESAT-STADIUS, KU Leuven
Dynamic Models, Autocorrelation and Forecasting
CHAPTER 3 RECURSIVE ESTIMATION FOR LINEAR MODELS
Collaborative Filtering Matrix Factorization Approach
لجنة الهندسة الكهربائية
Instructor :Dr. Aamer Iqbal Bhatti
METHOD OF STEEPEST DESCENT
The loss function, the normal equation,
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Presentation transcript:

CHAPTER 3 CHAPTER 3 R ECURSIVE E STIMATION FOR L INEAR M ODELS Organization of chapter in ISSO –Linear models Relationship between least-squares and mean-square –LMS and RLS estimation Applications in adaptive control –LMS, RLS, and Kalman filter for time-varying solution –Case study: Oboe reed data Slides for Introduction to Stochastic Search and Optimization (ISSO) by J. C. Spall

3-2 Basic Linear Model Consider estimation of vector  in model that is linear in  Model has classical linear form where z k is kth measurement, h k is corresponding “design vector,” and v k is unknown noise value Model used extensively in control, statistics, signal processing, etc. Many estimation/optimization criteria based on “squared- error”-type loss functions –Leads to criteria that are quadratic in  –Unique (global) estimate 

3-3 Least-Squares Estimation Most common method for estimating  in linear model is by method of least squares Criterion (loss function) has form where Z n = [z 1, z 2,…, z n ] T and H n is n  p concatenated matrix of h k T row vectors batchClassical batch least-squares estimate is recursivePopular recursive estimates (LMS, RLS, Kalman filter) may be derived from batch estimate

3-4 Geometric Interpretation of Least-Squares Estimate when p = 2 and n = 3

3-5 Recursive Estimation Batch form not convenient in many applications –E.g., data arrive over time and want “easy” way to update estimate at time k to estimate at time k+1 Least-mean-squares (LMS) method is very popular recursive method –Stochastic analogue of steepest descent algorithm LMS recursion: Convergence theory based on stochastic approximation (e.g., Ljung, et al., 1992; Gerencsér, 1995) –Less rigorous theory based on connections to steepest descent (ignores noise) (Widrow and Stearns, 1985; Haykin, 1996)

3-6 LMS in Closed-Loop Control Suppose process is modeled according to autoregressive (AR) form: where x k represents state,  and  i are unknown parameters, u k is control, and w k is noise Let target (“desired”) value for x k be d k Optimal control law known (minimizes mean-square tracking error): estimatesCertainty equivalence principle justifies substitution of parameter estimates for unknown true parameters –LMS used to estimate  and  i in closed-loop mode

3-7 LMS in Closed-Loop Control for First-Order AR Model

3-8 Recursive Least Squares (RLS) Alternative to LMS is RLS –Recall LMS is stochastic analogue of steepest descent (“first order” method) –RLS is stochastic analogue of Newton-Raphson (“second order” method)  faster convergence than LMS in practice RLS algorithm (2 recursions): Need P 0 and to initialize RLS recursions

3-9 Recursive Methods for Estimation of Time- Varying Parameters It is common to have the underlying true  evolve in time (e.g., target tracking, adaptive control, sequential experimental design, etc.) –Time-varying parameters implies  replaced with  k Consider modified linear model Prototype recursive form for estimating  k is where choice of A k and  k depends on specific algorithm

3-10 Three Important Algorithms for Estimation of Time-Varying Parameters LMS LMS –Goal is to minimize instantaneous squared-error criteria across iterations –General form for evolution of true parameters  k RLS RLS –Goal is to minimize weighted sum of squared errors –Sum criterion creates “inertia” not present in LMS –General form for evolution of  k Kalman filter Kalman filter –Minimizes instantaneous squared-error criteria –Requires precise statistical description of evolution of  k via state-space model Details for above algorithms in terms of prototype algorithm (previous slide) are in Section 3.3 of ISSO

3-11 Case Study: LMS and RLS with Oboe Reed Data …an ill wind that nobody blows good. —Comedian Danny Kaye in speaking of the oboe in the “The Secret Life of Walter Mitty” (1947) Section 3.4 of ISSO reports on linear and curvilinear models for predicting quality of oboe reeds –Linear model has 7 parameters; curvilinear has 4 parameters This study compares LMS and RLS with batch least- squares estimates reeddata-fit reeddata- test –160 data points for fitting models ( reeddata-fit ); 80 (independent) data points for testing models ( reeddata- test ) –reeddata-fitreeddata-test –reeddata-fit and reeddata-test data sets available from ISSO Web site

Oboe with Attached Reed

3-13 Comparison of Fitting Results for reeddata-fit and reeddata-test To test similarity of fit and test data sets, performed model fitting using test data set This comparison is for checking consistency of the two data sets; not for checking accuracy of LMS or RLS estimates Compared model fits for parameters in –Basic linear model (eqn. (3.25) in ISSO) (p = 7) –Curvilinear model (eqn. (3.26) in ISSO) (p = 4) Results on next slide for basic linear model

3-14 Comparison of Batch Parameter Estimates for Basic Linear Model. Approximate 95% Confidence Intervals Shown in [·, ·]

3-15 Comparison of Batch and RLS with Oboe Reed Data Compared batch and RLS using 160 data points in reeddata-fit and 80 data points for testing models in reeddata-test Two slides to follow present results –First slide compares parameter estimates in pure linear model –Second slide compares prediction errors for linear and curvilinear models

3-16 Batch and RLS Parameter Estimates for Basic Linear Model (Data from reeddata-fit )

3-17 Mean and Median Absolute Prediction Errors for the Linear and Curvilinear Models (Model fits from reeddata-fit; Prediction Errors from reeddata-test)