JFG de Freitas, M Niranjan and AH Gee

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Bayesian Belief Propagation
Mobile Robot Localization and Mapping using the Kalman Filter
A.M. Alonso, C. García-Martos, J. Rodríguez, M. J. Sánchez Seasonal dynamic factor model and bootstrap inference: Application to electricity market forecasting.
Use of Kalman filters in time and frequency analysis John Davis 1st May 2011.
Extended Kalman Filter (EKF) And some other useful Kalman stuff!
Modeling Uncertainty over time Time series of snapshot of the world “state” we are interested represented as a set of random variables (RVs) – Observable.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Presenter: Yufan Liu November 17th,
Kalman Filter CMPUT 615 Nilanjan Ray. What is Kalman Filter A sequential state estimator for some special cases Invented in 1960’s Still very much used.
1 Vertex fitting Zeus student seminar May 9, 2003 Erik Maddox NIKHEF/UvA.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
A brief Introduction to Particle Filters
Nonlinear and Non-Gaussian Estimation with A Focus on Particle Filters Prasanth Jeevan Mary Knox May 12, 2006.
Particle Filters for Mobile Robot Localization 11/24/2006 Aliakbar Gorji Roborics Instructor: Dr. Shiri Amirkabir University of Technology.
Particle Filtering for Non- Linear/Non-Gaussian System Bohyung Han
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
Overview and Mathematics Bjoern Griesbach
ROBOT MAPPING AND EKF SLAM
Principles of the Global Positioning System Lecture 13 Prof. Thomas Herring Room A;
Helsinki University of Technology Adaptive Informatics Research Centre Finland Variational Bayesian Approach for Nonlinear Identification and Control Matti.
Muhammad Moeen YaqoobPage 1 Moment-Matching Trackers for Difficult Targets Muhammad Moeen Yaqoob Supervisor: Professor Richard Vinter.
Colorado Center for Astrodynamics Research The University of Colorado STATISTICAL ORBIT DETERMINATION Project Report Unscented kalman Filter Information.
Markov Localization & Bayes Filtering
An Application Of The Divided Difference Filter to Multipath Channel Estimation in CDMA Networks Zahid Ali, Mohammad Deriche, M. Andan Landolsi King Fahd.
Computer vision: models, learning and inference Chapter 19 Temporal models.
Particle Filtering (Sequential Monte Carlo)
Computer vision: models, learning and inference Chapter 19 Temporal models.
Kalman Filter (Thu) Joon Shik Kim Computational Models of Intelligence.
Overview Particle filtering is a sequential Monte Carlo methodology in which the relevant probability distributions are iteratively estimated using the.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2014 Professor Brandon A. Jones Lecture 18: Minimum Variance Estimator.
-Arnaud Doucet, Nando de Freitas et al, UAI
Processing Sequential Sensor Data The “John Krumm perspective” Thomas Plötz November 29 th, 2011.
An Introduction to Kalman Filtering by Arthur Pece
NCAF Manchester July 2000 Graham Hesketh Information Engineering Group Rolls-Royce Strategic Research Centre.
Paper Reading Dalong Du Nov.27, Papers Leon Gu and Takeo Kanade. A Generative Shape Regularization Model for Robust Face Alignment. ECCV08. Yan.
Sequential Monte-Carlo Method -Introduction, implementation and application Fan, Xin
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Dept. E.E./ESAT-STADIUS, KU Leuven
Short Introduction to Particle Filtering by Arthur Pece [ follows my Introduction to Kalman filtering ]
Donald Gavel, Donald Wiberg, Center for Adaptive Optics, U.C. Santa Cruz Marcos Van Dam, Lawrence Livermore National Laboaratory Towards Strehl-Optimal.
Using Kalman Filter to Track Particles Saša Fratina advisor: Samo Korpar
Tracking with dynamics
Extended Kalman Filter
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
The Unscented Kalman Filter for Nonlinear Estimation Young Ki Baik.
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
DSP-CIS Part-III : Optimal & Adaptive Filters Chapter-9 : Kalman Filters Marc Moonen Dept. E.E./ESAT-STADIUS, KU Leuven
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
A Study on Speaker Adaptation of Continuous Density HMM Parameters By Chin-Hui Lee, Chih-Heng Lin, and Biing-Hwang Juang Presented by: 陳亮宇 1990 ICASSP/IEEE.
Predictive Automatic Relevance Determination by Expectation Propagation Y. Qi T.P. Minka R.W. Picard Z. Ghahramani.
Biointelligence Laboratory, Seoul National University
Probability Theory and Parameter Estimation I
Variational filtering in generated coordinates of motion
ASEN 5070: Statistical Orbit Determination I Fall 2014
PSG College of Technology
Course: Autonomous Machine Learning
Filtering and State Estimation: Basic Concepts
Hellenic Open University
Today (2/23/16) Learning objectives:
Robust Full Bayesian Learning for Neural Networks
Extended Kalman Filter
6.891 Computer Experiments for Particle Filtering
Principles of the Global Positioning System Lecture 13
Probabilistic Modelling of Brain Imaging Data
Extended Kalman Filter
Uncertainty Propagation
Presentation transcript:

JFG de Freitas, M Niranjan and AH Gee Hierarchical Bayesian-Kalman Models for Regularization and ARD in Sequential Learning JFG de Freitas, M Niranjan and AH Gee CUED/F-INFENG/TR 307 Nov 10, 1998

Abstract Sequential Learning Hierarchical bayesian modelling : model selection, noise estimation, parameter estimation Parameter estimation : Extended Kalman Filtering Minimum variance framework Noise estimation : adaptive regularization, ARD Adaptive noise estimation = Adaptive learning rate = smoothing regularization

Introduction Sequential Learning : Smoothing constraint : Non-stationary or expensive to get before training Smoothing constraint : A priori knowledge Contribution : Adaptive filtering = Regularized error function = Adaptive Learning rates

State Space Models, Regularization and Bayesian Inference Bayesian Framework : p(wk|Yk) From uncertainty in model parameter and measurement Regularization scheme for sequential learning First order Markov Process : wk+1=wk+dk Minimum variance estimation

Hierarchical Bayesian Sequential Modeling Parameter estimation can be done with EKF in slowly changing non-stationary environments.

Kalman Filter for Param. Estimation Linear Gauss-Markov process (Linear Dynamic System) Covriance Matrix: Q, R, P Bayesian Formulation Kalman equation :based on minimum variance of P

Extended Kalman Filter Linear estimation with Taylor series expansion

Noise Estimation and Regularization Limitation of Kalman filter Fixed a priori on Process Noise Q Large Q  Large K  more sensitive to noise or outlier 3 methods of updating noise covariance Adaptive Distributed Learning rates (multiple back propagation) Sequential evidence maximization with weight decay priors Sequential evidence maximization with updated priors Descending on a landscape with numerous peaks and throughs Varying speed, smoothing landscape, jumping while descending

Adaptive Distributed Learning Rates and Kalman Filtering Get speed, lose precision Assumption: UNCORRELATED model parameters. Update by back-propagation: (Sutton 1992b) Kalman Filter Equation Why Adaptive Learning rates?

Sequential Bayesian Regularization with Weight Decay Priors (Mackay 1992, 1994b)’s gaussian approximation By taylor series approximation Iteratively update of , Update of covariance

Sequential Evidence Maximization with Sequentially Updated Priors Maximizing evidence: Prob. of residuals = evidence function Maximizing evidence leads to k+12=E[k+12] Update equation Q=qIq.

Automatic Relevance Determination (Mackay 1995) Random correlation in finite data. ARDI (Mackay 1994a, 1995) Large c in case of irrelevant input Multiple learning rates = regularization coefficients = process noise hyper-parameters

Experiment1 Problem: Results: EKFEV, EKFMAP are not good in sequential environment. LIMITATION: Weight must be converged before noise covariance can be updated

Experiment 2: (time-varying, chaotic) Problem: Results: Tradeoff between regularization and tracking: EKFQ can do this well.

Experiment 4: Pricing Financial Options Problem: five pairs of call and put option contracts on the FTSE100 index(1994/2 ~ 1994/12) Results:

Conclusions Bayesian view of Kalman filtering Bayesian inference framework Estimating Drift function? Distributed learning rates = adaptive smoothing regularizer = adaptive noise parameter Mixture of Kalman filters?