Kalman Filters and Linear Dynamical Systems and Optimal Adaptation To A Changing Body (Koerding, Tenenbaum, Shadmehr)

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Internal models, adaptation, and uncertainty
Mobile Robot Localization and Mapping using the Kalman Filter
State Estimation and Kalman Filtering CS B659 Spring 2013 Kris Hauser.
EKF, UKF TexPoint fonts used in EMF.
Modeling Uncertainty over time Time series of snapshot of the world “state” we are interested represented as a set of random variables (RVs) – Observable.
Reducing Drift in Parametric Motion Tracking
Observers and Kalman Filters
Introduction to Mobile Robotics Bayes Filter Implementations Gaussian filters.
Sam Pfister, Stergios Roumeliotis, Joel Burdick
Probabilistic Robotics: Kalman Filters
Adam Rachmielowski 615 Project: Real-time monocular vision-based SLAM.
Stanford CS223B Computer Vision, Winter 2007 Lecture 12 Tracking Motion Professors Sebastian Thrun and Jana Košecká CAs: Vaibhav Vaish and David Stavens.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
Tracking using the Kalman Filter. Point Tracking Estimate the location of a given point along a sequence of images. (x 0,y 0 ) (x n,y n )
Probabilistic Robotics
Stanford CS223B Computer Vision, Winter 2007 Lecture 12 Tracking Motion Professors Sebastian Thrun and Jana Košecká CAs: Vaibhav Vaish and David Stavens.
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Tracking with Linear Dynamic Models. Introduction Tracking is the problem of generating an inference about the motion of an object given a sequence of.
Kalman Filtering Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics TexPoint fonts used in EMF. Read.
Overview and Mathematics Bjoern Griesbach
ROBOT MAPPING AND EKF SLAM
Bayesian Filtering for Robot Localization
Statistical learning and optimal control:
/09/dji-phantom-crashes-into- canadian-lake/
Computer vision: models, learning and inference Chapter 19 Temporal models.
From Bayesian Filtering to Particle Filters Dieter Fox University of Washington Joint work with W. Burgard, F. Dellaert, C. Kwok, S. Thrun.
Jamal Saboune - CRV10 Tutorial Day 1 Bayesian state estimation and application to tracking Jamal Saboune VIVA Lab - SITE - University.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
Young Ki Baik, Computer Vision Lab.
Human-Computer Interaction Kalman Filter Hanyang University Jong-Il Park.
Statistical learning and optimal control: A framework for biological learning and motor control Lecture 4: Stochastic optimal control Reza Shadmehr Johns.
Learning Theory Reza Shadmehr Optimal feedback control stochastic feedback control with and without additive noise.
Real-Time Simultaneous Localization and Mapping with a Single Camera (Mono SLAM) Young Ki Baik Computer Vision Lab. Seoul National University.
CS Statistical Machine learning Lecture 24
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
NCAF Manchester July 2000 Graham Hesketh Information Engineering Group Rolls-Royce Strategic Research Centre.
State Estimation and Kalman Filtering Zeeshan Ali Sayyed.
Tracking with dynamics
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
Robust Localization Kalman Filter & LADAR Scans
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
Probabilistic Robotics Bayes Filter Implementations Gaussian filters.
CS b659: Intelligent Robotics
2D Motion is just the Beginning
Tracking Objects with Dynamics
Probabilistic reasoning over time
Probabilistic Robotics
PSG College of Technology
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Probabilistic Robotics
Introduction to particle filter
Filtering and State Estimation: Basic Concepts
Introduction to particle filter
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Particle Filtering.
Lecture 17 Kalman Filter.
Probabilistic Map Based Localization
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Bayes and Kalman Filter
Kalman Filtering COS 323.
Kalman Filter: Bayes Interpretation
The Discrete Kalman Filter
Probabilistic reasoning over time
Nome Sobrenome. Time time time time time time..
Probabilistic Robotics Bayes Filter Implementations FastSLAM
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Presentation transcript:

Kalman Filters and Linear Dynamical Systems and Optimal Adaptation To A Changing Body (Koerding, Tenenbaum, Shadmehr)

Tracking {Cars, people} in {video images, GPS} Observations via sensors are noisy Recover true position Temporal task Position at t is determined in part by position at t-1 face tracking demo object tracking demo multiple person tracking demo (offline demo) Coifman et al.

Bayesian Formulation Prior Likelihood Posterior Predicted position of object based on previous position and velocity Likelihood Noisy observation Posterior Bayesian update

Graphical Model Formulation With Time X1 X2 X3 Z1 Z2 Z3 Xt: state (position, velocity) at time t Zt: observation (image, GPS data) at time t P(Xt|Xt-1): state dynamics (how vehicle/person moves) P(Zt|Xt): observation dynamics (how observations are generated)

HMM Versus State Space Model X is discrete; Z is often discrete Transition distribution is look up table State-space model X and Z are continuous vectors Dynamics (Xt→Xt+1, Xt→Zt) are arbitrary Linear state-space model State-space model with linear dynamics Kalman filter Linear state-space model with additive Gaussian noise X1 X2 X3 Z1 Z2 Z3 KF magic thing from engineering, but something we can grok now that we understand probabilistic models

Kalman Filter Generative Model Dynamics: linear stochastic difference equation Observation model Noise distributions What distributions do x and z have? X1 X2 X3 Z1 Z2 Z3 U1 U2 U3

Kalman Filter Assumes known dynamics and observation model, but that can be learned as well Offline: EM algorithm On-line: Gradient ascent in log likelihood Uses Prediction Smoothing (noise reduction) Uncertainty estimation Optimal: minimizes estimation error Credit: Hal Daumé III

Kalman Filter Inference X1 X2 X3 Z1 Z2 U1 U2 U3 X1 X2 X3 Z1 Z2 Z3 U1 U2 U3 Mean of predicted Xk-1 given all observations to k-1 Mean of predicted Xk (1) P(xk | z1...zk-1): Posterior predictive distribution Distribution is Gaussian with a mean (xhat) and variance (P) * explain subscripts: k|k-1, k-1|k-1 * dynamics of update linear transform of Gaussian is Gaussian scale up the mean by a factor, scale up the variance by factor squared (2) P(xk|z1...zk): Posterior distribution where we want to get to Steps: -- compute prediction error y~ -- observed z_k minus prediction Hx -- but Hx is a distribution not a point prediction -- so error is a distribution with mean y-tilde and covariance S K_k: Kalman gain -- I'll give you intuition here, not the derivation -- reliability of observation vs. internal prediction -- determines how much we move our state in direction of observation -- intuition -- consider scalar case S^-1 = 1/(HPH+R) K = PH/(H^2P+R) if R=0 = K = 1/H (just mapping observation to state space) as R gets larger, adjustment to X gets smaller -- note that it doesn't depend on data in any way Reliability of observation vs. internal prediction

Extensions Nonlinear dynamics Switched linear dynamics Extended Kalman filter Particle filter Switched linear dynamics E.g., running vs. walking vs. climbing Z1 Z2 Z3 X1 X2 X3 S1 S2 S3

Motor Adaptation Thought experiment Walking on the moon Many reasons for changing response of muscles Fatigue Disease Exercise Development Thought experiments: imagine experience and how you would react to it. Or bionic woman

Simple Example Saccadic eye movement to target Adaptation task Motor error -> gain adjustment error

Time Scale When muscles are perturbed, what is the time scale of the perturbation? Fatigue: a few minutes Disease/injury: months Exercise: months to years Development: lifetime Any adaptation to perturbation should last only as long as the time scale of the perturbation. Key idea Need to represent perturbations at multiple time scales Terminology Perturbation = Disturbance = Deviation of gain from default (1.0)

Generative Model of Disturbances: Random Walk At Two Time Scales

Kalman Filter Model Of Adaptation X1 X2 X3 Z1 Z2 Z3 X: vector of internal gain disturbances, one per time scale Z: net gain disturbance τ : something like number of time steps for state to decay τ=1 (fast time scale) vs. τ=100 (slow time scale) Large τ -> less memory of past, more noise -> rapid increase in uncertainty over time 𝜖 𝑖 ~ 𝓝(0,𝑐 𝜏 𝑖 − 1 2 )

Credit Assignment Net observed disturbance is sum of slow and fast Why is there a tendency to move ↑ instead of →? After prolonged exposure to disturbance: shift from fast to slow scale 30% adjustment to gain: could be due to fast, to slow, or some linear combination (dashed line) [dashed line is like Weiss model: in weiss model, uncertainty about whether motion was due to x displacement, y displacement, or some combination of the two] Hazy blue is uncertainty in dashed line due to observation noise Yellow: prior belief (gaussian blur) Red: posterior (gaussian) Tendency to move up = interpretation as fast disturbance Fast disturbance can change quickly -> more likely to be responsible If weather outside is freaky today it could be due to (a) storm passing through, or (b) global warming

Saccadic Gain Adaptation Two phases of gain Adaptation (30% shift) MONKEY DATA; Thick lines are exponential fits 30 different time scales, ranging from tau=1 to tau > 10k graphs at bottom compare one slow and one fast time scale. notice posterior sliding down line between #3 and #4. note in graph #5 that posterior is along diagonal but not at origin where it started: corresponds to fast has neg perturbation, but slow has pos perturbation Maintained perturbation leads to slow shift from fast time scale to slow time scale Note that final state is not the same as initial state, even though behavior is identical.

Multiday study: 1500 trials per day, followed by dark goggles. Phase 1 Phase 2 Multiday study: 1500 trials per day, followed by dark goggles. Features: (1) Day-to-day forgetting; (2) faster relearning [*] Multiday monkey study 1500 trials per day followed by dark goggles Thick lines are exponential fits to the noisy trial-by-trial data At end of session, monkey was blindfolded and held in darkness for rest of the day points to note: forgetting (recovery) from d1 to d2, faster relearning (steepness of curve) [*] Simulation Note that at end of phase 2, short term is compensating for long term Model claim: Forgetting occurs because monkey makes eye movements without feedback → greater uncertainty over time

Double Reversal Experiment Tgt jump: +35% -35% +35% Phase 1: target jump of +35% [control period] Adaptation continues until gain reaches ~1.25 Phase 2: target jump of -35% Adaptation continues until gain returns to 1.0 Phase 3: target jump of +35% again [test period] Lines fit to data at beginning of phases 1 and 3 Despite apparent washout from phase 2, faster relearning Slow states positive; fast states negative

Double Reversal Experiment II No intrasaccadic target steps Same sequence of three stages as previous experiment, but period of darkness between 2 and 3. Monkey shows spontaneous recovery in 3d stage New condition in which 3d stage had a target jump of +0degrees Monkey shows no spontaneous recovery MODEL: at end of darkness period, pos gain for slow states; fast states had decayed. Due to lack of feedback, all states are highly uncertain Uncertainty makes learning quicker -> rapid bump up due to locking in pos gain for slow states This bump up isn’t manifested if there’s no intrasaccadic step (ISS): system quickly locks in on Gain=1. During dark period, decay of to zero at all at all time scales. Faster decay for faster time scales. Because slow time scales are positive, net increase in gain Increased uncertainty -> rapid learning

Relation To Human Memory Effect of relearning Faster to relearn material than initial acquisition Effect of spacing Material is better remembered if the time between study sessions is increased

Relevance To AI / Machine Learning Robotics Any agent dealing with complex, nonstationary environment needs to adapt continuously But also needs to remember what had been learned earlier HCI If you’re going to design a system that adapts to users, it must have memory of user at multiple time scales