Donald Gavel, Donald Wiberg, Center for Adaptive Optics, U.C. Santa Cruz Marcos Van Dam, Lawrence Livermore National Laboaratory Towards Strehl-Optimal.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Properties of Least Squares Regression Coefficients
Probabilistic Reasoning over Time
The Simple Regression Model
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Use of Kalman filters in time and frequency analysis John Davis 1st May 2011.
The General Linear Model. The Simple Linear Model Linear Regression.
Session 2. Applied Regression -- Prof. Juran2 Outline for Session 2 More Simple Regression –Bottom Part of the Output Hypothesis Testing –Significance.
Visual Recognition Tutorial
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
The Simple Linear Regression Model: Specification and Estimation
280 SYSTEM IDENTIFICATION The System Identification Problem is to estimate a model of a system based on input-output data. Basic Configuration continuous.
Maximum likelihood (ML) and likelihood ratio (LR) test
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Prediction and model selection
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
Course AE4-T40 Lecture 5: Control Apllication
Linear and generalised linear models
Linear and generalised linear models
Basics of regression analysis
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Environmental Data Analysis with MatLab Lecture 7: Prior Information.
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Maximum likelihood (ML)
1 10. Joint Moments and Joint Characteristic Functions Following section 6, in this section we shall introduce various parameters to compactly represent.
Adaptive Signal Processing
Principles of the Global Positioning System Lecture 13 Prof. Thomas Herring Room A;
Principles of the Global Positioning System Lecture 11 Prof. Thomas Herring Room A;
Colorado Center for Astrodynamics Research The University of Colorado STATISTICAL ORBIT DETERMINATION Project Report Unscented kalman Filter Information.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
NSF Center for Adaptive Optics UCO Lick Observatory Laboratory for Adaptive Optics Tomographic algorithm for multiconjugate adaptive optics systems Donald.
Kalman Filter (Thu) Joon Shik Kim Computational Models of Intelligence.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Modern Navigation Thomas Herring
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Definitions Random Signal Analysis (Review) Discrete Random Signals Random.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Chapter 3: Maximum-Likelihood Parameter Estimation l Introduction l Maximum-Likelihood Estimation l Multivariate Case: unknown , known  l Univariate.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Machine Learning 5. Parametric Methods.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Computacion Inteligente Least-Square Methods for System Identification.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
The simple linear regression model and parameter estimation
Probability Theory and Parameter Estimation I
CH 5: Multivariate Methods
Hidden Markov chain models (state space model)
Filtering and State Estimation: Basic Concepts
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
Simple Linear Regression
Bayes and Kalman Filter
Principles of the Global Positioning System Lecture 11
Principles of the Global Positioning System Lecture 13
JFG de Freitas, M Niranjan and AH Gee
Parametric Methods Berlin Chen, 2005 References:
Kalman Filter: Bayes Interpretation
Presentation transcript:

Donald Gavel, Donald Wiberg, Center for Adaptive Optics, U.C. Santa Cruz Marcos Van Dam, Lawrence Livermore National Laboaratory Towards Strehl-Optimal Adaptive Optics Control

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., The goal of adaptive optics is to Maximize Strehl Max Strehl   minimize residual wavefront variance (Marechal’s aproximation) Phase correction by DM: Piston-removed atmospheric phase: vector of actuator commands vector of wavefront sensor readings actuator response functions aperture averaged residual

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Strehl-optimizing adaptive optics Define the cost function, J = mean square wavefront residual: J E is the estimation part: J C is the control part: is the conditional mean of the wavefront Wavefront estimation and control problems are separable (proven on subsequent pages): and where

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., The Conditional Mean The conditional mean is the expected value over the conditional distribution: The conditional probability distribution is defined via Bayes theorem:

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., The error in the conditional mean is uncorrelated to the data it is conditioned on: 3. The error in the conditional mean is uncorrelated to the conditional mean: 4. The error in the conditional mean is uncorrelated to the actuator commands: Properties of the conditional mean 1. The conditional mean is unbiased:

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Proof that J = J E +J C (the estimation and control problems are separable)

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., ) The conditional mean wavefront is the optimal estimate (minimizes J E ) Let for any 0 Proof: We show that any other wavefront estimate results in larger J E Therefore, minimizes J E

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., wavefront sensor operator: (average-gradient operator in the Hartmann slope sensor case) Calculating the conditional mean wavefront given wavefront sensor measurements Measurement noise For Gaussian distributed  and, it is straightforward to show (see next page) that the conditional mean of  must be a linear function of s : where since The measurement equation Cross-correlate both sides with s and solve for K so (known as the “normal” equation)

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Aside: Proof that the conditional mean is a linear function of measurements if the wavefront and measurement noise are Gaussian Bayesian conditional mean Gaussian distribution = maximum log-Likelihood of a-posteriori distribution = a linear (least squares) solution Measurement equation Measurement is a linear function of wavefront

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., ) The best-fit of the DM response functions to the conditional mean wavefront minimizes J C where and

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Comparing to Wallner’s 1 solution Combining the optimal estimator (1) and optimal controller (2) solutions gives Wallner’s “optimal correction” result: 1 E. P. Wallner, Optimal wave-front correction using slope measurements, JOSA, 73, where The two methods give the same result, a set of Strehl-optimizing actuator commands The conditional mean approach separates the problem into two independent problems: 1) statistically optimal estimation of the wavefront given noisy data 2) deterministic optimal control of the wavefront to its optimal estimate given the deformable mirror’s actuator influence functions We exploit the separation principle to derive a Strehl-optimizing closed-loop controller

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., The covariance statistics of  ( x ) (piston-removed phase over an aperture A) where

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., The g(x) function and a are “generic” under Kolmogorov statistics D  (x) = 6.88(|x|/r 0 ) 5/3 Circular aperture, diameter D Factor out parameters 6.88(D/r 0 ) 5/3 and integrals are computable numerically

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Towards a Strehl-optimizing control law for adaptive optics Remember our goal is to maximize Strehl = minimize wavefront variance in an adaptive optics system So the optimum controller uses the conditional mean, conditioned on all the previous data: But adaptive optic systems measure and control the wavefront in closed loop at sample times that are short compared to the wavefront correlation time.

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., We need to progress the conditional mean through time (the Kalman filter 2 concept) 1.Take a conditional mean at time t-1 and progress it forward to time t 2.Take data at time t 3.Instantaneously update the conditional mean, incorporating the new data 4.Progress forward to time step t+1 5.etc. 2 Kalman, R.E., A New Approach to Linear Filtering and Prediction Problems, J. Basic Eng., Trans. ASME, 82,1, 1960.

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Kalman filtering Update Time progress new data Time progress Update new data...

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Problems with calculating and progressing the conditional mean of an atmospheric wavefront through time The wavefront is defined on a Hilbert Space (continuous domain) at an infinite number of points, x  A (A = the aperture). The progression of wavefronts with time is not a well-defined process (Taylor’s frozen flow hypothesis, etc.) In addition to the estimate, the estimate’s error covariance must be updated at each time step. In the Hilbert Space, these are covariance bi-functions: c t (x,x’)=, x  A, x’  A.

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Justifying the extra effort of the optimal estimator/optimal controller If is interesting to compare “best possible” solutions to what we are getting now, with “non-optimal” controllers Determine if there is room for much improvement. Gain insights into the sensitivity of optimal solutions to modeling assumptions (e.g. knowledge of the wind, Cn2 profile, etc.) Preliminary analysis of tomographic (MCAO) reconstructors suggest that Weiner (statistically optimal) filtering may be necessary to keep the noise propagation manageable

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Updating a conditional mean given new data Say we are given a conditional mean wavefront given previous wavefront measurements And a measurement at time t The residual is uncorrelated to previous measurements, where Summarizing: Applying the normal equation on the two pieces of data e t and s t-1 : 0 0

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., …written in Wallner’s notation Estimate-update, given new data s t : Covariance-update: where the estimate error is defined: Hartmann sensor applied to the wavefront estimate Correlation of wavefront to measurement Correlation of measurement to itself

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., How it works in closed loop Wavefront sensor Best fit to DM Estimator Predictor 

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Closed-loop measurements need a correction term …since what the wavefront sensor sees is not exactly the same as s - s, the wavefront measurement prediction error DM Fitting error Measurement prediction error Measurement prediction error = Hartmann sensor residual + DM Fitting error (measured data) (can be computed from the wavefront estimate and knowledge of the DM) ^ 

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Time-progressing the conditional mean Given how do we determine ? Example 1: On a finite aperture, the phase screen is unchanging and frozen in place Estimates corrections accrue (the integrator “has a pole at zero”) If the noise covariance is non-zero, then the updates cause the estimate error covariance to decrease monotonically with t. Consequences:

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Time-progressing the conditional mean Example 2: The aperture A is infinite, and the phase screen is frozen flow, with wind velocity w Consequence: An infinite plane of phase estimates must be updated at each measurement

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Time-progressing the conditional mean Example 3: The aperture A is finite, and the phase screen is frozen flow, with wind velocity w A A’ as we might expect for x in the overlap region, A  A’ The problem is to determine the progression operator, F(x,x’), for x in the newly blown in region, A  A   A’ ) more on this approximation later

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., “Near Markov” approximation The property where w is random noise uncorrelated to  t-1 ( x ), is known as a Markov property. Phase  over the aperture however is not Markov, since some information in the “tail” portion, A’’ - (A’’  A’ ), which correlated to s t-1, is dropped off and ignored. The fractal nature of Kolmogorov statistics does not allow us to write a Markov difference equation governing  on a finite aperture. A A’ A’’ We will nevertheless proceed assuming the Markov property since the effect of neglecting  in A’’ - (A’’  A’ ) to estimates of  in A - (A  A’ ) is very small that is, the conditional mean on a finite sized aperture retains all of the relevant statistical information from the growing history of prior measurements. We see that if  obeyed a Markov property

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Validity of approximating wind-blown Kolmogorov turbulence as near-Markov contribution of neglected point in A’’ contribution of point in A’ AA’A’’ To predict this point using the estimate at this point what is the effect of neglecting this point? wind Information contained in points neglected by the near- Markov approximation is negligible

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., The progression operator from A’ to A G(x,x’’) solves We can then say that where q(x) is the error in the conditional mean  (x) -. q(x) is uncorrelated to the “data” (  (x’) ) and, consequently since the measurement at t-1 depends only on  (x’) and random measurement noise. We write the conditional mean of the wavefront in A, conditioned on knowing it in A’ Then i.e. Note: q(x) = 0 and G(x,x’) =  (x-x’-w) for x in the overlap A  A’ Also true in the overlap since q(x) = 0 there (a normal equation) A A’

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., In summary: The time-progression of the conditional mean is A A’ where F(x,x’) solves If we assume the wavefront phase covariance function is constant or slowly varying with time, then the Green’s function F(x,x’) need only be computed infrequently (e.g. in slowly varying seeing conditions) To solve this equation, we now need the cross-covariance statistics of the phase, piston-removed on two different apertures.

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Cross-covariance of Kolmogorov phase, piston-removed on two different apertures Where c and c’ are the centers of the respective apertures, and as before and A A’ also a “generic” function

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., The error covariance must also progress, since it is used in the update formulas where using the error in the conditional mean is and the error covariance is Q is defined simply to preserve the Kolmogorov turbulence strength on the subsequent aperture

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Simulations Nominal parameters –D = 3m, d = 43cm (D/d = 7) –r 0 ( =0.5  ) = 10cm ( r 0 ( =2  )  d ) –w = 11m/s  1 ms (w = D/300) –Noise = 0.1 arcsec rms Simulations –Wallner’s equations strictly applied, even though the wind is blowing –Strehl-optimal controller –Optimal controller with update matrix, K, set at converged value (allows pre- computing error covariances) –Sensitivity to assumed r0 –Sensitivity to assumed wind speed –Sensitivity to assumed wind direction

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Noise performance after convergence Strehl-optimal Single-step (Wallner)

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Convergence time history K matrix fixed at converged value K matrix optimal at each time step

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Sensitivity to r 0

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Sensitivity to wind speed and direction

IPAM Workshop on Estimation and Control Problems in Adaptive Optics, Jan., Conclusions Kalman filtering techniques can be applied to better optimize the closed-loop Strehl of adaptive optics wavefront controllers A-priori knowledge of r 0 and wind velocity is required Simulations show –Considerable improvement in performance over a single step optimized control law (Wallner) –Insensitivity to the exact knowledge of the seeing parameters over reasonably practical variations in these parameters