Computing a posteriori covariance in variational DA I.Gejadze, F.-X. Le Dimet, V.Shutyaev.

Slides:



Advertisements
Similar presentations
Introduction to Data Assimilation NCEO Data-assimilation training days 5-7 July 2010 Peter Jan van Leeuwen Data Assimilation Research Center (DARC) University.
Advertisements

Yi Heng Second Order Differentiation Bommerholz – Summer School 2006.
Instabilities of SVD Small eigenvalues -> m+ sensitive to small amounts of noise Small eigenvalues maybe indistinguishable from 0 Possible to remove small.
A Discrete Adjoint-Based Approach for Optimization Problems on 3D Unstructured Meshes Dimitri J. Mavriplis Department of Mechanical Engineering University.
The Inverse Regional Ocean Modeling System:
1 OR II GSLM Outline  some terminology  differences between LP and NLP  basic questions in NLP  gradient and Hessian  quadratic form  contour,
Ordinary Differential Equations
Data mining and statistical learning - lecture 6
Modern iterative methods For basic iterative methods, converge linearly Modern iterative methods, converge faster –Krylov subspace method Steepest descent.
1 L-BFGS and Delayed Dynamical Systems Approach for Unconstrained Optimization Xiaohui XIE Supervisor: Dr. Hon Wah TAM.
Numerical Optimization
Function Optimization Newton’s Method. Conjugate Gradients
Unconstrained Optimization Rong Jin. Recap  Gradient ascent/descent Simple algorithm, only requires the first order derivative Problem: difficulty in.
1/20 Accelerating minimizations in ensemble variational assimilation G. Desroziers, L. Berre Météo-France/CNRS (CNRM/GAME)
1 L-BFGS and Delayed Dynamical Systems Approach for Unconstrained Optimization Xiaohui XIE Supervisor: Dr. Hon Wah TAM.
A Concept of Environmental Forecasting and Variational Organization of Modeling Technology Vladimir Penenko Institute of Computational Mathematics and.
Retrieval Theory Mar 23, 2008 Vijay Natraj. The Inverse Modeling Problem Optimize values of an ensemble of variables (state vector x ) using observations:
ECIV 301 Programming & Graphics Numerical Methods for Engineers REVIEW II.
Advanced Topics in Optimization
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
1 Optimal solution error covariances in nonlinear problems of variational data assimilation Victor Shutyaev Institute of Numerical Mathematics, Russian.
Linear and generalised linear models
Basics of regression analysis
PETE 603 Lecture Session #29 Thursday, 7/29/ Iterative Solution Methods Older methods, such as PSOR, and LSOR require user supplied iteration.
Unconstrained Optimization Rong Jin. Logistic Regression The optimization problem is to find weights w and b that maximizes the above log-likelihood How.
Adaptive Signal Processing
RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.
Numerical Analysis 1 EE, NCKU Tien-Hao Chang (Darby Chang)
Computational Optimization
The Inverse Regional Ocean Modeling System: Development and Application to Data Assimilation of Coastal Mesoscale Eddies. Di Lorenzo, E., Moore, A., H.
Fast Low-Frequency Impedance Extraction using a Volumetric 3D Integral Formulation A.MAFFUCCI, A. TAMBURRINO, S. VENTRE, F. VILLONE EURATOM/ENEA/CREATE.
Optimization in Engineering Design Georgia Institute of Technology Systems Realization Laboratory 101 Quasi-Newton Methods.
Qualifier Exam in HPC February 10 th, Quasi-Newton methods Alexandru Cioaca.
Newton's Method for Functions of Several Variables Joe Castle & Megan Grywalski.
3D/4D-Var Methods Liang Xu (NRL) JCSDA Summer Colloquium on Satellite DA 1 3D-Var/4D-Var Solution Methods Liang Xu Naval Research Laboratory, Monterey,
Start Presentation October 4, 2012 Solution of Non-linear Equation Systems In this lecture, we shall look at the mixed symbolic and numerical solution.
Scientific Computing Partial Differential Equations Poisson Equation.
CSDA Conference, Limassol, 2005 University of Medicine and Pharmacy “Gr. T. Popa” Iasi Department of Mathematics and Informatics Gabriel Dimitriu University.
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
Stochastic Monte Carlo methods for non-linear statistical inverse problems Benjamin R. Herman Department of Electrical Engineering City College of New.
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
Tbilisi, GGSWBS'14 Optimization for inverse modelling Ketevan Kasradze 1 Hendrik Elbern 1,2
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Canonical Correlation Psy 524 Andrew Ainsworth. Matrices Summaries and reconfiguration.
Serge Andrianov Theory of Symplectic Formalism for Spin-Orbit Tracking Institute for Nuclear Physics Forschungszentrum Juelich Saint-Petersburg State University,
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION ASEN 5070 LECTURE 11 9/16,18/09.
INVERSE MODELING OF ATMOSPHERIC COMPOSITION DATA Daniel J. Jacob See my web site under “educational materials” for lectures on inverse modeling atmospheric.
© 2011 Autodesk Freely licensed for use by educational institutions. Reuse and changes require a note indicating that content has been modified from the.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Numerical Methods.
Lecture 26 Molecular orbital theory II
Errors, Uncertainties in Data Assimilation François-Xavier LE DIMET Université Joseph Fourier+INRIA Projet IDOPT, Grenoble, France.
Quality of model and Error Analysis in Variational Data Assimilation François-Xavier LE DIMET Victor SHUTYAEV Université Joseph Fourier+INRIA Projet IDOPT,
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Status of Reference Network Simulations John Dale ILC-CLIC LET Beam Dynamics Workshop 23 June 2009.
دانشگاه صنعتي اميركبير دانشكده مهندسي پزشكي استاد درس دكتر فرزاد توحيدخواه بهمن 1389 کنترل پيش بين-دکتر توحيدخواه MPC Stability-2.
Data assimilation for weather forecasting G.W. Inverarity 06/05/15.
Lecture II-3: Interpolation and Variational Methods Lecture Outline: The Interpolation Problem, Estimation Options Regression Methods –Linear –Nonlinear.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
The Ensemble Kalman filter
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Statistical Interpretation of Least Squares ASEN.
Filtering and State Estimation: Basic Concepts
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 – 14, Tuesday 8th November 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR)
Linear Algebra Lecture 16.
Pivoting, Perturbation Analysis, Scaling and Equilibration
Presentation transcript:

Computing a posteriori covariance in variational DA I.Gejadze, F.-X. Le Dimet, V.Shutyaev

What is a posteriori covariance and why one needs to compute it ? - 1 Model of evolution process Objective function (for the initial value control) Uncertainty in initial, boundary conditions, distributed coefficients, model error …. Uncertainty in the event measure - target state Measure of event Data assimilation is used to reduce uncertainty in controls / parameters, and eventually in the event measure can be defined in 2D-3D, can be a vector-function

What is a posteriori covariance and why one needs to compute it ? – 2(idea of adjoint sensitivities) Trivial relationship between (small) uncertainties in controls / parameters and in the event measure How to compute the gradient? 1. Direct method ( ) : for solve forward model with compute 2. Adjoint method: form the Lagrangian zero take variation integrate by parts If = zero and For initial value problemGenerally Gateaux derivative

What is a posteriori covariance and why one needs to compute it ? - 3 scalar - n-vector sensitivity vector Objective function - the covariance matrix of the background error or a priori covariance (measure of uncertainty in before DA) Discussion is mainly focused on the question: does the linearised error evolution model (adjoint sensitivity) approximate the non-linear model well? However, another issue is equally (if not more) important: do and how well we really now ? -must be a posteriori covariance of the estimation error (measure of uncertainty in after DA)

How to compute a-posteriori covariance?

Hessian Definitions: Gradient, where Hessian Function For large-scale problems, keeping Hessian or its inverse in the matrix form is not always feasible. If the state vector dimension is, the Hessian contains elements.

Hessian – operator-function product form 1 Control problem: Optimality system: Void optimality system for exact solution

Hessian – operator-function product form - 2 Non-linear optimality system for errors: Definition of errors: There exists a unique representation In the form of non-linear operator equation, we call - the Hessian operator. Hessian operator-function product definition: All operators are defined similarly.

Hessian – operator-function product form - 3 We do not know (and do not want to know), only. Hessian operator – is the only one to be inverted in error analysis for any inverse problem, therefore plays the key role in defining ! Since we do not store full matrices, we must find a way to define. Main Theorem Assume errors are normally distributed, unbiased and mutually uncorrelated, and, ; a) if b) if (tangent linear hypothesis is a ‘local’ sufficient condition, not necessary!

Optimal solution error covariance reminder Holds for any control or combination of controls!

Optimal solution error covariance Case I: 1) Errors are normally distributed, 2) is moderately non-linear or are small, i.e. Case II: 1) errors have arbitrary pdf, 2) ‘weak’ non-linear conditions hold a) pre-compute (define). b) produce single errors implementation using certain pdf generators. c) compute (no inversion involved)! d) generate ensemble of Case III: 1) Errors have arbitrary pdf, 2) is strongly non-linear (chaotic?) or/and are very big All as above, but compute by iterative procedure using as a pre-conditioner e) consider pdf of, which is not normal! f) use it, if you can.

Methods for computing the inverse Hessian Direct methods : 1. Fisher information matrix: Requires full matrix storage and algebraic inversion. 2. Sensitivity matrix method (A. Bennett), not suitable for large space-temporal data sets, i.e. good for 3DVar; 3. Finite difference method. Not suitable when constraints are partial differential equations due to large truncation errors; requires full matrix storage and algebraic matrix Inversion. Iterative methods: 4. Lanczos type methods Require only the product. The convergence could be erratic and slow if eigenvalues are no well separated. 5. Quasi-Newton BFGS/LBFGS (generate inverse Hessian as a collateral result) Require only the product. The convergence seems uniform. Required storage can be controlled. Exact method (J. Greenstadt, J.Bard)

Preconditioned LBFGS Solve by LBFGS the following auxiliary control problem: BFGS algorithm: - pairs of vectors to be kept in memory. LBFGS keeps only M latest pairs and the diagonal. for chosen. Retrieve pairs which define projected inverse Hessian. Then, How to get ? For example one can use Cholesky factors.

Stupid, yet 100% non-linear ensemble method 1. Consider function as the exact solution to the problem 2. Assume that and. 3. The solution of the control problem is 4. Start ensemble loop k=1,…,m. Generate, which correspond to the given pdf. Put Solve the non-linear optimization problem defined in 3). Compute. 5. End ensemble loop. 6. Compute statistic

NUMERICAL EXAMPLES: case 1 Model (non-linear convection-diffusion): - driving bc

NUMERICAL EXAMPLES: case 1 upstream boundary downstream boundary

NUMERICAL EXAMPLES: case 2

NUMERICAL EXAMPLES: case 3

SUMMARY 1.Hessian plays the crucial role in the analysis of the inverse problem solution errors as the only invertible operator; 2.If errors are normally distributed and constraints are linear the inverse Hessian is itself the covariance operator (matrix) of the optimal solution error; 3.If the problem is moderately non-linear, the inverse Hessian could be a good approximation of the optimal solution error covariance far beyond the validity of the tangent linear hypothesis. Higher order terms could be considered in problem canonical decomposition. 4. Inverse Hessian can be well approximated by a sequence of quasi-Newton updates (LBFGS) using the operator-vector product only. This sequence seems to converge uniformly; 5. Preconditioning dramatically accelerates the computing; 6. The computational cost of computing inverse Hessian should not exceed the cost of data assimilation procedure itself. 7. Inverse Hessian is useful for uncertainty analysis, experiment design, adaptive measuring techniques, etc.

PUBLICATIONS 1.I.Yu. Gejadze, V.P. Shutyaev, An optimal control problem of initial data restoration, Comput. Math. & Math. Physics, 39/9 (1999), pp F-X. Le-Dimet, V.P. Shutyaev, On deterministic error analysis in variational data assimilation, Non-linear Processes in Geophysics 14(2005), pp F.-X. Le-Dimet, V.P. Shutyaev, I.Yu. Gejadze, On optimal solution error in variational data assimilation: theoretical aspects. Russ. J. Numer. Analysis and Math. Modelling (2006), v21/2, pp I. Gejadze, F-X. Le-Dimet and V. Shutyaev. On analysis error covariances in variational data assimilation, SIAM J. Sci. Comp. (2008), v.30, no.4, I. Gejadze, F-X. Le-Dimet and V. Shutyaev. Optimal solution error covariances in variational data assimilation problems, SIAM J. Sci. Comp. (2008), to be published. Covariance matrix of the initial control problem for the diffusion equation. 3-sensor configuration