Data Assimilation.

Slides:



Advertisements
Similar presentations
Introduction to Data Assimilation NCEO Data-assimilation training days 5-7 July 2010 Peter Jan van Leeuwen Data Assimilation Research Center (DARC) University.
Advertisements

Forecasting OPS 370.
Simple Regression. Major Questions Given an economic model involving a relationship between two economic variables, how do we go about specifying the.
Chapter 3 Simple Regression. What is in this Chapter? This chapter starts with a linear regression model with one explanatory variable, and states the.
Chapter 4 Multiple Regression.
Development of Empirical Models From Process Data
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
Linear and generalised linear models
UNBIASED ESTIAMTION OF ANALYSIS AND FORECAST ERROR VARIANCES
Statistical Treatment of Data Significant Figures : number of digits know with certainty + the first in doubt. Rounding off: use the same number of significant.
Linear and generalised linear models
Economics Prof. Buckles
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Chapter 2 Data Handling.
David Corne, and Nick Taylor, Heriot-Watt University - These slides and related resources:
Understanding the 3DVAR in a Practical Way
Observing Strategy and Observation Targeting for Tropical Cyclones Using Ensemble-Based Sensitivity Analysis and Data Assimilation Chen, Deng-Shun 3 Dec,
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
CSDA Conference, Limassol, 2005 University of Medicine and Pharmacy “Gr. T. Popa” Iasi Department of Mathematics and Informatics Gabriel Dimitriu University.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss High-resolution data assimilation in COSMO: Status and.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
© Department of Statistics 2012 STATS 330 Lecture 20: Slide 1 Stats 330: Lecture 20.
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
V Bandi and R Lahdelma 1 Forecasting. V Bandi and R Lahdelma 2 Forecasting? Decision-making deals with future problems -Thus data describing future must.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Quadratic Regression ©2005 Dr. B. C. Paul. Fitting Second Order Effects Can also use least square error formulation to fit an equation of the form Math.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Sampling Design and Analysis MTH 494 Lecture-21 Ossam Chohan Assistant Professor CIIT Abbottabad.
Regression Analysis: A statistical procedure used to find relations among a set of variables B. Klinkenberg G
Computacion Inteligente Least-Square Methods for System Identification.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
1 Ka-fu Wong University of Hong Kong A Brief Review of Probability, Statistics, and Regression for Forecasting.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
AP Statistics From Randomness to Probability Chapter 14.
Virtual University of Pakistan
SUR-2250 Error Theory.
US Customary Measurement System
Physics 114: Lecture 13 Probability Tests & Linear Fitting
Linear Algebra Review.
Chapter 7. Classification and Prediction
Statistical Data Analysis - Lecture /04/03
12. Principles of Parameter Estimation
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
QM222 A1 Nov. 27 More tips on writing your projects
US Customary Measurement System
Chapter 2 Minimum Variance Unbiased estimation
US Customary Measurement System
FSOI adapted for used with 4D-EnVar
Estimation Error and Portfolio Optimization
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Key Words and Introduction to Expressions
Honors Statistics From Randomness to Probability
Data Mining (and machine learning)
Error statistics in data assimilation
US Customary Measurement System
US Customary Measurement System
Integration of sensory modalities
US Customary Measurement System
Simple Linear Regression
Estimation Error and Portfolio Optimization
Psych 231: Research Methods in Psychology
Seasonal Forecasting Using the Climate Predictability Tool
Parameterizations in Data Assimilation
Measurements & Error Analysis
US Customary Measurement System
12. Principles of Parameter Estimation
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Data Assimilation

Data assimilation in different tongues “Data assimilation” - GFD “State estimation” - nonlinear dynamics “Inverse modeling” - geophysics et al. “Signal processing” – engineering “Chaos synchronization” – physics ….Well, the main reason for that is that data assimilation is a body of techniques that many fields have relied upon and developed on their own in some fashion. In fact, the term “data assimilation” itself is largely synonymous with other terms, though practitioners in other fields will sometimes try to distinguish their meaning. Don’t be fooled. For all intents and purposes, DA is equivalent to “state estimation”, “inverse modeling”, and in some instances, “signal processing.” At root, one finds that all fields are essentially blending and combining multiple sources of information to get a “best estimate” or an “analysis.” DA is generally credited with Gauss’s invention of least squares in the early 1800s, and ever since then, all such uses and instantiations of DA have generally been in accordance with some sound theoretical underpinning…. Though it is often seen as such from the outside (often as a consequence of its complicated reputation), there is no “magic” in DA! When people use DA today, they are generally doing some form of regression at some point in their process. Part of the confusion is that this regression viewpoint can be arrived at from several different starting points. At root, it is blending/combining multiple sources of information to get a “best estimate.”

Goals of data assimilation Keep a numerical model “close” to a set of observations over time Provide appropriate initial conditions for a forecast Provide an estimate of analysis errors Propagate information from observations to unobserved locations Tell us something about how the model behaves

Review: main components Relating a gridded model state to observations Introducing them over some number of times (could be as few as one) (Initialization step)

Single-observation examples Single-time updates: Cressman Statistical Weather is 4D Sequential Continuous

Weighted-average of nearby observations based on the distance First steps: Cressman Weighted-average of nearby observations based on the distance squared

Why not Cressman? If we have a preliminary estimate of the analysis with a good quality, we do not want to replace it by values provided from poor quality observations.  When going away from an observation, it is not clear how to relax the analysis toward the arbitrary state, i.e. how to decide on the shape of the function.  An analysis should respect some basic known properties of the true system, like smoothness of the fields, or relationship between the variables (e.g. hydrostatic balance). This is not guaranteed by the Cressman method: random observation errors could generate unphysical features in the analysis. No background field!

A practical route to data assimilation…statistics What temperature is it here? To consider the more popular route to data assimilation, consider the very practical question of “what temperature is it here?” Suppose you want to perform a temperature-sensitive event, and you want to know if it is actually, say, 60 degrees F or 62. You have your Mom’s beautiful, antique Galilean thermometer, and that gives you one reading, but you know it has some resolution (i.e., accuracy) issues. You invite your friend over to help measure the temperature, and he brings over his outdoor thermometer, which you have to read the number off the dial, again with resolution issues. Also, of course, no measurement is perfect, and both sources are expected to have (random) instrumental errors involved….. So, what temperature is it? Both measurements are sources of information, so how do we come up with a best estimate of the temperature that utilizes the information content in both estimates?

Form a linear combination of estimates ? With two different estimates of the same quantity, we might think it reasonable to form a linear combination of them to come up with our best estimate of the temperature, here denoted xa for “analysis,” a specialized term in data assimilation for “best estimate.” Having written our analysis like this, it remains only to find alpha -- so that we can determine whether our analysis is like this or like this. We can write it as a linear combination in the “canonical” fashion, or with some foresight, we can alternately write it in a “correction” fashion, where we are correcting the “initial guess” of x2 by the information content within x1.

Include the background as one “obs” Generalize to many observations, including the background (first guess)

Statistical assimilation Observations have errors The forecast has errors The truth xt is unknown

Statistical assimilation Best linear estimate: combination between the background xf at observation locations and the observations yo themselves Think in terms of averages We do not know the truth, so we look for the maximum likelihood estimate, or the minimum-variance estimate

Statistical assimilation We don’t know truth, so we can’t know the errors. We have at least a chance of estimating the error variances Making these estimates is the heart of statistical data assimilation

Statistical assimilation The goal can be restated: find the best α that minimizes the analysis error on average The analysis is a combination between the observation and the forecast at the observation locations

Statistical assimilation It turns out that the best estimate is achieved when: Observation increment Analysis increment

RTFDDA How does it relate to statistical assimilation?

G includes it all Distance to the obs Time from the obs Expected obs error Quality control

Obs-nudging: Weighting Functions G = Wqf Whorizontal Wvertical Wtime

The fourth dimension: time Sequential assimilation Continuous assimilation Putting it all together

Variational methods 3DVAR 4DVAR Very complex, maybe not better.

Sequential assimilation This is RTFDDA!

Complexity