Calibration of Computer Simulators using Emulators.

Slides:



Advertisements
Similar presentations
Bayesian tools for analysing and reducing uncertainty Tony OHagan University of Sheffield.
Advertisements

Using an emulator. Outline So we’ve built an emulator – what can we use it for? Prediction What would the simulator output y be at an untried input x.
Interfacing physical experiments and computer models Preliminary remarks Tony O’Hagan.
Running a model's adjoint to obtain derivatives, while more efficient and accurate than other methods, such as the finite difference method, is a computationally.
Experiments and Variables
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Designing Ensembles for Climate Prediction
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Experimental Design, Response Surface Analysis, and Optimization
Uncertainty Analysis Using GEM-SA. GEM-SA course - session 42 Outline Setting up the project Running a simple analysis Exercise More complex analyses.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Validating uncertain predictions Tony O’Hagan, Leo Bastos, Jeremy Oakley, University of Sheffield.
Gaussian Processes I have known
Gaussian process emulation of multiple outputs Tony O’Hagan, MUCM, Sheffield.
Climate case study. Outline The challenge The simulator The data Definitions and conventions Elicitation Expert beliefs about climate parameters Expert.
Sensitivity Analysis for Complex Models Jeremy Oakley & Anthony O’Hagan University of Sheffield, UK.
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
CRASH UQ Program: Overview & Results James Paul Holloway CRASH Annual Review Fall 2010.
1 Simulation Modeling and Analysis Verification and Validation.
The Calibration Process
Chapter 6 Numerical Interpolation
1/49 EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2008 Chapter 9 Estimation: Additional Topics.
Lecture 11 Implementation Issues – Part 2. Monte Carlo Simulation An alternative approach to valuing embedded options is simulation Underlying model “simulates”
Calibration & Curve Fitting
Neural Networks Lecture 8: Two simple learning algorithms
Session 3: Calibration Using observations of the real process.
Engineering subprogramme, 7 November 2006 Tony O’Hagan.
Gaussian process modelling
Calibration and Model Discrepancy Tony O’Hagan, MUCM, Sheffield.
14/11/2000 Association of British Insurers 1 Rounding: the problems in data conversion and wrapping Mike Bell Enid Jones Association of British Insurers.
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
SCIENTIFIC METHOD CHEM 04 A series of logical steps to follow to solve problems Define the Scientific Method:
General Principle of Monte Carlo Fall 2013 By Yaohang Li, Ph.D.
CJT 765: Structural Equation Modeling Class 7: fitting a model, fit indices, comparingmodels, statistical power.
Metrology Adapted from Introduction to Metrology from the Madison Area Technical College, Biotechnology Project (Lisa Seidman)
Copyright 2012 John Wiley & Sons, Inc. Chapter 7 Budgeting: Estimating Costs and Risks.
Artificial Intelligence Techniques Multilayer Perceptrons.
Using an emulator. Outline So we’ve built an emulator – what can we use it for? Prediction What would the simulator output y be at an untried input x?
Fast Simulators for Assessment and Propagation of Model Uncertainty* Jim Berger, M.J. Bayarri, German Molina June 20, 2001 SAMO 2001, Madrid *Project of.
Statistics and the Verification Validation & Testing of Adaptive Systems Roman D. Fresnedo M&CT, Phantom Works The Boeing Company.
17 May 2007RSS Kent Local Group1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan CTCD, Sheffield.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
Adaptive Hybrid EnKF-OI for State- Parameters Estimation in Contaminant Transport Models Mohamad E. Gharamti, Johan Valstar, Ibrahim Hoteit European Geoscience.
Computing Simulation in Orders Based Transparent Parallelizing Pavlenko Vitaliy Danilovich, Odessa National Polytechnic University Burdeinyi Viktor Viktorovych,
Slide 1 Marc Kennedy, Clive Anderson, Anthony O’Hagan, Mark Lomas, Ian Woodward, Andreas Heinemayer and John Paul Gosling Quantifying uncertainty in the.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Chapter 1 Addition and Subtraction within 1, Edition.
- 1 - Calibration with discrepancy Major references –Calibration lecture is not in the book. –Kennedy, Marc C., and Anthony O'Hagan. "Bayesian calibration.
 Simulation enables the study of complex system.  Simulation is a good approach when analytic study of a system is not possible or very complex.  Informational,
1 MONALISA Compact Straightness Monitor Simulation and Calibration Week 2 Report By Patrick Gloster.
Simulation. Types of simulation Discrete-event simulation – Used for modeling of a system as it evolves over time by a representation in which the state.
Rick Walker Evaluation of Out-of-Tolerance Risk 1 Evaluation of Out-of-Tolerance Risk in Measuring and Test Equipment Rick Walker Fluke - Hart Scientific.
Introduction to emulators Tony O’Hagan University of Sheffield.
8 Sept 2006, DEMA2006Slide 1 An Introduction to Computer Experiments and their Design Problems Tony O’Hagan University of Sheffield.
Probabilistic methods for aggregate and cumulative exposure to pesticides Marc Kennedy Risk and Numerical Sciences team
Section 9.4 – Solving Differential Equations Symbolically Separation of Variables.
Marc Kennedy, Tony O’Hagan, Clive Anderson,
Fuzzy Systems Simulation Session 5
Lecture – 2 on Data structures
The Calibration Process
Essentials of Modern Business Statistics (7e)
CJT 765: Structural Equation Modeling
Copyright 2012 John Wiley & Sons, Inc. Chapter 7 Budgeting: Estimating Costs and Risks.
CSCI 5822 Probabilistic Models of Human and Machine Learning
Some General and Fundamental Requirements for Designing
6-3 Solving Systems Using Elimination
More about Posterior Distributions
Matlab.
Warmup Does the function f(x) = 3x2 + 6x have a maximum or a minimum value?
Presentation transcript:

Calibration of Computer Simulators using Emulators

Recap –Emulators We are concerned with complex, non-linear simulators In this session we will look at calibration of such simulators We will heavily depend on emulators An emulator is a Gaussian process (or second order process) that interpolates the simulator output Emulators are fast EGU short course - session 42

Calibration Simulator users often want to tune the simulator using observations of the real system Adjust the input parameters so that the simulator output matches observations as well as possible Two very important points 1. Calibration will reduce uncertainty about x but will not eliminate it 2. It is necessary to understand how the simulator relates to reality Model discrepancy EGU short course - session 43

Calibration and Assimilation Calibration is concerned with the values of the inputs that are consistent with the data. Assimilation is concerned with producing the best forecast/hindcast Calibration changes the simulator inputs Assimilation changes the simulator state variables EGU short course - session 44

Model discrepancy Simulator output y = f(x) will not equal the real system value z Even with best/correct inputs x Model discrepancy is the difference z – f(x) As discussed in Session 1, model discrepancy is due to Wrong or incomplete science Programming errors, rounding errors Inaccuracy in numerically solving systems of equations Ignoring model discrepancy leads to poor calibration Over-fitting of parameter estimates Over-confidence in the fitted values EGU short course - session 45

History matching History matching (a term taken from the petroleum industry) means finding sets of inputs that given simulator outputs that are ‘compatible’ with data Calibration means finding a best value (or a distribution) for the inputs given the data EGU short course - session 46

Implausibility Define a measure of implausibility ( I mp ) If the implausibility is greater then ±3 those values of the inputs are deemed implausible Because this is a function of the emulator not the original simulator runs we calculate it everywhere in input space EGU short course - session 47

Waves of Implausibility Wave 1: Apply the implausibility measure. Mark part of input space as implausible Wave 2: Add extra points in the not implausible region and rebuild the emulator. Repeat the implausibility measure Wave 3+: Repeat until the implausible region ceases to grow EGU short course - session 4 8

A 1-d example EGU short course - session 49

10

EGU short course - session 411

EGU short course - session 412

EGU short course - session 413

EGU short course - session 414

EGU short course - session 415

Example -Galform EGU short course - session 416

Example - Galform Galform is a simulator of Galaxy formation It has 17 inputs The amount of not implausible space in each wave is None of the original 1000 member LHC was an acceptable fit to the data EGU short course - session 417 Wave 114.9% Wave 25.9% Wave 31.6% Wave 40.26% Wave %

Calibration In history matching we were simply looking for regions of input space that were not implausible given the data. In calibration we want to find the ‘best input’ x (and it associated uncertainty) EGU short course - session 418

Kennedy and O’Hagan(2001) ζ is the real system z=ζ+ε is data on the real system ( ε~N(0,σ 2 ) ) y=f(x) is the simulator output d=ζ-y is the model discrepancy ζ=f(x)+d Build an emulator for f and simultaneously model the discrepancy as a GP ζ=f * (x)+d * z=f * (x)+d * +ε EGU short course - session 419

Kennedy and O’Hagan (2001) -2 We can now perform an uncertainty analysis This shows how much we have learned about the simulator inputs from the data The mean/mode of the posteriors give us our estimate of the best value for the inputs EGU short course - session 420

Model Discrepancy Revisited We have seen that we can use the model discrepancy to calibrate/history match the simulator We can also look at the discrepancy between different simulators This is particularly interesting if we have hierarchies of simulators EGU short course - session 421

Hierarchies of Simulators Often we have hierarchies of simulators Usually the resolution is increasing but additional processes could be added EGU short course - session 422

Hierarchies of Simulators Rather than emulate each simulator separately Emulate simulator 1 and then emulate the difference between each level Need to have some runs at common inputs Need few runs of expensive complex simulators EGU short course - session 423

Reified Simulators EGU short course - session 424

Reified Simulators EGU short course - session 425

Reified Simulators EGU short course - session 426

Reified Simulators Reified simulators are ‘imaginary’ simulators that we impose between our simulators and reality They are the ‘best’ simulator we could produce Model discrepancy is split into two: 1. The discrepancy between the current simulator and the reified simulator 2. The discrepancy between the reified simulator and reality Reification does not reduce the discrepancy. It might make it easier to elicit. EGU short course - session 427

Overview Emulators are useful tools in the calibration of complex simulators Two methods have been described: History Matching – ruling out implausible regions of input space Calibration – Finding ‘best fit’ input values Reification may be useful in eliciting the relationship between simulators and reality EGU short course - session 428