Testing modeling frameworks or… how can we improve predictive skill from our data? Mary C. Hill, University of Kansas; Emeritus, U.S. Geological Survey.

Slides:



Advertisements
Similar presentations
Ground Water Mounding & P Evaluations
Advertisements

Bridging the Gap Between Statistics and Engineering Statistical calibration of CFD simulations in Urban street canyons with Experimental data Liora Malki-Epshtein.
Objectives 10.1 Simple linear regression
Kyle Withers University of Arizona Acknowledgements: James Callegary USGS Space Grant Symposium April 18, 2009 Using Geophysical and GIS Methods to Develop.
Forecasting Using the Simple Linear Regression Model and Correlation
VII. Observation-Parameter Statistics
Model calibration using. Pag. 5/3/20152 PEST program.
This presentation can be downloaded at – This work is carried out within the SWITCH-ON.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Prediction, Correlation, and Lack of Fit in Regression (§11. 4, 11
1 Annual Report by Work Group 2 on Uncertainty Analysis and Parameter Estimation Tom Nicholson, U.S. Nuclear Regulatory Commission Mary Hill, U.S. Geological.
Simple Linear Regression and Correlation
Incorporation of Magnetic Resonance Sounding data into groundwater models through coupled and joint inversion 8th Annual Meeting of DWRIP 2014 JANUARY.
Using the Maryland Biological Stream Survey Data to Test Spatial Statistical Models A Collaborative Approach to Analyzing Stream Network Data Andrew A.
Uncertainty analysis and Model Validation.
HYDRUS_1D Sensitivity Analysis Limin Yang Department of Biological Engineering Sciences Washington State University.
The Calibration Process
Ch. 14: The Multiple Regression Model building
Simple Linear Regression Analysis
Quantify prediction uncertainty (Book, p ) Prediction standard deviations (Book, p. 180): A measure of prediction uncertainty Calculated by translating.
Inference for regression - Simple linear regression
Geology 5670/6670 Inverse Theory 26 Jan 2015 © A.R. Lowry 2015 Read for Wed 28 Jan: Menke Ch 4 (69-88) Last time: Ordinary Least Squares (   Statistics)
Calibration Guidelines 1. Start simple, add complexity carefully 2. Use a broad range of information 3. Be well-posed & be comprehensive 4. Include diverse.
Bayesian and Geostatistical Approaches to Inverse Problems Peter K. Kitanidis Civil and Environmental Engineering Stanford University.
Diploma in Statistics Introduction to Regression Lecture 2.21 Introduction to Regression Lecture Review of Lecture 2.1 –Homework –Multiple regression.
The Islamic University of Gaza Faculty of Engineering Approaches to Groundwater Modeling Conceptual Model.
IV. Sensitivity Analysis for Initial Model 1. Sensitivities and how are they calculated 2. Fit-independent sensitivity-analysis statistics 3. Scaled sensitivities.
Uncertainty Analysis and Model “Validation” or Confidence Building.
III. Ground-Water Management Problem Used for the Exercises.
Using Ground-Water Model Predictions and the ppr and opr Statistics to Guide Data Collection.
Practical Statistical Analysis Objectives: Conceptually understand the following for both linear and nonlinear models: 1.Best fit to model parameters 2.Experimental.
(Zheng and Bennett) Steps in Transport Modeling Calibration step (calibrate flow model & transport model) Adjust parameter values Traditional approach.
VI. Evaluate Model Fit Basic questions that modelers must address are: How well does the model fit the data? Do changes to a model, such as reparameterization,
GROUND-WATER AVAILABILITY Planning for Sustainable Future Water Supplies South of the C&D Canal Delaware Geological Survey – January 2007.
Calibration Guidelines 1. Start simple, add complexity carefully 2. Use a broad range of information 3. Be well-posed & be comprehensive 4. Include diverse.
V. Nonlinear Regression Objective-Function Surfaces Thus far, we have: Parameterized the forward model Obtained head and flow observations and their weights.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Onion creek catchment modeling using PIHM By, Harish Sangireddy PhD Candidate The University of Texas at Austin.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
A More Accurate and Powerful Tool for Managing Groundwater Resources and Predicting Land Subsidence: an application to Las Vegas Valley Zhang, Meijing.
17 May 2007RSS Kent Local Group1 Quantifying uncertainty in the UK carbon flux Tony O’Hagan CTCD, Sheffield.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
Mathematical Models & Optimization?
3. Scale needed for numerical accuracy Even if the grid can closely reproduce features, accurate solution can require further refinement. Generally achievable.
Calibration Guidelines 1. Start simple, add complexity carefully 2. Use a broad range of information 3. Be well-posed & be comprehensive 4. Include diverse.
Experiences in assessing deposition model uncertainty and the consequences for policy application Rognvald I Smith Centre for Ecology and Hydrology, Edinburgh.
IX. Transient Model Nonlinear Regression and Statistical Analysis.
9. Testing Model Linearity Modified Beale’s Measure (p ) Total model nonlinearity (p. 144) Intrinsic model nonlinearity (p. 145) For total and.
CE 3354 Engineering Hydrology Lecture 21: Groundwater Hydrology Concepts – Part 1 1.
Regularised Inversion and Model Predictive Uncertainty Analysis.
How Good is a Model? How much information does AIC give us? –Model 1: 3124 –Model 2: 2932 –Model 3: 2968 –Model 4: 3204 –Model 5: 5436.
The Islamic University of Gaza Faculty of Engineering Civil Engineering Department EENV 5326 Groundwater Modeling.
- 1 - Calibration with discrepancy Major references –Calibration lecture is not in the book. –Kennedy, Marc C., and Anthony O'Hagan. "Bayesian calibration.
IX. Transient Forward Modeling. Ground-Water Management Issues Recall the ground-water management issues for the simple flow system considered in the.
A Framework and Methods for Characterizing Uncertainty in Geologic Maps Donald A. Keefer Illinois State Geological Survey.
Anders Nielsen Technical University of Denmark, DTU-Aqua Mark Maunder Inter-American Tropical Tuna Commission An Introduction.
Sandeep Bisht Assistant Director Basin planning
AUTOMATED PARAMETER ESTIMATION Model Parameterization, Inverse Modeling, PEST.
Chapter 13 Simple Linear Regression
Chapter 14 Introduction to Multiple Regression
Multiple Linear Regression
Chapter 11: Simple Linear Regression
The Calibration Process
Chapter 6 Calibration and Application Process
Calibration.
BA 275 Quantitative Business Methods
Prepared by Lee Revere and John Large
Linear Model Selection and regularization
Neural networks (1) Traditional multi-layer perceptrons
Presentation transcript:

Testing modeling frameworks or… how can we improve predictive skill from our data? Mary C. Hill, University of Kansas; Emeritus, U.S. Geological Survey Laura Foglia, Technical University of Darmstadt (soon at UC Davis) Ming Ye, Florida state University Lu Dan, Oak Ridge National Laboratory

Frameworks for Model Analysis Look for surprises Surprises can indicate New understanding about reality or Model limitations or problems Fast methods are convenient -- enable routine evaluation. Computationally frugal, parallelizable methods

What do I mean by a modeling framework? Methods used to – Parameterize system characteristics – Compare to data – Adjust for data (calibration, tuning) – Sensitivity analysis – Uncertainty quantification

What do I mean by a modeling framework? Methods used to – Parameterize system characteristics Define many interpolation points. Estimate a parameter value at each point. Define zones or simple interpolations and estimate relatively few parameter values

What do I mean by a modeling framework? Methods used to – Parameterize system characteristics – Compare to data Graphical Statistics like sum of squared residuals

What do I mean by a modeling framework? Methods used to – Parameterize system characteristics – Compare to data – Adjust for data (calibration, tuning)

What do I mean by a modeling framework? Methods used to – Parameterize system characteristics – Compare to data – Adjust for data (calibration, tuning) – Sensitivity analysis Head identifier Observation importance using leverage, Cook’s D Head, in meters Foglia et al, in preparation

What do I mean by a modeling framework? Methods used to – Parameterize system characteristics – Compare to data – Adjust for data (calibration, tuning) – Sensitivity analysis – Uncertainty quantification Lu et al WRR

Overview of framework methods Problems: -Many method - Tower of Babel? -Colossal computational demands – Necessary? Hill , Groundwater

Testing modeling frameworks Test using cross-validation

Maggia Valley, Switzerland 11 Goal: Integrated hydrologic model to help manage ecology of this altered hydrologic system. Use component sw and gw models to test methods for the integrated model. 1954: Dams closed upstream

Maggia Valley, southern Switzerland Series of studies to identify and test a computationally frugal protocol for model development 1. Test frugal sensitivity analysis (SA) with cross-validation – MODFLOW Model w stream (dry) (Foglia GW) 2. Demonstrate frugal optimal calibration method – TOPKAPI Rainfall-Runoff model (Foglia WRR) 3. Here - SA and calibration methods enable test of computationally frugal model selection criteria MODFLOW GW model (wet) (Foglia WRR) 4. Integrated hydrologic model – SW and GW Software: UCODE (Poeter , 2014) MMA (MultiModel Analysis) (Poeter and Hill 2007) 12

13 Model outline and 35 Obs Locations 13 Flow data (use gains/losses) Head data

Model structures 14 Streamflow Routing (SFR2) vs. River (RIV) Recharge – Zero, constant, TOPKAPI Bedrock boundary – Old & New: Before and after gravimetric survey. – Depth in south x 2. K – Based on map of surficial quaternary geology – 5 combinations, from 1 parameter to 6 SFR RIV K Range: 5e-4 to 1e-3 64 alternative models

15 Cross validation experiment 8 heads omitted for cross-validation experiment 15 3 gains/losses omitted for cross-validation experiment Head data Flow data (use gains/losses)

16 How important are the omitted observations in the alternative models? Use SA measure leverage for models calibrated with all 35 obs. High leverage = important obs V1_diff R1_diff S1_diff S2_diff S3_diff L1_diff Observation number or name Leverage Each symbol represents a different model

17 How important are the omitted observations in the alternative models? Quickly evaluate using SA measure leverage. Identifies observations important and unimportant to parameters V1_diff R1_diff S1_diff S2_diff S3_diff L1_diff Observation number or name Leverage Omitted obs are very important to not important. Omitting these obs should provide a good test of methods of analyzing alternative models

Tests conducted Test model selection criteria AIC, BIC, and KIC – Forund problems – especially for KIC Compared results with a basic premise related to model complexity – Led to some specific suggestions about model development – Present this part of the analysis here

Our results vs basic premise? 19 Prediction error Model Fit Lines connect models that are the same except for K

Our results vs basic premise? 20 Prediction error Model Fit Lines connect models that are the same except for K Lack of improvement with added K parameters indicates the K parameterization has little relation to the processes affecting the data Solid: SFR Hollow: RIV

Our results vs basic premise? 21 Prediction error Model Fit Lines connect models that are the same except for K Solid: SFR Hollow: RIV

Our results vs basic premise? 22 Prediction error Model Fit Lines connect models that are the same except for K Solid: SFR Hollow: RIV

Our results vs basic premise? 23 Prediction error Model Fit Lines connect models that are the same except for K Sensitivity analysis indicated the models with the most parameters were poorly posed. Suggests we can use these methods to detect poor models. Solid: SFR Hollow: RIV

What do poor predictions look like?

Identify important parameters Results from the integrated hydrologic model Groundwater model + distributed hydrologic model + new observations for low flows

Soil depth Saturated K Porosity Mannings Overland Rivers B_ETP1 K KRiv Rech Identify important parameters

Soil depth Saturated K Porosity Mannings Overland Rivers B_ETP1 K KRiv Rech Identify important parameters GW observations are quite important to some of the SW parameters SW obs are interpreted as GW flow (gain/loss) obs, which are important

Conclusions from this test The surficial geologic mapping did not provide a useful way to parameterize the K Sensitivity analysis can be used to determine when the model detail exceeds what can be supported by data. Models unsupported by data are likely to have poor predictive ability 5 suggestions for model development

Testing modeling frameworks Enabled by computationally frugal methods, which are enabled by computationally robust models This is where you, the model developers come in!

Do your models have numerical daemons??? Need important nonlinearities. Keep them! Get rid of unrealistic, difficult numerics Kavetski Kuczera 2007 WRR; Hill GW; test with DELSA Rakovec 2014 WRR

Soil depth Saturated K Porosity Mannings Overland Rivers B_ETP1 K KRiv Rech Identify important parameters GW observations are quite important to some of the SW parameters SW obs are interpreted as GW flow (gain/loss) obs, which are important

Perspective In 100 years, we may very well be thought of as being in the Model T stage of model development A prediction: In the future, simply getting a close model fit will not be enough. It will also be necessary to show what obs, pars, and processes are important to the good model fit. This increased model transparency is like the safety equipment in automobile development.

I am looking for a test case Test a way to detect and diagnose numerical daemons. High profile model applications preferred.

AGU: New Technical Committee on Hydrologic Uncertainty SIAM: New SIAM-UQ section and journal

Quantifying Uncertainty Lu Ye Hill 2012 WRR Investigate methods that take 10s to 100,000s of model runs (Linear to MCMC) New UCODE_2014 with MCMC supports these ways of calculating uncertainty intervals

Test Case MODFLOW 3.8 km x 6 km x 80m Top: free-surface No-flow sides, bottom Obs: 54 heads, 2 flows, lake level, lake budget Parameters: 5 for recharge KRB confining unit KV net lake recharge, vertical anisotropy K : 1 (HO), 3 (3Z), or 21 (INT) Data on K at yellow dots True 3Z Int

Calibrated model, Predictions, Confidence Intervals HO3ZInt Standard error of regression 1.49 ( ) 1.27 ( ) 1.00 ( ) 1 Intrinsic nonlinearity RN2RN2 True Model runs 100 1, ,000 Statistics to use when the truth is unknown HO 3Z INT Name of alternative model…….. Lesson about uncertainty  May be more important to consider the uncertainty from many models than use a computationally demanding way to calculate uncertainty for one model