Validation Dr Andy Evans. Preparing to model Verification Calibration/Optimisation Validation Sensitivity testing and dealing with error.

Slides:



Advertisements
Similar presentations
The Research Consumer Evaluates Measurement Reliability and Validity
Advertisements

Testing Theories: Three Reasons Why Data Might not Match the Theory.
Materials for Lecture 11 Chapters 3 and 6 Chapter 16 Section 4.0 and 5.0 Lecture 11 Pseudo Random LHC.xls Lecture 11 Validation Tests.xls Next 4 slides.
Modeling and Simulation Monte carlo simulation 1 Arwa Ibrahim Ahmed Princess Nora University.
Propagation of Error Ch En 475 Unit Operations. Quantifying variables (i.e. answering a question with a number) 1. Directly measure the variable. - referred.
1 Summarizing Performance Data Confidence Intervals Important Easy to Difficult Warning: some mathematical content.
Correlation and Autocorrelation
Statistics & Modeling By Yan Gao. Terms of measured data Terms used in describing data –For example: “mean of a dataset” –An objectively measurable quantity.
Determining the # Of PCs Remembering the process Some cautionary comments Statistical approaches Mathematical approaches “Nontrivial factors” approaches.
1 Validation and Verification of Simulation Models.
Simple Correlation Scatterplots & r Interpreting r Outcomes vs. RH:
The Calibration Process
Analysis of Individual Variables Descriptive – –Measures of Central Tendency Mean – Average score of distribution (1 st moment) Median – Middle score (50.
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan Simple Linear Regression SECTIONS 9.3 Confidence and prediction intervals.
Chemometrics Method comparison
Inferential statistics Hypothesis testing. Questions statistics can help us answer Is the mean score (or variance) for a given population different from.
Copyright © 2012 Pearson Education. All rights reserved Copyright © 2012 Pearson Education. All rights reserved. Chapter 10 Sampling Distributions.
Principal Components An Introduction
1 Summarizing Performance Data Confidence Intervals Important Easy to Difficult Warning: some mathematical content.
Analysis of Variance. ANOVA Probably the most popular analysis in psychology Why? Ease of implementation Allows for analysis of several groups at once.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 6 – Multiple comparisons, non-normality, outliers Marshall.
Bivariate Relationships Analyzing two variables at a time, usually the Independent & Dependent Variables Like one variable at a time, this can be done.
Programming for Geographical Information Analysis: Advanced Skills Lecture 10: Modelling II: The Modelling Process Dr Andy Evans.
1 Validation & Verification Chapter VALIDATION & VERIFICATION Very Difficult Very Important Conceptually distinct, but performed simultaneously.
Correlation1.  The variance of a variable X provides information on the variability of X.  The covariance of two variables X and Y provides information.
The Modelling Process Dr Andy Evans. This lecture The modelling process: Identify interesting patterns Build a model of elements you think interact and.
Hydrologic Modeling: Verification, Validation, Calibration, and Sensitivity Analysis Fritz R. Fiedler, P.E., Ph.D.
Testing Theories: Three Reasons Why Data Might not Match the Theory Psych 437.
Evidence Based Medicine
Bootstrapping (And other statistical trickery). Reminder Of What We Do In Statistics Null Hypothesis Statistical Test Logic – Assume that the “no effect”
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 1, Slide 1 Chapter 01 Stats Starts Here.
Reinsurance of Long Tail Liabilities Dr Glen Barnett and Professor Ben Zehnwirth.
Extension to Multiple Regression. Simple regression With simple regression, we have a single predictor and outcome, and in general things are straightforward.
Generic Approaches to Model Validation Presented at Growth Model User’s Group August 10, 2005 David K. Walters.
VI. Evaluate Model Fit Basic questions that modelers must address are: How well does the model fit the data? Do changes to a model, such as reparameterization,
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
The Modelling Process Dr Andy Evans. This lecture The modelling process: Identify interesting patterns Build a model of elements you think interact and.
Accuracy Assessment Having produced a map with classification is only 50% of the work, we need to quantify how good the map is. This step is called the.
The Correlational Research Strategy
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
Model validation Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.
Regression Analysis A statistical procedure used to find relations among a set of variables.
Propagation of Error Ch En 475 Unit Operations. Quantifying variables (i.e. answering a question with a number) 1. Directly measure the variable. - referred.
Data Analysis Econ 176, Fall Populations When we run an experiment, we are always measuring an outcome, x. We say that an outcome belongs to some.
Statistics Presentation Ch En 475 Unit Operations.
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan 11/6/12 Simple Linear Regression SECTIONS 9.1, 9.3 Inference for slope (9.1)
Chapter 10 Verification and Validation of Simulation Models
Stat 112 Notes 9 Today: –Multicollinearity (Chapter 4.6) –Multiple regression and causal inference.
M ONTE C ARLO SIMULATION Modeling and Simulation CS
Hypothesis Testing An understanding of the method of hypothesis testing is essential for understanding how both the natural and social sciences advance.
The Modelling Process Dr Andy Evans. This lecture The modelling process: Identify interesting patterns Build a model of elements you think interact and.
1 Summarizing Performance Data Confidence Intervals Important Easy to Difficult Warning: some mathematical content.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
ABM: Issues Dr Andy Evans. Structuring a model Models generally comprise: Objects. Environment. I/O code. Data reporting code. Some kind of time sequencing.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Statistics Presentation Ch En 475 Unit Operations.
Regression Analysis: A statistical procedure used to find relations among a set of variables B. Klinkenberg G
King Faisal University جامعة الملك فيصل Deanship of E-Learning and Distance Education عمادة التعلم الإلكتروني والتعليم عن بعد [ ] 1 جامعة الملك فيصل عمادة.
Review Statistical inference and test of significance.
Introduction Sample surveys involve chance error. Here we will study how to find the likely size of the chance error in a percentage, for simple random.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 7 Scatterplots, Association, and Correlation.
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
Data measurement, probability and Spearman’s Rho
Selecting the Best Measure for Your Study
A Spatial Analysis of the Central London Office Market
Modelling Dr Andy Evans In this lecture we'll look at modelling.
Statistics Review ChE 477 Winter 2018 Dr. Harding.
Natalie Robinson Centre for Evidence-based Veterinary Medicine
Introduction to modelling
Presentation transcript:

Validation Dr Andy Evans

Preparing to model Verification Calibration/Optimisation Validation Sensitivity testing and dealing with error

Validation Can you quantitatively replicate known data? Important part of calibration and verification as well. Need to decide on what you are interested in looking at. Visual or “face” validation eg. Comparing two city forms. One-number statistic eg. Can you replicate average price? Spatial, temporal, or interaction match eg. Can you model city growth block-by-block?

Validation If we can’t get an exact prediction, what standard can we judge against? Randomisation of the elements of the prediction. eg. Can we do better at geographical prediction of urban areas than randomly throwing them at a map. Doesn’t seem fair as the model has a head start if initialised with real data. Business-as-usual If we can’t do better than no prediction, we’re not doing very well. But, this assumes no known growth, which the model may not.

Visual comparison

Comparison stats: space and class Could compare number of geographical predictions that are right against chance randomly right: Kappa stat. Construct a confusion matrix / contingency table: for each area, what category is it in really, and in the prediction. Fraction of agreement = ( ) / ( ) = 0.6 Probability Predicted A = ( ) / ( ) = 0.5 Probability Real A = (10 + 5) / ( ) = 0.3 Probability of random agreement on A = 0.3 * 0.5 = 0.15 Predicted APredicted B Real A10 areas5 areas Real B15 areas20 areas

Comparison stats Equivalents for B: Probability Predicted B = (5 + 20) / ( ) = 0.5 Probability Real B = ( ) / ( ) = 0.7 Probability of random agreement on B = 0.5 * 0.7 = 0.35 Probability of not agreeing = = 0.65 Total probability of random agreement = = 0.5 Total probability of not random agreement = 1 – ( ) = 0.5 κ = fraction of agreement - probability of random agreement probability of not agreeing randomly = 0.1 / 0.50 = 0.2

Comparison stats Tricky to interpret κStrength of Agreement < 0None 0.0 — 0.20Slight 0.21 — 0.40Fair 0.41 — 0.60Moderate 0.61 — 0.80Substantial 0.81 — 1.00Almost perfect

Comparison stats The problem is that you are predicting in geographical space and time as well as categories. Which is a better prediction?

Comparison stats The solution is a fuzzy category statistic and/or multiscale examination of the differences (Costanza, 1989). Scan across the real and predicted map with a larger and larger window, recalculating the statistics at each scale. See which scale has the strongest correlation between them – this will be the best scale the model predicts at? The trouble is, scaling correlation statistics up will always increase correlation coefficients.

Correlation and scale Correlation coefficients tend to increase with the scale of aggregations. Robinson (1950) compared illiteracy in those defined as in ethnic minorities in the US census. Found high correlation in large geographical zones, less at state level, but none at individual level. Ethnic minorities lived in high illiteracy areas, but weren’t necessarily illiterate themselves. More generally, areas of effect overlap: Road accidents Dog walkers

Comparison stats So, we need to make a judgement – best possible prediction for the best possible resolution.

Comparison stats: time-series correlation This is kind of similar to the cross-correlation of time series, in which the standard difference between two datasets is lagged by increasing increments. r lag

Comparison stats: Graph / SIM flows Make an origin-destination matrix for model and reality. Compare the two using some difference statistic. Only probably is all the zero origins/destinations, which tend to reduce the significance of the statistics, not least if they give an infinite percentage increase in flow. Knudsen and Fotheringham (1986) test a number of different statistics and suggest Standardised Root Mean Squared Error is the most robust.

Preparing to model Verification Calibration/Optimisation Validation Sensitivity testing and dealing with error

Errors Model errors Data errors: Errors in the real world Errors in the model Ideally we need to know if the model is a reasonable version of reality. We also need to know how it will respond to minor errors in the input data.

Sensitivity testing Tweak key variables in a minor way to see how the model responds. The model maybe ergodic, that is, insensitive to starting conditions after a long enough run. If the model does respond strongly is this how the real system might respond, or is it a model artefact? If it responds strongly what does this say about the potential errors that might creep into predictions if your initial data isn't perfectly accurate? Is error propagation a problem? Where is the homeostasis?

Prediction If the model is deterministic, one run will be much like another. If the model is stochastic (ie. includes some randomisation), you’ll need to run in multiple times. In addition, if you’re not sure about the inputs, you may need to vary them to cope with the uncertainty: Monte Carlo testing runs 1000’s of models with a variety of potential inputs, and generates probabilistic answers.

Analysis Models aren’t just about prediction. They can be about experimenting with ideas. They can be about testing ideas/logic of theories. They can be to hold ideas.