Automating The Selection of a Simulation Warm-up Period Stewart Robinson, Katy Hoad, Ruth Davies Warwick Business School Cardiff University - October 2008.

Slides:



Advertisements
Similar presentations
Chapter 5 Multiple Linear Regression
Advertisements

Part II – TIME SERIES ANALYSIS C3 Exponential Smoothing Methods © Angel A. Juan & Carles Serrat - UPC 2007/2008.
AGIFORS--RM Study Group New York City, March 2000 Lawrence R. Weatherford, PhD University of Wyoming Unconstraining Methods.
Modellierung großer Netze in der Logistik SFB 559 Initial Transient Period Detection Using Parallel Replications F. Bause, M. Eickhoff LS Informatik IV,
Automating the Analysis of Simulation Output Data Katy Hoad Stewart Robinson, Ruth Davies, Mark Elder Funded by EPSRC and SIMUL8.
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
ETM 607 – Output Analysis: Estimation of Absolute Performance
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 7: Box-Jenkins Models – Part II (Ch. 9) Material.
Confidence Intervals, Effect Size and Power
Output analyses for single system
Non-Seasonal Box-Jenkins Models
Validation and Monitoring Measures of Accuracy Combining Forecasts Managing the Forecasting Process Monitoring & Control.
Automating estimation of warm-up length Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School WSC08 The AutoSimOA Project A 3 year, EPSRC funded.
The AutoSimOA Project Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School WSC 07 A 3 year, EPSRC funded project in collaboration with SIMUL8.
Some more issues of time series analysis Time series regression with modelling of error terms In a time series regression model the error terms are tentatively.
Data Sources The most sophisticated forecasting model will fail if it is applied to unreliable data Data should be reliable and accurate Data should be.
Automating estimation of warm-up length Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School Simulation Workshop - April 2008 The AutoSimOA.
Automated Analysis of Simulation Output Data and the AutoSimOA Project
Automating the Analysis of Simulation Output Data Katy Hoad, Stewart Robinson, Ruth Davies SSIG Meeting, 24th October 2007
Classification of Discrete Event Simulation Models and Output Data: Creating a Sufficient Model Set. Katy Hoad Stewart Robinson,
Lecture 9 Output Analysis for a Single Model. 2  Output analysis is the examination of data generated by a simulation.  Its purpose is to predict the.
The AutoSimOA Project Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School OR49 Sept 07 A 3 year, EPSRC funded project in collaboration with.
Automating the Analysis of Simulation Output Data Stewart Robinson, Katy Hoad, Ruth Davies OR48, September 2006.
Automating the Analysis of Simulation Output Data Katy Hoad Stewart Robinson, Ruth Davies, Mark Elder Funded by EPSRC and SIMUL8.
Part II – TIME SERIES ANALYSIS C2 Simple Time Series Methods & Moving Averages © Angel A. Juan & Carles Serrat - UPC 2007/2008.
X-12 ARIMA Eurostat, Luxembourg Seasonal Adjustment.
Automating the Analysis of Simulation Output Data Stewart Robinson, Katy Hoad, Ruth Davies ORGS Meeting, 4th October 2007
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
SIMULATION. Simulation Definition of Simulation Simulation Methodology Proposing a New Experiment Considerations When Using Computer Models Types of Simulations.
1 Simulation Modeling and Analysis Output Analysis.
191 Drawing Statistical Inference from Simulation Runs......the "fun" stuff!
FINAL REPORT: OUTLINE & OVERVIEW OF SURVEY ERRORS
Relationships Among Variables
Copyright ©2011 Pearson Education 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft Excel 6 th Global Edition.
AutoSimOA : A Framework for Automated Analysis of Simulation Output Stewart Robinson Katy Hoad, Ruth Davies Funded by.
1 1 Slide © 2003 South-Western/Thomson Learning™ Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 Terminating Statistical Analysis By Dr. Jason Merrick.
Mean Tests & X 2 Parametric vs Nonparametric Errors Selection of a Statistical Test SW242.
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft.
Steady-State Statistical Analysis By Dr. Jason Merrick.
The Examination of Residuals. The residuals are defined as the n differences : where is an observation and is the corresponding fitted value obtained.
Elementary Statistical Methods André L. Souza, Ph.D. The University of Alabama Lecture 22 Statistical Power.
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
The next step in performance monitoring – Stochastic monitoring (and reserving!) NZ Actuarial Conference November 2010.
University of Ottawa - Bio 4118 – Applied Biostatistics © Antoine Morin and Scott Findlay 08/10/ :23 PM 1 Some basic statistical concepts, statistics.
Multiple Regression Selecting the Best Equation. Techniques for Selecting the "Best" Regression Equation The best Regression equation is not necessarily.
CS Learning Rules1 Learning Sets of Rules. CS Learning Rules2 Learning Rules If (Color = Red) and (Shape = round) then Class is A If (Color.
Time Series Data Analysis - I Yaji Sripada. Dept. of Computing Science, University of Aberdeen2 In this lecture you learn What are Time Series? How to.
Budget-based Control for Interactive Services with Partial Execution 1 Yuxiong He, Zihao Ye, Qiang Fu, Sameh Elnikety Microsoft Research.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 4): Power Fall, 2008.
McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved. 1.
Chapter 10 Verification and Validation of Simulation Models
MINING COLOSSAL FREQUENT PATTERNS BY CORE PATTERN FUSION FEIDA ZHU, XIFENG YAN, JIAWEI HAN, PHILIP S. YU, HONG CHENG ICDE07 Advisor: Koh JiaLing Speaker:
Multiple Regression Selecting the Best Equation. Techniques for Selecting the "Best" Regression Equation The best Regression equation is not necessarily.
1 OUTPUT ANALYSIS FOR SIMULATIONS. 2 Introduction Analysis of One System Terminating vs. Steady-State Simulations Analysis of Terminating Simulations.
Summarizing Risk Analysis Results To quantify the risk of an output variable, 3 properties must be estimated: A measure of central tendency (e.g. µ ) A.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Short term forecast of travel times on the Danish highway network based on TRIM data Klaus Kaae Andersen Thomas Kaare Christensen Bo Friis Nielsen Informatics.
Chapter 10 Verification and Validation of Simulation Models Banks, Carson, Nelson & Nicol Discrete-Event System Simulation.
Module 4 Forecasting Multiple Variables from their own Histories EC 827.
Metaheuristics for the New Millennium Bruce L. Golden RH Smith School of Business University of Maryland by Presented at the University of Iowa, March.
Demand Management and Forecasting Chapter 11 Portions Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
MODEL DIAGNOSTICS By Eni Sumarminingsih, Ssi, MM.
Chapter 4 Basic Estimation Techniques
Chapter 10 Verification and Validation of Simulation Models
More About Tests Notes from
M. Kezunovic (P.I.) S. S. Luo D. Ristanovic Texas A&M University
Wednesday After Lunch: Add randomness to the Flowers Model
Presentation transcript:

Automating The Selection of a Simulation Warm-up Period Stewart Robinson, Katy Hoad, Ruth Davies Warwick Business School Cardiff University - October 2008 The AutoSimOA Project A 3 year, EPSRC funded project in collaboration with SIMUL8 Corporation.

Research Aim To create an automated system for advising a non-expert user on how to obtain accurate measures of model performance i.e. warm-up, run-length and number of replications For implementation into simulation software AutoSimOA = Automated SIMulation Output Analysis

Simulation model Warm-up analysis Run-length analysis Replications analysis Use replications or long-run? Recommendation possible? Recommend- ation Output data Analyser Obtain more output data

The Initial Bias Problem Model may not start in a “typical” state. Can cause initial bias in the output. Many methods proposed for dealing with initial bias, e.g.: –Initial steady-state conditions –Run model for ‘long’ time –Deletion of initial transient data (‘warm-up period’)

The Initial Bias Problem

This project uses: Deletion of the initial transient data by specifying a warm- up period (or truncation point). The question is: How do you estimate the length of the warm-up period required?

Methods fall into 5 main types : 1. Graphical Methods. 2. Heuristic Approaches. 3. Statistical Methods. 4. Initialisation Bias Tests. 5. Hybrid Methods.

Literature search – 44 methods Summary of methods and literature references on project web site:

Short-listing warm-up methods for automation using literature Short-listing Criteria » Accuracy & robustness » Ease of automation » Generality » Computer running time

Short-listing results: reasons for rejection of methods

Statistical methods: –Goodness of Fit (GoF) test –Algorithm for a static data set (ASD) –Algorithm for a Dynamic data set (ADD) Heuristics: –MSER-5 –Kimbler’s Double Exponential Smoothing –Euclidean Distance Method (ED) Short-listing results: 6 Methods taken forward to testing

Preliminary testing of shortlisted methods Rejected methods: –ASD & ADD required a prohibitively large number of replications –GoF & Kimbler’s method consistently and severely underestimated truncation point. –ED failed to give any result on majority of occasions MSER-5 most accurate and robust method

MSER-5 warm-up method

Further Testing of MSER-5 1.Artificial data – controllable & comparable  initial bias functions  steady state functions 2.Full factorial design. 3.Set of performance criteria.

ParametersLevels Data TypeSingle run Data averaged over 5 reps Error typeN(1,1), Exp(1) Auto- correlation None, AR(1), AR(2), MA(2), AR(4), ARMA(5,5) Bias Severity1, 2, 4 Bias Length0%, 10%, 40%, 100% Bias directionPositive, Negative Bias shape7 shapes Artificial Data Parameters

Mean Shift: Linear: Quadratic: Exponential: Oscillating (decreasing):

Add Initial Bias to Steady state: Superpostion: Bias Fn, a(t), added onto end of steady state function: e.g. 2. Full factorial design: 3048 types of artificial data set MSER-5 run with each type 100 times

i.Coverage of true mean. ii.Closeness of estimated truncation point (Lsol) to true truncation point (L). iii.Percentage bias removed by truncation. iv.Analysis of the pattern & frequency of rejections of Lsol (i.e. Lsol > n/2). 3. Performance Criteria

MSER-5 Results Does the true mean fall into the 95% CI for the estimated mean? Non-truncated data sets Truncated data sets % of cases yes 7.7% yesno0% no 19.8% noyes72.5% i. Coverage of true mean.

ii. Closeness of Lsol to L: Wide range of Lsol values. e.g. (Positive bias functions, single run data, N(1,1) errors, MA(2) auto-correlation, bias severity value of 2 and true L = 100.)

iii. Percentage bias removed by truncation.

Effect of data parameters on bias removal No significant effect: Error type Bias direction Significant effect: Data type Auto-correlation type Bias shape Bias severity Bias length

More bias removed by using averaged replications rather than a single run.

The stronger the auto-correlation, the less accurate the bias removal. Effect greatly reduced by using averaged data.

The more sharply the initial bias declines, the more likely MSER-5 is to underestimate the warm-up period and to remove increasingly less bias.

As the bias severity increases, MSER-5 removes an increasingly higher percentage of the bias.

Longer bias removed slightly more efficiently than shorter bias. Shorter bias - more overestimations - partly due to longer bias overestimations being more likely to be rejected.

Rejections caused by: high auto-correlation, bias close to or over n/2, smooth end to data = ‘end point’ rejection. Averaged data slightly increases probability of getting ‘end point’ rejection but increases probability of more accurate L estimates. iv. Lsol rejections

Giving more data to MSER-5 in an iterative fashion produces a valid Lsol value where previously the Lsol value had been rejected. e.g. ARMA(5,5)

Lsol valuesPercentage of cases Lsol = 071% Lsol ≤ 5093% Testing MSER-5 with data that has no initial bias. Want Lsol = 0 Lsol > 50 mainly due to highest auto-correlated data sets - AR(1) & ARMA(5,5). Rejected Lsol values: 5.6% of the 2400 Lsol values produced. 93% from the highest auto-correlated data ARMA(5,5).

Testing MSER-5 with data that has 100% bias. Want 100% rejection rate: Actual rate = 61%

Summary MSER-5 most promising method for automation –Not model or data type specific. –No estimation of parameters needed. –Can function without user intervention. –Shown to perform robustly and effectively for the majority of data sets tested. –Quick to run. –Fairly simple to understand.

Heuristic framework around MSER-5 Iterative procedure for procuring more data when required. ‘Failsafe’ mechanism - to deal with possibility of data not in steady state; insufficient data provided when highly auto- correlated. Being implemented in SIMUL8.