INCLUDING UNCERTAINTY MODELS FOR SURROGATE BASED GLOBAL DESIGN OPTIMIZATION The EGO algorithm STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION GROUP Thanks.

Slides:



Advertisements
Similar presentations
Spatial point patterns and Geostatistics an introduction
Advertisements

STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION GROUP
Pattern Recognition and Machine Learning
Kriging.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to form.
Global Optimization General issues in global optimization Classification of algorithms The DIRECT algorithm – Divided rectangles – Exploration and Exploitation.
Model calibration using. Pag. 5/3/20152 PEST program.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
Toolbox example with three surrogates Data: clc; clear all; X = [ ]; Y = sin(X); NbVariables = 1; NbPointsTraining = length(X);
Games of Prediction or Things get simpler as Yoav Freund Banter Inc.
Local surrogates To model a complex wavy function we need a lot of data. Modeling a wavy function with high order polynomials is inherently ill-conditioned.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
CMPUT 466/551 Principal Source: CMU
Optimal Design Laboratory | University of Michigan, Ann Arbor 2011 Design Preference Elicitation Using Efficient Global Optimization Yi Ren Panos Y. Papalambros.
Including Uncertainty Models for Surrogate-based Global Optimization
The loss function, the normal equation,
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Surrogate model based design optimization
Sparse vs. Ensemble Approaches to Supervised Learning
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Class 5: Thurs., Sep. 23 Example of using regression to make predictions and understand the likely errors in the predictions: salaries of teachers and.
An Optimal Learning Approach to Finding an Outbreak of a Disease Warren Scott Warren Powell
Stat 112: Lecture 18 Notes Chapter 7.1: Using and Interpreting Indicator Variables. Visualizing polynomial regressions in multiple regression Review Problem.
Assigning Numbers to the Arrows Parameterizing a Gene Regulation Network by using Accurate Expression Kinetics.
Applications in GIS (Kriging Interpolation)
Curve fit noise=randn(1,30); x=1:1:30; y=x+noise ………………………………… [p,s]=polyfit(x,y,1);
Chapter 6-2 Radial Basis Function Networks 1. Topics Basis Functions Radial Basis Functions Gaussian Basis Functions Nadaraya Watson Kernel Regression.
Mathematical Programming in Support Vector Machines
Space-Filling DOEs Design of experiments (DOE) for noisy data tend to place points on the boundary of the domain. When the error in the surrogate is due.
Objectives of Multiple Regression
CHAPTER 15 S IMULATION - B ASED O PTIMIZATION II : S TOCHASTIC G RADIENT AND S AMPLE P ATH M ETHODS Organization of chapter in ISSO –Introduction to gradient.
PATTERN RECOGNITION AND MACHINE LEARNING
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
July 11, 2001Daniel Whiteson Support Vector Machines: Get more Higgs out of your data Daniel Whiteson UC Berkeley.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
1 FORECASTING Regression Analysis Aslı Sencer Graduate Program in Business Information Systems.
Local surrogates To model a complex wavy function we need a lot of data. Modeling a wavy function with high order polynomials is inherently ill-conditioned.
WB1440 Engineering Optimization – Concepts and Applications Engineering Optimization Concepts and Applications Fred van Keulen Matthijs Langelaar CLA H21.1.
Spatial Statistics in Ecology: Continuous Data Lecture Three.
Managerial Economics Demand Estimation & Forecasting.
Virtual Vector Machine for Bayesian Online Classification Yuan (Alan) Qi CS & Statistics Purdue June, 2009 Joint work with T.P. Minka and R. Xiang.
Spatially Assessing Model Error Using Geographically Weighted Regression Shawn Laffan Geography Dept ANU.
1 Using Multiple Surrogates for Metamodeling Raphael T. Haftka (and Felipe A. C. Viana University of Florida.
ISCG8025 Machine Learning for Intelligent Data and Information Processing Week 3 Practical Notes Application Advice *Courtesy of Associate Professor Andrew.
Some Aspects of Bayesian Approach to Model Selection Vetrov Dmitry Dorodnicyn Computing Centre of RAS, Moscow.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
CS 8625 High Performance Computing Dr. Hoganson Copyright © 2003, Dr. Ken Hoganson CS8625 Class Will Start Momentarily… CS8625 High Performance.
1 Simple Linear Regression and Correlation Least Squares Method The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES.
Logistic Regression Saed Sayad 1www.ismartsoft.com.
Evolving RBF Networks via GP for Estimating Fitness Values using Surrogate Models Ahmed Kattan Edgar Galvan.
Optimization formulation Optimization methods help us find solutions to problems where we seek to find the best of something. This lecture is about how.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
A Hybrid Optimization Approach for Automated Parameter Estimation Problems Carlos A. Quintero 1 Miguel Argáez 1, Hector Klie 2, Leticia Velázquez 1 and.
Kriging - Introduction Method invented in the 1950s by South African geologist Daniel Krige (1919-) for predicting distribution of minerals. Became very.
Questions from lectures
Goal We present a hybrid optimization approach for solving global optimization problems, in particular automated parameter estimation models. The hybrid.
CSE 4705 Artificial Intelligence
A Simple Artificial Neuron
Probabilistic Models for Linear Regression
Statistical Methods For Engineers
CS 179 Project Ideas.
Curve fit metrics When we fit a curve to data we ask:
Curve fit metrics When we fit a curve to data we ask:
Cancer Clustering Phenomenon
Linear Discrimination
Optimization under Uncertainty
Presentation transcript:

INCLUDING UNCERTAINTY MODELS FOR SURROGATE BASED GLOBAL DESIGN OPTIMIZATION The EGO algorithm STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION GROUP Thanks to Felipe A. C. Viana

2 BACKGROUND: SURROGATE MODELING Differences are larger in regions of low point density. Surrogates replace expensive simulations by simple algebraic expressions fit to data. Kriging (KRG) Polynomial response surface (PRS) Support vector regression Radial basis neural networks Example: is an estimate of.

3 BACKGROUND: UNCERTAINTY Some surrogates also provide an uncertainty estimate: standard error, s(x). Example: kriging and polynomial response surface. These are used in EGO

4 KRIGING FIT AND THE IMPROVEMENT QUESTION First we sample the function and fit a kriging model. We note the present best solution (PBS) At every x there is some chance of improving on the PBS. Then we ask: Assuming an improvement over the PBS, where is it likely be largest?

5 WHAT IS EXPECTED IMPROVEMENT? Consider the point x=0.8, and the random variable Y, which is the possible values of the function there. Its mean is the kriging prediction, which is slightly above zero.

6 EXPLORATION AND EXPLOITATION EGO maximizes E[I(x)] to find the next point to be sampled. The expected improvement balances exploration and exploitation because it can be high either because of high uncertainty or low surrogate prediction. When can we say that the next point is “exploration?”

7 THE BASIC EGO WORKS WITH KRIGING (a) Kriging (b) Support vector regression We want to run EGO with the most accurate surrogate. But we have no uncertainty model for SVR Considering the root mean square error, :

8 IMPORTATION AT A GLANCE

9 HARTMAN3 EXAMPLE Hartman3 function (initially fitted with 20 points): After 20 iterations (i.e., total of 40 points), improvement (I) over initial best sample:

10 TWO OTHER DESIGNS OF EXPERIMENTS FIRST: SECOND:

11 SUMMARY OF THE HARTMAN3 EXAMPLE In 34 DOEs (out of 100) KRG outperforms RBNN (in those cases, the difference between the improvements has mean of only 0.8%). Box plot of the difference between improvement offered by different surrogates (out of 100 DOEs)

12 EGO WITH MULTIPLE SURROGATES Traditional EGO uses kriging to generate one point at a time. We use multiple surrogates to get multiple points.

13 POTENTIAL OF EGO WITH MULTIPLE SURROGATES Hartman3 function (100 DOEs with 20 points) Overall, surrogates are comparable in performance.

14 POTENTIAL OF EGO WITH MULTIPLE SURROGATES “krg” runs EGO for 20 iterations adding one point at a time. “ krg-svr ” and “ krg-rbnn ” run 10 iterations adding two points. Multiple surrogates offer good results in half of the time!!!