Tea Break!.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION GROUP
Copula Regression By Rahul A. Parsa Drake University &
Logistic Regression Psy 524 Ainsworth.
Kriging.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to form.
Engineering optimization dilemma Optimization algorithms developed by mathematicians are normally based on linear and quadratic approximations Usually.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
A Short Introduction to Curve Fitting and Regression by Brad Morantz
Including Uncertainty Models for Surrogate-based Global Optimization
Gizem ALAGÖZ. Simulation optimization has received considerable attention from both simulation researchers and practitioners. Both continuous and discrete.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Surrogate model based design optimization
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Dimensional reduction, PCA
Additional Topics in Regression Analysis
Genetic Algorithms in Materials Processing N. Chakraborti Department of Metallurgical & Materials Engineering Indian Institute of Technology Kharagpur.
Fitting a Model to Data Reading: 15.1,
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Principles of the Global Positioning System Lecture 13 Prof. Thomas Herring Room A;
Principles of the Global Positioning System Lecture 11 Prof. Thomas Herring Room A;
Muhammad Moeen YaqoobPage 1 Moment-Matching Trackers for Difficult Targets Muhammad Moeen Yaqoob Supervisor: Professor Richard Vinter.
Geo479/579: Geostatistics Ch13. Block Kriging. Block Estimate  Requirements An estimate of the average value of a variable within a prescribed local.
Surrogate-based constrained multi-objective optimization Aerospace design is synonymous with the use of long running and computationally intensive simulations,
Gaussian process modelling
Efficient Model Selection for Support Vector Machines
Least-Mean-Square Training of Cluster-Weighted-Modeling National Taiwan University Department of Computer Science and Information Engineering.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Modern Navigation Thomas Herring
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
2005MEE Software Engineering Lecture 11 – Optimisation Techniques.
1 Using Multiple Surrogates for Metamodeling Raphael T. Haftka (and Felipe A. C. Viana University of Florida.
VI. Regression Analysis A. Simple Linear Regression 1. Scatter Plots Regression analysis is best taught via an example. Pencil lead is a ceramic material.
“Isolating Failure Causes through Test Case Generation “ Jeremias Rößler Gordon Fraser Andreas Zeller Alessandro Orso Presented by John-Paul Ore.
Estimating and Testing Hypotheses about Means James G. Anderson, Ph.D. Purdue University.
Stats 845 Applied Statistics. This Course will cover: 1.Regression –Non Linear Regression –Multiple Regression 2.Analysis of Variance and Experimental.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav.
Derivative-Enhanced Variable Fidelity Kriging Approach Dept. of Mechanical Engineering, University of Wyoming, USA Wataru YAMAZAKI 23 rd, September, 2010.
Tutorial I: Missing Value Analysis
Geology 5670/6670 Inverse Theory 20 Feb 2015 © A.R. Lowry 2015 Read for Mon 23 Feb: Menke Ch 9 ( ) Last time: Nonlinear Inversion Solution appraisal.
A comparative approach for gene network inference using time-series gene expression data Guillaume Bourque* and David Sankoff *Centre de Recherches Mathématiques,
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
A Hybrid Optimization Approach for Automated Parameter Estimation Problems Carlos A. Quintero 1 Miguel Argáez 1, Hector Klie 2, Leticia Velázquez 1 and.
1 Design of experiment for computer simulations Let X = (X 1,…,X p )  R p denote the vector of input values chosen for the computer program Each X j is.
Structural & Multidisciplinary Optimization Group Deciding How Conservative A Designer Should Be: Simulating Future Tests and Redesign Nathaniel Price.
Computacion Inteligente Least-Square Methods for System Identification.
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
CWR 6536 Stochastic Subsurface Hydrology Optimal Estimation of Hydrologic Parameters.
Kriging - Introduction Method invented in the 1950s by South African geologist Daniel Krige (1919-) for predicting distribution of minerals. Became very.
March 9th, 2015 Maxime Lapalme Nazim Ould-Brahim
12. Principles of Parameter Estimation
Fundamentals of estimation and detection in signals and images
Boosting and Additive Trees
Collaborative Filtering Matrix Factorization Approach
Filtering and State Estimation: Basic Concepts
Principles of the Global Positioning System Lecture 11
Principles of the Global Positioning System Lecture 13
Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014
12. Principles of Parameter Estimation
Applications of CFD 2008 Neil W. Bressloff N. W. Bressloff
Probabilistic Surrogate Models
Stochastic Methods.
Presentation transcript:

Tea Break!

Coming up: Fixing problems with expected improvement et al. Noisy data ‘Noisy’ deterministic data Multi-fidelity expected improvement Multi-objective expected improvement

Different parameter values have a big effect on expected improvement

The Fix

A one-stage approach combines the search of the model parameters with that of the infill criterion Choose a goal value, g, of the objective function The merit of sampling at a new point x is based on the likelihood of the observed data conditional on passing though x with function value g At each x theta is chosen to maximize the conditional likelihood

g=-5

Avoiding underestimating the error At a given x, Kriging predictor is most likely value How much lower could the output be, e.g. how much error? Approach: Hypothesise that at x the function has a value y Maximize the likelihood of the data (by varying theta) conditional on passing through the point x,y Keep reducing y until the change in the likelihood is more than can be accepted by a likelihood ratio test Difference between Kriging prediction and lowest value is measure of error, which is robust to poor theta estimation

Example For limit=0.975, chi-squared critical = 5.0, lowest value fails likelihood ratio test

Use to compute a new one-stage error bound Should provide better error estimates with sparse sampling/ deceptive functions Will converge upon standard error estimate for well sampled problems

Comparison with standard error estimates

New one-stage expected improvement One-stage error estimate embedded within usual expected improvement formulation Now a constrained optimization problem with more dimensions (>2k+1) All the usual benefits of expected improvement, but now better!?

EI using robust error estimate

EI using robust error: passive vibration isolating truss example

Difficult design landscape

Deceptive sample E[I(x)] E[I(x,yh,θ)]

Lucky sample E[I(x)] E[I(x,yh,θ)]

A Quicker Way

Problem is when theta is underestimated Make one adjustment to theta, not at every point Procedure Maximize likelihood to find model parameters Maximize the thetas subject to likelihood not degrading too much (based on likelihood ratio test) Maximize EI using conservative thetas for standard error calculation

Truss problem Luck sample (top) deceptive sample (bottom)

8 variable truss problem

10 runs of 8 variable truss problem

Noisy Data

‘Noisy’ data Many data sets are corrupted by noise In computational engineering, deterministic ‘noise’ ‘Noise’ in aerofoil drag data due to discretization of Euler equations

Failure of interpolation based infill Surrogate becomes excessively snaky Error estimates increase Search becomes too global

Regression, by adding constant λ to diagonal of correlation matrix, improves model

A few issues with error estimates Interpolation error=0 at sample point: at x=xi But not for regression:

EI is no longer a global search

‘Noisy’ Deterministic Data

Want ‘error’=0 at sample points Answer is to ‘re-interpolate points from the regressing model Equivalent to using in the interpolating error equation

Re-interpolation error estimate Errors due to noise removed Only modelling errors included

Now EI is global method again

Note of caution when calculating EI as:

Two variable aerofoil example Same as missing data problem Course mesh causes ‘noise’

Interpolation – very global

Re-interpolation – searches local basins, but finds global optimum

Multi-fidelity data

Can use partially converged CFD as low fidelity (tunable) model

Multi-level convergence wing optimization

Co-kriging Expensive data modelled as scaled cheap data based process plus difference process So, have covariance matrix:

One variable example

Multi-fidelity geometry example 12 geometry variables 10 full car RANS simulations 15h each 120 rear wing only RANS simulations 1.5h each

Rear wing only Full car

Kriging models Visualisation of four most important variables Based on 20 full car simulations correct data, but not enough? Based on 120 rear wing simulations right trends, but incorrect data?

Co-Kriging, all data

Design improvement

Multi-objective EI

Pareto optimization We want to identify a set of non-dominated solutions These define the Pareto front We can formulate an expectation of improvement on the current non-dominated solutions

Multi-dimensional Gaussian process Consider a 2 objective problem The random variables Y1 and Y2 have a 2D probability density function:

Probability of improving on one point Need to integrate the 2D pdf:

Integrating under all non-dominated solutions: The EI is the first moment of this integral about the Pareto front (see book)

Pareto solutions

Summary Surrogate based optimization offers answers to, or ways to get round, many problems associated with real world optimization This seemingly blunt tool must, however, be used with precision as there are many traps to fall into Co-Kriging seems like a great way to combine multi-fidelity data How best to optimize with stochastic noisy data? Only consider modelling error and use multiple evaluations to drive down random error? or forgo global exploration?

References All Matlab code at www.wiley.com/go/forrester (or email me) A. I. J. Forrester, A. Sóbester, A. J. Keane, Engineering Design via Surrogate Modelling: A Practical Guide, John Wiley & Sons, Chichester, 240 pages, ISBN 978-0-470-06068-1.   A. I. J. Forrester, A. J. Keane, Recent advances in surrogate-based optimization, Progress in Aerospace Sciences, 45, 50-79, (doi:10.1016/j.paerosci.2008.11.001) A. I. J. Forrester, A. Sóbester, A. J. Keane, Multi-fidelity optimization via surrogate modelling. Proc. R. Soc. A, 463(2088):3251–3269, 2007.  A. I. J. Forrester, A. Sóbester, A. J. Keane, Optimization with missing data, Proc. R. Soc. A, 462(2067), 935-945, (doi:10.1098/rspa.2005.1608). A. I. J. Forrester, N. W. Bressloff, A. J. Keane, Design and analysis of ‘noisy’ computer experiments, AIAA journal, 44(10), 2331-2339, (doi:10.2514/1.20068). All Matlab code at www.wiley.com/go/forrester (or email me)

Gratuitous publicity