Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.

Slides:



Advertisements
Similar presentations
Sampling plans for linear regression
Advertisements

Kriging.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to form.
Mech4740 Numerical Methods, Hong Kong University of Science and Technology. 1 Part Five Curve Fitting.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
Regression Greg C Elvers.
Local surrogates To model a complex wavy function we need a lot of data. Modeling a wavy function with high order polynomials is inherently ill-conditioned.
P. Venkataraman Mechanical Engineering P. Venkataraman Rochester Institute of Technology DETC2011 –47658 Determining ODE from Noisy Data 31 th CIE, Washington.
MATH 685/ CSI 700/ OR 682 Lecture Notes
Forecasting JY Le Boudec 1. Contents 1.What is forecasting ? 2.Linear Regression 3.Avoiding Overfitting 4.Differencing 5.ARMA models 6.Sparse ARMA models.
WFM 6202: Remote Sensing and GIS in Water Management © Dr. Akm Saiful IslamDr. Akm Saiful Islam WFM 6202: Remote Sensing and GIS in Water Management Akm.
Curve Fitting and Interpolation: Lecture (IV)
Function Approximation
Curve-Fitting Regression
Class 5: Thurs., Sep. 23 Example of using regression to make predictions and understand the likely errors in the predictions: salaries of teachers and.
ECIV 301 Programming & Graphics Numerical Methods for Engineers Lecture 26 Regression Analysis-Chapter 17.
ECIV 301 Programming & Graphics Numerical Methods for Engineers Lecture 25 Regression Analysis-Chapter 17.
Petter Mostad Linear regression Petter Mostad
ECIV 301 Programming & Graphics Numerical Methods for Engineers Lecture 24 Regression Analysis-Chapter 17.
Lecture 5 Curve fitting by iterative approaches MARINE QB III MARINE QB III Modelling Aquatic Rates In Natural Ecosystems BIOL471 © 2001 School of Biological.
Chi Square Distribution (c2) and Least Squares Fitting
Chapter 6 Numerical Interpolation
ECIV 301 Programming & Graphics Numerical Methods for Engineers Lecture 21 CURVE FITTING Chapter 18 Function Interpolation and Approximation.
Stat Notes 4 Chapter 3.5 Chapter 3.7.
Curve fit noise=randn(1,30); x=1:1:30; y=x+noise ………………………………… [p,s]=polyfit(x,y,1);
Classification and Prediction: Regression Analysis
Chapter 6-2 Radial Basis Function Networks 1. Topics Basis Functions Radial Basis Functions Gaussian Basis Functions Nadaraya Watson Kernel Regression.
Least-Squares Regression
CpE- 310B Engineering Computation and Simulation Dr. Manal Al-Bzoor
Gaussian process modelling
Local and Local-Global Approximations Local algebraic approximations – Variants on Taylor series Local-Global approximations – Variants on “fudge factor”
Kalman Filter (Thu) Joon Shik Kim Computational Models of Intelligence.
Time Series Data Analysis - I Yaji Sripada. Dept. of Computing Science, University of Aberdeen2 In this lecture you learn What are Time Series? How to.
MECN 3500 Inter - Bayamon Lecture 9 Numerical Methods for Engineering MECN 3500 Professor: Dr. Omar E. Meza Castillo
Local surrogates To model a complex wavy function we need a lot of data. Modeling a wavy function with high order polynomials is inherently ill-conditioned.
WB1440 Engineering Optimization – Concepts and Applications Engineering Optimization Concepts and Applications Fred van Keulen Matthijs Langelaar CLA H21.1.
Curve-Fitting Regression
MECN 3500 Inter - Bayamon Lecture 3 Numerical Methods for Engineering MECN 3500 Professor: Dr. Omar E. Meza Castillo
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
1 Using Multiple Surrogates for Metamodeling Raphael T. Haftka (and Felipe A. C. Viana University of Florida.
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
Lecture 16 - Approximation Methods CVEN 302 July 15, 2002.
Numerical Analysis. Numerical Analysis or Scientific Computing Concerned with design and analysis of algorithms for solving mathematical problems that.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
Curve Fitting Pertemuan 10 Matakuliah: S0262-Analisis Numerik Tahun: 2010.
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
INCLUDING UNCERTAINTY MODELS FOR SURROGATE BASED GLOBAL DESIGN OPTIMIZATION The EGO algorithm STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION GROUP Thanks.
Machine Learning 5. Parametric Methods.
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
Optimization formulation Optimization methods help us find solutions to problems where we seek to find the best of something. This lecture is about how.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
Kriging - Introduction Method invented in the 1950s by South African geologist Daniel Krige (1919-) for predicting distribution of minerals. Became very.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Sampling plans for linear regression
Questions from lectures
CSE 4705 Artificial Intelligence
Stat 112 Notes 4 Today: Review of p-values for one-sided tests
MATH 2140 Numerical Methods
POLYNOMIAL INTERPOLATION
Curve fit metrics When we fit a curve to data we ask:
Curve fit metrics When we fit a curve to data we ask:
Artificial Intelligence 10. Neural Networks
Machine learning overview
Empirical Bayesian Kriging and EBK Regression Prediction – Robust Kriging as Geoprocessing Tools Eric Krause.
Regression and Correlation of Data
Presentation transcript:

Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture deals with the following case: –The data is noisy. –The functional form of the true function is known. –The data is dense enough to allow us some noise filtering. The objective is to answer the two questions.

Curve fit We sample the function y=x (in red) at x=1,2,…,30, add noise with standard deviation 1 and fit a linear polynomial (blue). How would you check the statement that fit is more accurate than the data? With dense data, functional form is clear. Fit serves to filter out noise

Regression The process of fitting data with a curve by minimizing the mean square difference from the data is known as regression Term originated from first paper to use regression dealt with a phenomenon called regression to the mean The polynomial regression on the previous slide is a simple regression, where we know or assume the functional shape and need to determine only the coefficients.

Surrogate (metamodel) The algebraic function we fit to data is called surrogate, metamodel or approximation. Polynomial surrogates were invented in the 1920s to characterize crop yields in terms of inputs such as water and fertilizer. They were called then “response surface approximations.” The term “surrogate” captures the purpose of the fit: using it instead of the data for prediction. Most important when data is expensive and noisy, especially for optimization.

Surrogates for fitting simulations Great interest now in fitting computer simulations Computer simulations are also subject to noise (numerical) Simulations are exactly repeatable, so noise is hidden. Some surrogates (e.g. polynomial response surfaces) cater mostly to noisy data. Some (e.g. Kriging) interpolate data.

Surrogates of given functional form Noisy response Linear approximation Rational approximation Data from n y experiments Error (fit) metrics

Linear Regression Functional form For linear approximation Error or difference between data and surrogate Rms error Minimize rms error e T e=(y-Xb T ) T (y-Xb T ) Differentiate to obtain Beware of ill-conditioning !

Example

Other metric fits Rms fitAv. Err. fitMax err. fit RMS error Av. error Max error

Three lines

Original 30-point curve fit With dense data difference due to metrics is small. Rms fitAv. Err. fitMax err. fit RMS error Av. error Max error

problems 1.Find other metrics for a fit beside the three discussed in this lecture. 2.Redo the 30-point example with the surrogate y=bx. Use the same data. 3. Redo the 30-point example using only every third point (x=3,6,…). You can consider the other 20 points as test points used to check the fit. Compare the difference between the fit and the data points to the difference between the fit and the test points. It is sufficient to do it for one fit metric. Source: Smithsonian Institution Number: