Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.

Slides:



Advertisements
Similar presentations
Young Modulus Example The pairs (1,1), (2,2), (4,3) represent strain (millistrains) and stress (ksi) measurements. Estimate Young modulus using the three.
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Kin 304 Regression Linear Regression Least Sum of Squares
Statistical Techniques I EXST7005 Simple Linear Regression.
Uncertainty in fall time surrogate Prediction variance vs. data sensitivity – Non-uniform noise – Example Uncertainty in fall time data Bootstrapping.
Kriging.
Probability & Statistical Inference Lecture 9
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Linear regression models
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
The Simple Linear Regression Model: Specification and Estimation
9. SIMPLE LINEAR REGESSION AND CORRELATION
Multivariate Data Analysis Chapter 4 – Multiple Regression.
Petter Mostad Linear regression Petter Mostad
Intro to Statistics for the Behavioral Sciences PSYC 1900
Simple Linear Regression Analysis
Quantitative Business Analysis for Decision Making Simple Linear Regression.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Curve fit noise=randn(1,30); x=1:1:30; y=x+noise ………………………………… [p,s]=polyfit(x,y,1);
Simple Linear Regression Analysis
Interpolation, extrapolation and regression Interpolation is mathematically contrasted to regression or least-squares fit As important is the contrast.
Relationships Among Variables
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Introduction to Linear Regression and Correlation Analysis
LINEAR REGRESSION Introduction Section 0 Lecture 1 Slide 1 Lecture 5 Slide 1 INTRODUCTION TO Modern Physics PHYX 2710 Fall 2004 Intermediate 3870 Fall.
CORRELATION & REGRESSION
1 FORECASTING Regression Analysis Aslı Sencer Graduate Program in Business Information Systems.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
Correlation and Prediction Error The amount of prediction error is associated with the strength of the correlation between X and Y.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Chapter 13 Multiple Regression
Section 12.3 Regression Analysis HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2008 by Hawkes Learning Systems/Quant Systems, Inc. All.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
11/23/2015Slide 1 Using a combination of tables and plots from SPSS plus spreadsheets from Excel, we will show the linkage between correlation and linear.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
1 Introduction What does it mean when there is a strong positive correlation between x and y ? Regression analysis aims to find a precise formula to relate.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Machine Learning 5. Parametric Methods.
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Hypothesis Testing Example 3: Test the hypothesis that the average content of containers of a particular lubricant is 10 litters if the contents of random.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
The simple linear regression model and parameter estimation
Questions from lectures
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
10.2 Regression If the value of the correlation coefficient is significant, the next step is to determine the equation of the regression line which is.
Kin 304 Regression Linear Regression Least Sum of Squares
BPK 304W Regression Linear Regression Least Sum of Squares
Slope of the regression line:
Correlation and Regression
Statistical Methods For Engineers
BA 275 Quantitative Business Methods
6-1 Introduction To Empirical Models
Curve fit metrics When we fit a curve to data we ask:
Curve fit metrics When we fit a curve to data we ask:
Linear Regression Summer School IFPRI
Regression and Correlation of Data
Presentation transcript:

Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple determination –Adjusted coefficient of multiple determination Prediction accuracy measures –Model independent: Cross validation error –Model dependent: Standard error

Linear Regression

Coefficient of multiple determination Equivalence of surrogate with data is often measured by how much of the variance in the data is captured by the surrogate. Coefficient of multiple determination and adjusted version

R 2 does not reflect accuracy Compare y1=x to y2=0.1x plus same noise (normally distributed with zero mean and standard deviation of 1. Estimate the average errors between the function (red) and surrogate (blue). R 2 = R 2 =0.3016

Cross validation Validation consists of checking the surrogate at a set of validation points. This may be considered wasteful because we do not use all the points for fitting the best possible surrogate. Cross validation divides data into n g groups. Fit the approximation to n g -1 groups, and use last group to estimate error. Repeat for each group. When each group consists of one point, error often called PRESS (prediction error sum of squares) Calculate error at each point and then present r.m.s error For linear regression can be shown that

Model based error for linear regression The common assumptions for linear regression –Surrogate is in functional form of true function –The data is contaminated with normally distributed error with the same standard deviation at every point. –The errors at different points are not correlated. Under these assumptions, the noise standard deviation (called standard error) is estimated as. Similarly, the standard error in the coefficients is

Comparison of errors

Top hat question We sample the function y=x with noise at x=0, 1, 2 to get 0.5, 0.5, 2.5. Assume that the linear regression fit is y=0.8x. What are the noise (epsilon), the discrepancy (e), the cross-validation error, and the actual error at x=2.

Prediction variance Linear regression model Define then With some algebra Standard error

Example of prediction variance For a linear polynomial RS y=b 1 +b 2 x 1 +b 3 x 2 find the prediction variance in the region (a) For data at three vertices (omitting (1,1))

Interpolation vs. Extrapolation At origin. At 3 vertices. At (1,1)

Standard error contours

Data at four vertices Now And Error at vertices At the origin minimum is How can we reduce error without adding points?

Graphical Comparison of Standard Errors Three pointsFour points

Problems The pairs (0,0), (1,1), (2,1) represent strain (millistrains) and stress (ksi) measurements. –Estimate Young’s modulus using regression. –Calculate the error in Young modulus using cross validation both from the definition and from the formula on Slide 5. Repeat the example of y=x, using only data at x=3,6,9,…,30. Use the same noise values as given for these points in the notes for Slide 4.