Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.

Slides:



Advertisements
Similar presentations
Young Modulus Example The pairs (1,1), (2,2), (4,3) represent strain (millistrains) and stress (ksi) measurements. Estimate Young modulus using the three.
Advertisements

Sampling plans for linear regression
3.3 Hypothesis Testing in Multiple Linear Regression
Kin 304 Regression Linear Regression Least Sum of Squares
Uncertainty in fall time surrogate Prediction variance vs. data sensitivity – Non-uniform noise – Example Uncertainty in fall time data Bootstrapping.
Kriging.
Cost of surrogates In linear regression, the process of fitting involves solving a set of linear equations once. For moving least squares, we need to.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Greg C Elvers.
Linear regression models
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
1 G Lect 14b G Lecture 14b Within subjects contrasts Types of Within-subject contrasts Individual vs pooled contrasts Example Between subject.
Data mining in 1D: curve fitting
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
Petter Mostad Linear regression Petter Mostad
Linear Regression Example Data
Curve fit noise=randn(1,30); x=1:1:30; y=x+noise ………………………………… [p,s]=polyfit(x,y,1);
Simple Linear Regression Analysis
Interpolation, extrapolation and regression Interpolation is mathematically contrasted to regression or least-squares fit As important is the contrast.
Objectives of Multiple Regression
T-test Mechanics. Z-score If we know the population mean and standard deviation, for any value of X we can compute a z-score Z-score tells us how far.
Simple Linear Regression Models
1 FORECASTING Regression Analysis Aslı Sencer Graduate Program in Business Information Systems.
Review of Statistical Models and Linear Regression Concepts STAT E-150 Statistical Methods.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
1 Chapter 3 Multiple Linear Regression Multiple Regression Models Suppose that the yield in pounds of conversion in a chemical process depends.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Chapter 13 Multiple Regression
Section 12.3 Regression Analysis HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2008 by Hawkes Learning Systems/Quant Systems, Inc. All.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
1 Introduction What does it mean when there is a strong positive correlation between x and y ? Regression analysis aims to find a precise formula to relate.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Principles of Extrapolation
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
Lecture 10 Introduction to Linear Regression and Correlation Analysis.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
Univariate Point Estimation Confidence Interval Estimation Bivariate: Linear Regression Multivariate: Multiple Regression 1 Chapter 4: Statistical Approaches.
Hypothesis Testing Example 3: Test the hypothesis that the average content of containers of a particular lubricant is 10 litters if the contents of random.
Curve fit metrics When we fit a curve to data we ask: –What is the error metric for the best fit? –What is more accurate, the data or the fit? This lecture.
The Least Squares Regression Line. The problem with drawing line of best fit by eye is that the line drawn will vary from person to person. Instead, use.
Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple.
The simple linear regression model and parameter estimation
Questions from lectures
Regression Analysis AGEC 784.
Statistical Data Analysis - Lecture /04/03
10.2 Regression If the value of the correlation coefficient is significant, the next step is to determine the equation of the regression line which is.
Kin 304 Regression Linear Regression Least Sum of Squares
CSE 4705 Artificial Intelligence
Chapter 12: Regression Diagnostics
BPK 304W Regression Linear Regression Least Sum of Squares
Correlation and Regression
Statistical Methods For Engineers
Curve fit metrics When we fit a curve to data we ask:
Pattern Recognition and Machine Learning
Curve fit metrics When we fit a curve to data we ask:
11C Line of Best Fit By Eye, 11D Linear Regression
Adequacy of Linear Regression Models
Adequacy of Linear Regression Models
3.2. SIMPLE LINEAR REGRESSION
Regression and Correlation of Data
3 basic analytical tasks in bivariate (or multivariate) analyses:
Presentation transcript:

Global predictors of regression fidelity A single number to characterize the overall quality of the surrogate. Equivalence measures –Coefficient of multiple determination –Adjusted coefficient of multiple determination Prediction accuracy measures –Model independent: Cross validation error –Model dependent: Standard error

Linear Regression

Coefficient of multiple determination Equivalence of surrogate with data is often measured by how much of the variance in the data is captured by the surrogate. Coefficient of multiple determination and adjusted version

R 2 does not reflect accuracy Compare y1=x to y2=0.1x plus same noise (normally distributed with zero mean and standard deviation of 1. Estimate the average errors between the function (red) and surrogate (blue). R 2 = R 2 =0.3016

Cross validation Validation consists of checking the surrogate at a set of validation points. This may be considered wasteful because we do not use all the points for fitting the best possible surrogate. Cross validation divides data into n g groups. Fit the approximation to n g -1 groups, and use last group to estimate error. Repeat for each group. When each group consists of one point, error often called PRESS (prediction error sum of squares) Calculate error at each point and then present r.m.s error For linear regression can be shown that

Model based error for linear regression

Comparison of errors

Problems regression accuracy The pairs (0,0), (1,1), (2,1) represent strain (millistrains) and stress (ksi) measurements. –Estimate Young’s modulus using regression. –Calculate the error in Young modulus using cross validation both from the definition and from the formula on Slide 5. SolutionSolution Repeat the example of y=x, using only data at x=3,6,9,…,30. Use the same noise values as given for these points in the notes for Slide 4. Solution Solution

Prediction variance in Linear Regression Assumptions on noise in linear regression allow us to estimate the prediction variance due to the noise at any point. Prediction variance is usually large when you are far from a data point. We distinguish between interpolation, when we are in the convex hull of the data points, and extrapolation where we are outside. Extrapolation is associated with larger errors, and in high dimensions it usually cannot be avoided.

Prediction variance Linear regression model Define then With some algebra Standard error

Example of prediction variance For a linear polynomial RS y=b 1 +b 2 x 1 +b 3 x 2 find the prediction variance in the region (a) For data at three vertices (omitting (1,1))

Interpolation vs. Extrapolation At origin. At 3 vertices. At (1,1)

Standard error contours

Data at four vertices Now And Error at vertices At the origin minimum is

Graphical Comparison of Standard Errors Three pointsFour points

Problems prediction variance Redo analytically the four point example, when the data points are not at the corners but inside the domain, at What does the difference in the results tells you? For a grid of 3x3 data points, compare the standard error contours for linear and quadratic fits.