Regression Regression relationship = trend + scatter

Slides:



Advertisements
Similar presentations
Kin 304 Regression Linear Regression Least Sum of Squares
Advertisements

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Chapter 10 Regression. Defining Regression Simple linear regression features one independent variable and one dependent variable, as in correlation the.
Regression Regression: Mathematical method for determining the best equation that reproduces a data set Linear Regression: Regression method applied with.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
Statistics for the Social Sciences
Regression and Correlation
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
Business Statistics - QBM117 Least squares regression.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
1 1 Slide Simple Linear Regression Chapter 14 BA 303 – Spring 2011.
Linear Regression Analysis
Lecture 3: Bivariate Data & Linear Regression 1.Introduction 2.Bivariate Data 3.Linear Analysis of Data a)Freehand Linear Fit b)Least Squares Fit c)Interpolation/Extrapolation.
Relationship of two variables
Simple Linear Regression
Residuals and Residual Plots Most likely a linear regression will not fit the data perfectly. The residual (e) for each data point is the ________________________.
VCE Further Maths Least Square Regression using the calculator.
Biostatistics Unit 9 – Regression and Correlation.
Linear Regression Least Squares Method: the Meaning of r 2.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Holt Algebra Curve Fitting with Linear Models 2-7 Curve Fitting with Linear Models Holt Algebra 2 Lesson Presentation Lesson Presentation.
Modeling a Linear Relationship Lecture 47 Secs – Tue, Apr 25, 2006.
Creating a Residual Plot and Investigating the Correlation Coefficient.
Modeling a Linear Relationship Lecture 44 Secs – Tue, Apr 24, 2007.
Correlation The apparent relation between two variables.
9.1B – Computing the Correlation Coefficient by Hand
7-3 Line of Best Fit Objectives
9.2 Linear Regression Key Concepts: –Residuals –Least Squares Criterion –Regression Line –Using a Regression Equation to Make Predictions.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Math 4030 – 11b Method of Least Squares. Model: Dependent (response) Variable Independent (control) Variable Random Error Objectives: Find (estimated)
Curve Fitting Pertemuan 10 Matakuliah: S0262-Analisis Numerik Tahun: 2010.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Simple Linear Regression The Coefficients of Correlation and Determination Two Quantitative Variables x variable – independent variable or explanatory.
Method 3: Least squares regression. Another method for finding the equation of a straight line which is fitted to data is known as the method of least-squares.
REGRESSION MODELS OF BEST FIT Assess the fit of a function model for bivariate (2 variables) data by plotting and analyzing residuals.
The Least Squares Regression Line. The problem with drawing line of best fit by eye is that the line drawn will vary from person to person. Instead, use.
1 Objective Given two linearly correlated variables (x and y), find the linear function (equation) that best describes the trend. Section 10.3 Regression.
1. Analyzing patterns in scatterplots 2. Correlation and linearity 3. Least-squares regression line 4. Residual plots, outliers, and influential points.
The simple linear regression model and parameter estimation
Lesson 4.5 Topic/ Objective: To use residuals to determine how well lines of fit model data. To use linear regression to find lines of best fit. To distinguish.
Sections Review.
Statistics 101 Chapter 3 Section 3.
distance prediction observed y value predicted value zero
Regression and Correlation
Kin 304 Regression Linear Regression Least Sum of Squares
Ch12.1 Simple Linear Regression
BPK 304W Regression Linear Regression Least Sum of Squares
(Residuals and
Suppose the maximum number of hours of study among students in your sample is 6. If you used the equation to predict the test score of a student who studied.
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
2-7 Curve Fitting with Linear Models Holt Algebra 2.
Linear Regression.
Linear regression Fitting a straight line to observations.
Least Squares Method: the Meaning of r2
Examining Relationships
Residuals and Residual Plots
M248: Analyzing data Block D UNIT D2 Regression.
Regression All about going back (regressing) from the points to the function from which they came. We are looking at linear regression.
11C Line of Best Fit By Eye, 11D Linear Regression
Adequacy of Linear Regression Models
Ch 4.1 & 4.2 Two dimensions concept
Adequacy of Linear Regression Models
Algebra Review The equation of a straight line y = mx + b
A medical researcher wishes to determine how the dosage (in mg) of a drug affects the heart rate of the patient. Find the correlation coefficient & interpret.
Lesson 2.2 Linear Regression.
Descriptive Statistics Univariate Data
Ch 9.
Presentation transcript:

Regression Regression relationship = trend + scatter 8 y = 5 + 2x data point (8, 25) 25 21 prediction error Regression is about fitting a line or curve to bivariate data to predict the value of a variable y based on the value of an independent variable x. Regression relationship = trend + scatter Observed value = predicted value + prediction error

Residual A line of best fit will be used to predict a value of y for a given value of x. The difference between the measured value y and the predicted value ŷ is called the residual. Residual = y-ŷ Residual = observed value – predicted value

Regression Line Obviously, we would like all these residuals to be as small as possible. A technique is least squares regression minimises the sum of the squares of the residuals, the line found by this technique is therefore called the least squares regression line of y on x, or simply the regression line.

Complete the table below Data Point (8, 25) (3, 7) (-2, -3) (x, y) Observed y-value 25 y Fitted line Predicted value / fitted value 21 Prediction error / residual 4 y -

Complete the table below Data Point (8, 25) (3, 7) (-2, -3) (x, y) Observed y-value 25 7 -3 y Fitted line Predicted value / fitted value 21 19 -1 Prediction error / residual 4 -12 -2 y -

The Least Squares Regression Line Which line? Choose the line with smallest sum of squared prediction errors. Minimise the sum of squared prediction errors Minimise

The Least Squares Regression Line There is one and only one least squares regression line for every linear regression for the least squares line but it is also true for many other lines is on the least squares line Calculator or computer gives the equation of the least squares line

Residuals Plot The pattern of residuals allows you to see if your regression line is a good fit for the data and how reliable interpolation and extrapolation will be. If the model is a good fit, the residuals will oscillate closely above and below the zero line.

Correlation coefficient = 0.8352 This is √0.6975 Temperature oF Chirps per second   69.4 15.4 69.7 14.7 71.6 16.0 75.2 15.5 76.3 14.4 79.6 15.0 80.6 17.1 82.0 82.6 17.2 83.3 16.2 83.5 17.0 84.3 18.4 88.6 20.0 93.3 19.8 Correlation coefficient = 0.8352 This is √0.6975

Predicted chirps per second Observed chirps per second Residuals Temperature oF Chirps per second   69.4 15.4 69.7 14.7 71.6 16.0 75.2 15.5 76.3 14.4 79.6 15.0 80.6 17.1 82.0 82.6 17.2 83.3 16.2 83.5 17.0 84.3 18.4 88.6 20.0 93.3 19.8 Predicted chirps per second Observed chirps per second Residuals   14.4 15.4 1.0 14.5 14.7 0.2 14.9 16.0 1.1 15.6 15.5 -0.1 15.9 -1.5 16.6 15.0 -1.6 16.8 17.1 0.3 -0.8 0.0 17.2 17.3 16.2 -1.1 17.4 17.0 -0.4 17.6 18.4 0.8 18.5 20.0 1.5 19.5 19.8 The regression line is: y = 0.2119x - 0.3091 which is what we use to get the predicted value of y. Eg. x = 69.4 oF y = 0.2119(69.4) – 0.3091 = 14.4 chirps per second Residual = Observed – Predicted Value Value

Predicted chirps per second Observed chirps per second Residuals Predicted chirps per second Observed chirps per second Residuals   14.4 15.4 1.0 14.5 14.7 0.2 14.9 16.0 1.1 15.6 15.5 -0.1 15.9 -1.5 16.6 15.0 -1.6 16.8 17.1 0.3 -0.8 0.0 17.2 17.3 16.2 -1.1 17.4 17.0 -0.4 17.6 18.4 0.8 18.5 20.0 1.5 19.5 19.8 The plot of the residuals shows that they are randomly scattered, so in this case a linear model is appropriate.