The Squared Correlation r2 – What Does It Tell Us?

Slides:



Advertisements
Similar presentations
Coefficient of Determination- R²
Advertisements

AP Statistics Section 3.2 C Coefficient of Determination
The Role of r2 in Regression Target Goal: I can use r2 to explain the variation of y that is explained by the LSRL. D4: 3.2b Hw: pg 191 – 43, 46, 48,
Statistics and Quantitative Analysis U4320
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Statistics Measures of Regression and Prediction Intervals.
 Coefficient of Determination Section 4.3 Alan Craig
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Definition  Regression Model  Regression Equation Y i =  0 +  1 X i ^ Given a collection of paired data, the regression equation algebraically describes.
ASSESSING THE STRENGTH OF THE REGRESSION MODEL. Assessing the Model’s Strength Although the best straight line through a set of points may have been found.
Probability & Statistics for Engineers & Scientists, by Walpole, Myers, Myers & Ye ~ Chapter 11 Notes Class notes for ISE 201 San Jose State University.
Regression Analysis In regression analysis we analyze the relationship between two or more variables. The relationship between two or more variables could.
BCOR 1020 Business Statistics
Lecture 19 Simple linear regression (Review, 18.5, 18.8)
Linear Regression and Correlation Topic 18. Linear Regression  Is the link between two factors i.e. one value depends on the other.  E.g. Drivers age.
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
Anthony Greene1 Regression Using Correlation To Make Predictions.
AP STATISTICS LESSON 3 – 3 LEAST – SQUARES REGRESSION.
Further Topics in Regression Analysis Objectives: By the end of this section, I will be able to… 1) Explain prediction error, calculate SSE, and.
Linear Regression Least Squares Method: the Meaning of r 2.
Class 4 Simple Linear Regression. Regression Analysis Reality is thought to behave in a manner which may be simulated (predicted) to an acceptable degree.
Regression. Population Covariance and Correlation.
Chapter 5 Residuals, Residual Plots, & Influential points.
Educ 200C Wed. Oct 3, Variation What is it? What does it look like in a data set?
Regression Analysis Relationship with one independent variable.
Section 3.2C. The regression line can be found using the calculator Put the data in L1 and L2. Press Stat – Calc - #8 (or 4) - enter To get the correlation.
STA291 Statistical Methods Lecture LINEar Association o r measures “closeness” of data to the “best” line. What line is that? And best in what terms.
Standard Deviation Lecture 18 Sec Tue, Feb 15, 2005.
Review Lecture 51 Tue, Dec 13, Chapter 1 Sections 1.1 – 1.4. Sections 1.1 – 1.4. Be familiar with the language and principles of hypothesis testing.
Dept of Bioenvironmental Systems Engineering National Taiwan University Lab for Remote Sensing Hydrology and Spatial Modeling STATISTICS Linear Statistical.
AP Statistics HW: p. 165 #42, 44, 45 Obj: to understand the meaning of r 2 and to use residual plots Do Now: On your calculator select: 2 ND ; 0; DIAGNOSTIC.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 3 Association: Contingency, Correlation, and Regression Section 3.3 Predicting the Outcome.
Residuals Recall that the vertical distances from the points to the least-squares regression line are as small as possible.  Because those vertical distances.
Chapter Thirteen Bivariate Correlation and Regression Chapter Thirteen.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Section 9.3 Measures of Regression and Prediction Intervals.
AP STATISTICS LESSON 3 – 3 (DAY 2) The role of r 2 in regression.
© 2001 Prentice-Hall, Inc.Chap 13-1 BA 201 Lecture 19 Measure of Variation in the Simple Linear Regression Model (Data)Data.
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Warm-up Get a sheet of computer paper/construction paper from the front of the room, and create your very own paper airplane. Try to create planes with.
The simple linear regression model and parameter estimation
Lecture #26 Thursday, November 17, 2016 Textbook: 14.1 and 14.3
Linear Regression Using Excel
10.3 Coefficient of Determination and Standard Error of the Estimate
Relationship with one independent variable
Simple Linear Regression
Coefficient of Determination
Linear Regression.
AP STATISTICS LESSON 3 – 3 (DAY 2)
Least-Squares Regression
^ y = a + bx Stats Chapter 5 - Least Squares Regression
Least Squares Method: the Meaning of r2
Calculating the Least Squares Regression Line
Correlation and Regression-III
Relationship with one independent variable
Calculating the Least Squares Regression Line
Standard Deviation Lecture 20 Sec Fri, Feb 23, 2007.
Correlation and Regression
Least-Squares Regression
HW# : Complete the last slide
Linear Regression and Correlation
Standard Deviation Lecture 20 Sec Fri, Sep 28, 2007.
Linear Regression and Correlation
Standard Deviation Lecture 18 Sec Wed, Oct 4, 2006.
Simple Linear Regression
Simple Linear Regression
The Squared Correlation r2 – What Does It Tell Us?
MGS 3100 Business Analysis Regression Feb 18, 2016
Chapter 14, part C Goodness of Fit..
Presentation transcript:

The Squared Correlation r2 – What Does It Tell Us? Lecture 51 Sec. 13.9 Mon, Dec 12, 2005

Residual Sum of Squares Recall that the line of “best” fit was that line with the smallest sum of squared residuals. This is also called the residual sum of squares:

Other Sums of Squares There are two other sums of squares associated with y. The regression sum of squares: The total sum of squares:

Other Sums of Squares The regression sum of squares, SSR, measures the variability in y that is predicted by the model, i.e., the variability in y^. The total sum of squares, SST, measures the observed variability in y.

Example – SST, SSR, and SSE Plot the data in Example 13.14, p. 800, withy. 20 18 16 14 12 10 8 8 10 12 14 16

Example – SST, SSR, and SSE The deviations of y fromy (observed). 20 18 16 14 12 10 8 8 10 12 14 16

Example – SST, SSR, and SSE The deviations of y^ fromy (predicted). 20 18 16 14 12 10 8 8 10 12 14 16

Example – SST, SSR, and SSE The deviations of y from y^ (residual deviations). 20 18 16 14 12 10 8 8 10 12 14 16

The Squared Correlation It turns out that It also turns out that

Explaining Variation One goal of regression is to “explain” the variation in y. For example, if x were height and y were weight, how would we explain the variation in weight? That is, why do some people weigh more than others? Or if x were the hours spent studying for a math test and y were the score on the test, how would we explain the variation in scores? That is, why do some people score higher than others?

Explaining Variation A certain amount of the variation in y can be explained by the variation in x. Some people weigh more than others because they are taller. Some people score higher on math tests because they studied more. But that is never the full explanation. Not all taller people weigh more. Not everyone who studies longer scores higher.

Explaining Variation High degree of correlation between x and y  variation in x explains most of the variation in y. Low degree of correlation between x and y  variation in x explains only a little of the variation in y. In other words, the amount of variation in y that is explained by the variation in x should be related to r.

Explaining Variation Statisticians consider the predicted variation SSR to be the amount of variation in y (SST) that is explained by the model. The remaining variation in y, i.e., residual variation SSE, is the amount that is not explained by the model.

Explaining Variation SST = SSE + SSR

Explaining Variation SST = SSE + SSR Total variation in y (to be explained)

Explaining Variation SST = SSE + SSR Total variation in y (to be explained) Variation in y that is explained by the model

unexplained by the model Explaining Variation Variation in y that is unexplained by the model SST = SSE + SSR Total variation in y Variation in y that is explained by the model

Example – SST, SSR, and SSE The total (observed) variation in y. 20 18 16 14 12 10 8 8 10 12 14 16

Example – SST, SSR, and SSE The variation in y that is explained by the model. 20 18 16 14 12 10 8 8 10 12 14 16

Example – SST, SSR, and SSE The variation in y that is not explained by the model. 20 18 16 14 12 10 8 8 10 12 14 16

Explaining Variation Therefore, r2 is the proportion of variation in y that is explained by the model and 1 – r2 is the proportion that is not explained by the model.

TI-83 – Calculating r2 To calculate r2 on the TI-83, Follow the procedure that produces the regression line and r. In the same window, the TI-83 reports r2.

Let’s Do It! Let’s Do It! 13.3, p. 819 – Oil-Change Data. Do part (b) on the TI-83. How much of the variation in repair costs is explained by frequency of oil change?