The Least-Squares Regression Line

Slides:



Advertisements
Similar presentations
Lesson 10: Linear Regression and Correlation
Advertisements

AP Statistics.  Least Squares regression is a way of finding a line that summarizes the relationship between two variables.
CHAPTER 3 Describing Relationships
McGraw-Hill/Irwin Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 13 Linear Regression and Correlation.
Overview 4.2 Introduction to Correlation 4.3 Introduction to Regression.
Chapter 6 & 7 Linear Regression & Correlation
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.1 Using Several Variables to Predict a Response.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.2 Extending the Correlation and R-Squared for Multiple.
Correlation and Regression. Section 9.1  Correlation is a relationship between 2 variables.  Data is often represented by ordered pairs (x, y) and.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 3 Describing Relationships 3.2 Least-Squares.
CHAPTER 5 CORRELATION & LINEAR REGRESSION. GOAL : Understand and interpret the terms dependent variable and independent variable. Draw a scatter diagram.
LEAST-SQUARES REGRESSION 3.2 Least Squares Regression Line and Residuals.
1 Simple Linear Regression and Correlation Least Squares Method The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Describing the Relation between Two Variables 4.
Chapters 8 Linear Regression. Correlation and Regression Correlation = linear relationship between two variables. Summarize relationship with line. Called.
Least Square Regression Line. Line of Best Fit Our objective is to fit a line in the scatterplot that fits the data the best As just seen, the best fit.
Chapter 4 Basic Estimation Techniques
Chapter 13 Multiple Regression
Linear Regression Special Topics.
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
LSRL Least Squares Regression Line
SIMPLE LINEAR REGRESSION MODEL
2.1 Equations of Lines Write the point-slope and slope-intercept forms
Chapter 13 Multiple Regression
Linear Regression/Correlation
Correlation and Regression
Ice Cream Sales vs Temperature
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Least-Squares Regression
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
The Weather Turbulence
Unit 4 Vocabulary.
Least-Squares Regression
Chapter 3 Describing Relationships Section 3.2
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Least-Squares Regression
4-5 Scatter Plots and Lines of Best Fit
Least-Squares Regression
Chapter 3: Describing Relationships
Least-Squares Regression
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Ch 4.1 & 4.2 Two dimensions concept
CHAPTER 3 Describing Relationships
3.2 – Least Squares Regression
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Regression and Categorical Predictors
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
9/27/ A Least-Squares Regression.
Created by Erin Hodgess, Houston, Texas
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Presentation transcript:

The Least-Squares Regression Line Section 4.2

Objectives Compute the least-squares regression line Use the least-squares regression line to make predictions Interpret predicted values, the slope, and the 𝑦-intercept of the least-squares regression line

Compute the least-squares regression line Objective 1 Compute the least-squares regression line

House Size Versus Sales Price The table presents the size in square feet and selling price in thousands of dollars for a sample of houses. Size (Square Feet) Selling Price ($1000s) 2521 400 2555 426 2735 428 2846 435 3028 469 3049 475 3198 488 455 In the previous section, we concluded that there is a strong positive linear association between size and sales price. We can use these data to predict the selling price of a house based on its size.

Least-Squares Regression Line The figures present scatterplots of the previous data, each with a different line superimposed. It is clear that the line in the figure on the left fits better than the line in the figure on the right. The reason is that the vertical distances are, on the whole, smaller. The line that fits best is the line for which the sum of squared vertical distances is as small as possible. This line is called the least-squares regression line.

Equation of the Least-Squares Regression Line Given ordered pairs (𝑥,𝑦), with sample means 𝑥 and 𝑦 , sample standard deviations 𝑠 𝑥 and 𝑠 𝑦 , and correlation coefficient 𝑟, the equation of the least-squares regression line for predicting 𝑦 from 𝑥 is 𝑦 = 𝑏 0 + 𝑏 1 𝑥 where 𝑏 1 =𝑟 𝑠 𝑦 𝑠 𝑥 is the slope and 𝑏 0 = 𝑦 − 𝑏 1 𝑥 . The variable we want to predict (in this case, selling price) is called the outcome variable, or response variable, and the variable we are given is called the explanatory variable, or predictor variable. In the equation of the least-squares regression line, 𝑥 represents the explanatory variable and 𝑦 represents the outcome variable.

Example – Least-Squares Regression Line Compute the least-squares regression line for predicting selling price from size. Size (Square Feet) Selling Price ($1000s) 2521 400 2555 426 2735 428 2846 435 3028 469 3049 475 3198 488 455 We first compute the sample means, standard deviations, and correlation coefficient. We obtain 𝑥 =2891.25 𝑦 =447.0 𝑠 𝑥 =269.49357 𝑠 𝑦 =29.68405 𝑟=0.9005918 The slope of the least-squares regression line is 𝑏 1 =𝑟 𝑠 𝑦 𝑠 𝑥 =0.9005918 29.68405 269.49357 = 0.09919796. The 𝒚-intercept is given by: 𝑏 0 = 𝑦 − 𝑏 1 𝑥 = 447.0 – 0.09919796(2891.25) = 160.1939 The equation of the least-squares regression line is 𝑦 = 𝑏 0 + 𝑏 1 𝑥=160.1939+0.0992𝑥.

Use the least-squares regression line to make predictions Objective 2 Use the least-squares regression line to make predictions

Predicted Value We can use the least-squares regression line to predict the value of the outcome variable by substituting a value for the explanatory variable in the equation of the least-squares regression line. Example: The equation of the least-squares regression line for predicting selling price from size is 𝑦 =160.1939+0.0992𝑥. Predict the selling price of a house of size 2800 sq. ft. Solution: 𝑦 =160.1939+0.0992 2800 =438.0 (2800, 438.0)

Objective 3 Interpret predicted values, the slope, and the 𝑦-intercept of the least-squares regression line

Interpreting the Predicted Value 𝑦 The predicted value 𝑦 can be used to estimate the average outcome for a given value of 𝑥. For any given 𝑥, the value 𝑦 is an estimate of the average 𝑦-value for all points with that 𝑥-value. Example: Use the equation of the least-squares regression line for predicting selling price from size 𝑦 =160.1939+0.0992 to estimate the average price of all houses whose size is 3000 sq. feet. Solution: 𝑦 =160.1939+0.0992=160.1939+0.0992 3000 =457.8 We estimate the average price for a house of 3000 sq. feet to be 457.8 thousand dollars.

Interpreting the 𝑦-Intercept 𝑏 0 The 𝑦-intercept 𝑏 0 is the point where the line crosses the 𝑦-axis. This has a practical interpretation only when the data contain both positive and negative values of 𝑥. If the data contain both positive and negative 𝑥-values, then the 𝑦-intercept is the estimated outcome when the value of the explanatory variable 𝑥 is 0. If the 𝑥-values are all positive or all negative, then the 𝑦-intercept 𝑏 0 does not have a useful interpretation.

Interpreting the Slope 𝑏 1 If the 𝑥-values of two points on a line differ by 1, their 𝑦-values will differ by an amount equal to the slope of the line. This fact enables us to interpret the slope 𝑏 1 of the least-squares regression line. If the values of the explanatory variable for two individuals differ by 1, their predicted values will differ by 𝑏 1 . If the values of the explanatory variable differ by an amount 𝑑, then their predicted values will differ by 𝑏 1 ∙𝑑. Example: Two houses differ in size by 150 square feet. By how much should we predict their prices to differ? Solution: The slope of the least-squares regression line is 𝑏 1 =0.0992. We predict the prices to differ by (0.0992)(150) = 14.9 thousand dollars.

You Should Know… The definitions of outcome or response variables and explanatory or predictor variables How to compute the least-squares regression line How to use the least-squares regression line to make predictions How to interpret the predicted value 𝑦 , the 𝑦-intercept 𝑏 0 , and the slope 𝑏 1 of a least-squares regression line