CHAPTER 8 Linear Regression. Residuals Slide 8 - 2  The model won’t be perfect, regardless of the line we draw.  Some points will be above the line.

Slides:



Advertisements
Similar presentations
Copyright © 2010 Pearson Education, Inc. Slide The correlation between two scores X and Y equals 0.8. If both the X scores and the Y scores are converted.
Advertisements

Chapter 8 Linear Regression
Linear Regression.  The following is a scatterplot of total fat versus protein for 30 items on the Burger King menu:  The model won’t be perfect, regardless.
Chapter 8 Linear regression
Chapter 8 Linear regression
Linear Regression Copyright © 2010, 2007, 2004 Pearson Education, Inc.
Copyright © 2010 Pearson Education, Inc. Chapter 8 Linear Regression.
Copyright © 2010 Pearson Education, Inc. Slide
Residuals Revisited.   The linear model we are using assumes that the relationship between the two variables is a perfect straight line.  The residuals.
Chapter 8 Linear Regression.
Extrapolation: Reaching Beyond the Data
Warm up Use calculator to find r,, a, b. Chapter 8 LSRL-Least Squares Regression Line.
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 8 Linear Regression © 2010 Pearson Education 1.
Statistics Residuals and Regression. Warm-up In northern cities roads are salted to keep ice from freezing on the roadways between 0 and -9.5 ° C. Suppose.
CHAPTER 8: LINEAR REGRESSION
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 7 Linear Regression.
Chapter 5: Regression1 Chapter 5 Relationships: Regression.
Chapter 8: Linear Regression
AP Statistics Chapter 8: Linear Regression
Chapter 8: Linear Regression
Introduction to Linear Regression and Correlation Analysis
1 Chapter 3: Examining Relationships 3.1Scatterplots 3.2Correlation 3.3Least-Squares Regression.
Inferences for Regression
CHAPTER 8 LEAST SQUARES LINE. IMPORTANT TERMS – WRITE THESE DOWN WE WILL WORK ON THE DEFINITIONS AS WE GO! Model Linear Model Predicted value Residuals.
Chapter 151 Describing Relationships: Regression, Prediction, and Causation.
Notes Bivariate Data Chapters Bivariate Data Explores relationships between two quantitative variables.
Notes Bivariate Data Chapters Bivariate Data Explores relationships between two quantitative variables.
 The equation used to calculate Cab Fare is y = 0.75x where y is the cost and x is the number of miles traveled. 1. What is the slope in this equation?
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 4 Section 2 – Slide 1 of 20 Chapter 4 Section 2 Least-Squares Regression.
Slide 8- 1 Copyright © 2010 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Business Statistics First Edition.
Objective: Understanding and using linear regression Answer the following questions: (c) If one house is larger in size than another, do you think it affects.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 8 Linear Regression *The Linear Model *Residuals *Best Fit Line *Correlation and the Line *Predicated Values *Regression.
Chapter 8 Linear Regression. Slide 8- 2 Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression (3)
Chapter 8 Linear Regression
Examining Bivariate Data Unit 3 – Statistics. Some Vocabulary Response aka Dependent Variable –Measures an outcome of a study Explanatory aka Independent.
 Find the Least Squares Regression Line and interpret its slope, y-intercept, and the coefficients of correlation and determination  Justify the regression.
1 Association  Variables –Response – an outcome variable whose values exhibit variability. –Explanatory – a variable that we use to try to explain the.
Chapter 8 Linear Regression HOW CAN A MODEL BE CREATED WHICH REPRESENTS THE LINEAR RELATIONSHIP BETWEEN TWO QUANTITATIVE VARIABLES?
Chapter 8 Linear Regression. Objectives & Learning Goals Understand Linear Regression (linear modeling): Create and interpret a linear regression model.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
Chapter 8 Linear Regression. Fat Versus Protein: An Example 30 items on the Burger King menu:
Linear Regression Chapter 8. Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King.
Copyright © 2010 Pearson Education, Inc. Chapter 8 Linear Regression.
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 7, Slide 1 Chapter 7 Linear Regression.
Statistics 8 Linear Regression. Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King.
Honors Statistics Chapter 8 Linear Regression. Objectives: Linear model Predicted value Residuals Least squares Regression to the mean Regression line.
Part II Exploring Relationships Between Variables.
AP Statistics Chapter 8 Linear Regression. Objectives: Linear model Predicted value Residuals Least squares Regression to the mean Regression line Line.
Training Activity 4 (part 2)
Finding the Best Fit Line
Inference for Regression
Chapter 8 Linear Regression.
(Residuals and
Chapter 7 Linear Regression.
Chapter 8 Linear Regression Copyright © 2010 Pearson Education, Inc.
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
Finding the Best Fit Line
Lecture Slides Elementary Statistics Thirteenth Edition
Chapter 8 Part 2 Linear Regression
Looking at Scatterplots
No notecard for this quiz!!
Chapter 8 Part 1 Linear Regression
Chi-square Test for Goodness of Fit (GOF)
Algebra Review The equation of a straight line y = mx + b
Presentation transcript:

CHAPTER 8 Linear Regression

Residuals Slide  The model won’t be perfect, regardless of the line we draw.  Some points will be above the line and some will be below.  The estimate made from a model is the predicted value (denoted as ).

Residuals (cont.) Slide  The difference between the observed value and its associated predicted value is called the residual.  To find the residuals, we always subtract the predicted value from the observed one:

Residuals (cont.)  A negative residual means the predicted value’s too big (an overestimate).  A positive residual means the predicted value’s too small (an underestimate).  In the figure, the estimated fat of the BK Broiler chicken sandwich is 36 g, while the true value of fat is 25 g, so the residual is –11 g of fat. Slide 8 - 4

“Best Fit” Means Least Squares Slide  Some residuals are positive, others are negative, and, on average, they cancel each other out.  So, we can’t assess how well the line fits by adding up all the residuals.  The smaller the sum, the better the fit.  The line of best fit is the line for which the sum of the squared residuals is smallest, the least squares line.

The Regression Line in Real Units Slide  Remember from Algebra that a straight line can be written as:   In Statistics we use a slightly different notation:  We write to emphasize that the points that satisfy this equation are just our predicted values, not the actual data values.  This model says that our predictions from our model follow a straight line.  If the model is a good one, the data values will scatter closely around it.

The Regression Line in Real Units(cont.) Slide  We write b 1 and b 0 for the slope and intercept of the line.  b 1 is the slope, which tells us how rapidly changes with respect to x.  b 0 is the y-intercept, which tells where the line crosses (intercepts) the y -axis.  Our slope is always in units of y per unit of x.  Our intercept is always in units of y.

Residuals Revisited (cont.) Slide  Residuals help us to see whether the model makes sense.  When a regression model is appropriate, nothing interesting should be left behind.  After we fit a regression model, we usually plot the residuals in the hope of finding…nothing.

Residuals Revisited (cont.) Slide  The residuals for the BK menu regression look appropriately boring:

R 2 —The Variation Accounted For (cont.) Slide  The squared correlation, r 2, gives the fraction of the data’s variance accounted for by the model.  Thus, 1 – r 2 is the fraction of the original variance left in the residuals.  For the BK model, r 2 = = 0.69, so 31% of the variability in total fat has been left in the residuals.  When interpreting a regression model you need to Tell what R 2 means.  In the BK example, 69% of the variation in total fat is accounted for by variation in the protein content.

Reporting R 2 Slide  Along with the slope and intercept for a regression, you should always report R 2 so that readers can judge for themselves how successful the regression is at fitting the data.

Assumptions and Conditions Slide  Quantitative Variables Condition:  Regression can only be done on two quantitative variables (and not two categorical variables), so make sure to check this condition.  Straight Enough Condition:  The linear model assumes that the relationship between the variables is linear.  A scatterplot will let you check that the assumption is reasonable.

Assumptions and Conditions (cont.) Slide  If the scatterplot is not straight enough, stop here.  You can’t use a linear model for any two variables, even if they are related.  They must have a linear association or the model won’t mean a thing.  Some nonlinear relationships can be saved by re- expressing the data to make the scatterplot more linear.

Assumptions and Conditions (cont.) Slide  Outlier Condition:  Watch out for outliers.  Outlying points can dramatically change a regression model.  Outliers can even change the sign of the slope, misleading us about the underlying relationship between the variables.  If the data seem to clump or cluster in the scatterplot, that could be a sign of trouble worth looking into further.