Warm up Use calculator to find r,, a, b. Chapter 8 LSRL-Least Squares Regression Line.

Slides:



Advertisements
Similar presentations
Linear Regression (LSRL)
Advertisements

Linear Regression.  The following is a scatterplot of total fat versus protein for 30 items on the Burger King menu:  The model won’t be perfect, regardless.
Chapter 8 Linear regression
Chapter 8 Linear regression
Linear Regression Copyright © 2010, 2007, 2004 Pearson Education, Inc.
Copyright © 2010 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 8 Linear Regression.
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 8 Linear Regression © 2010 Pearson Education 1.
Statistics Residuals and Regression. Warm-up In northern cities roads are salted to keep ice from freezing on the roadways between 0 and -9.5 ° C. Suppose.
CHAPTER 8: LINEAR REGRESSION
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 7 Linear Regression.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
LSRL Least Squares Regression Line
 The equation used to calculate Cab Fare is y = 0.75x where y is the cost and x is the number of miles traveled. 1. What is the slope in this equation?
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 8 Linear Regression. Slide 8- 2 Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression (3)
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 3 Describing Relationships 3.2 Least-Squares.
CHAPTER 5 Regression BPS - 5TH ED.CHAPTER 5 1. PREDICTION VIA REGRESSION LINE NUMBER OF NEW BIRDS AND PERCENT RETURNING BPS - 5TH ED.CHAPTER 5 2.
 Find the Least Squares Regression Line and interpret its slope, y-intercept, and the coefficients of correlation and determination  Justify the regression.
Chapter 3-Examining Relationships Scatterplots and Correlation Least-squares Regression.
Chapter 8 Linear Regression HOW CAN A MODEL BE CREATED WHICH REPRESENTS THE LINEAR RELATIONSHIP BETWEEN TWO QUANTITATIVE VARIABLES?
 Chapter 3! 1. UNIT 7 VOCABULARY – CHAPTERS 3 & 14 2.
CHAPTER 8 Linear Regression. Residuals Slide  The model won’t be perfect, regardless of the line we draw.  Some points will be above the line.
Linear Regression Day 1 – (pg )
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 3 Association: Contingency, Correlation, and Regression Section 3.3 Predicting the Outcome.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
Least Squares Regression Lines Text: Chapter 3.3 Unit 4: Notes page 58.
Unit 4 Lesson 3 (5.3) Summarizing Bivariate Data 5.3: LSRL.
Chapter 8 Linear Regression. Fat Versus Protein: An Example 30 items on the Burger King menu:
Linear Regression Chapter 8. Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King.
Chapter 7 Linear Regression. Bivariate data x – variable: is the independent or explanatory variable y- variable: is the dependent or response variable.
Chapter 5 Lesson 5.2 Summarizing Bivariate Data 5.2: LSRL.
Copyright © 2010 Pearson Education, Inc. Chapter 8 Linear Regression.
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
Unit 3 Correlation. Homework Assignment For the A: 1, 5, 7,11, 13, , 21, , 35, 37, 39, 41, 43, 45, 47 – 51, 55, 58, 59, 61, 63, 65, 69,
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 7, Slide 1 Chapter 7 Linear Regression.
Statistics 8 Linear Regression. Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King.
Honors Statistics Chapter 8 Linear Regression. Objectives: Linear model Predicted value Residuals Least squares Regression to the mean Regression line.
Part II Exploring Relationships Between Variables.
Describing Bivariate Relationships. Bivariate Relationships When exploring/describing a bivariate (x,y) relationship: Determine the Explanatory and Response.
Chapter 3 LSRL. Bivariate data x – variable: is the independent or explanatory variable y- variable: is the dependent or response variable Use x to predict.
Chapter 5 LSRL. Bivariate data x – variable: is the independent or explanatory variable y- variable: is the dependent or response variable Use x to predict.
Training Activity 4 (part 2)
CHAPTER 3 Describing Relationships
Unit 4 LSRL.
LSRL.
Chapter 4.2 Notes LSRL.
Least Squares Regression Line.
Chapter 5 LSRL.
LSRL Least Squares Regression Line
Chapter 4 Correlation.
Chapter 3.2 LSRL.
Chapter 8 Linear Regression.
Least Squares Regression Line LSRL Chapter 7-continued
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 5 LSRL.
Chapter 5 LSRL.
Chapter 5 LSRL.
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
Presentation transcript:

Warm up Use calculator to find r,, a, b

Chapter 8 LSRL-Least Squares Regression Line

 Find the Least Squares Regression Line and interpret its slope, y-intercept, and the coefficients of correlation and determination AP Statistics Objectives Ch8

Bivariate data x – variable: is the independent or explanatory variable y- variable: is the dependent or response variable Use x to predict y

Copyright © 2010, 2007, 2004 Pearson Education, Inc. Slide Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King menu:

Copyright © 2010, 2007, 2004 Pearson Education, Inc. Slide The Linear Model The correlation in this example is It says “There seems to be a linear association between these two variables,” but it doesn’t tell what that association is. We can say more about the linear relationship between two quantitative variables with a model. A model simplifies reality to help us understand underlying patterns and relationships.

Copyright © 2010, 2007, 2004 Pearson Education, Inc. Slide The Linear Model (cont.) The linear model is just an equation of a straight line through the data. The points in the scatterplot don’t all line up, but a straight line can summarize the general pattern with only a couple of parameters. The linear model can help us understand how the values are associated.

b – is the slope –it is the amount by which y increases when x increases by 1 unit a – is the y-intercept –it is the height of the line when x = 0 –in some situations, the y-intercept has no meaning - (y-hat) means the predicted y Be sure to put the hat on the y

Least Squares Regression Line LSRL bestThe line that gives the best fit to the data set minimizesThe line that minimizes the sum of the squares of the deviations from the line

X Y Scattergram 1.Plot of All (X i, Y i ) Pairs 2.Suggests How Well Model Will Fit

Thinking Challenge How would you draw a line through the points? How do you determine which line ‘fits best’?

Thinking Challenge How would you draw a line through the points? How do you determine which line ‘fits best’?

Thinking Challenge How would you draw a line through the points? How do you determine which line ‘fits best’?

Thinking Challenge How would you draw a line through the points? How do you determine which line ‘fits best’?

Thinking Challenge How would you draw a line through the points? How do you determine which line ‘fits best’?

Thinking Challenge How would you draw a line through the points? How do you determine which line ‘fits best’?

Thinking Challenge How would you draw a line through the points? How do you determine which line ‘fits best’?

Sum of the squares = y =.5(0) + 4 = 4 0 – 4 = -4 (0,0) (3,10) (6,2) (0,0) y =.5(3) + 4 = – 5.5 = 4.5 y =.5(6) + 4 = 7 2 – 7 = -5

Copyright © 2010, 2007, 2004 Pearson Education, Inc. Slide Residuals The model won’t be perfect, regardless of the line we draw. Some points will be above the line and some will be below. The estimate made from a model is the predicted value (denoted as ).

Copyright © 2010, 2007, 2004 Pearson Education, Inc. Slide Residuals (cont.) The difference between the observed value and its associated predicted value is called the residual. To find the residuals, we always subtract the predicted value from the observed one:

(0,0) (3,10) (6,2) Sum of the squares = 54 Use a calculator to find the line of best fit Find y - y -3 6 What is the sum of the deviations from the line? Will it always be zero? minimizes LSRL The line that minimizes the sum of the squares of the deviations from the line is the LSRL.

Types of Regression Models Positive Linear Relationship Negative Linear Relationship Relationship NOT Linear No Relationship

Slope: unitx increase/decreaseby For each unit increase in x, there is an approximate increase/decrease of b in y. Interpretations

The ages (in months) and heights (in inches) of seven children are given. x y Find the LSRL. Interpret the slope and correlation coefficient in the context of the problem.

Slope: unitx increase/decreaseby For each unit increase in x, there is an approximate increase/decrease of b in y. Interpretations

Correlation coefficient: There is a _________________ association between the ________ Slope: _______________ ______ _______________________________ For an increase in _______________, there is an approximate ______ of _______________________________

Correlation coefficient: strong, positive, linear age and height of children There is a strong, positive, linear association between the age and height of children. Slope: age of one month increase.34 inches in heights of children. For an increase in age of one month, there is an approximate increase of.34 inches in heights of children.

The ages (in months) and heights (in inches) of seven children are given. x y Predict the height of a child who is 4.5 years old. Predict the height of someone who is 20 years old.

Extrapolation should notThe LSRL should not be used to predict y for values of x outside the data set. It is unknown whether the pattern observed in the scatterplot continues outside this range.

The ages (in months) and heights (in inches) of seven children are given. x y Calculate x & y. Plot the point (x, y) on the LSRL. Will this point always be on the LSRL?

non-resistant The correlation coefficient and the LSRL are both non-resistant measures.

Formulas – on chart

The following statistics are found for the variables posted speed limit and the average number of accidents. Find the LSRL & predict the number of accidents for a posted speed limit of 50 mph.

Chapter 8 r a) b) c) d)

Chapter 8 r a) b) c) d)

Chapter 8 r a) b) c) d)

Chapter 8 r a) b) c) d)

Chapter 8 r a) b) c) d)

Chapter 8 r a) b) c) d)

Chapter 8 r a) b) c) d)

Chapter 8 r a) b) c) d)

Chapter 8 R 2 – Must also be interpreted when describing a regression model “With the linear regression model, _____% of the variability in _______ (response variable) is accounted for by variation in ________ (explanatory variable)” The remaining variation is due to the residuals

R 2 = +1 Examples of Approximate R 2 Values y x y x R 2 = 1 Perfect linear relationship between x and y: 100% of the variation in y is explained by variation in x

Examples of Approximate R 2 Values y x y x 0 < R 2 < 1 Weaker linear relationship between x and y: Some but not all of the variation in y is explained by variation in x

Examples of Approximate R 2 Values R 2 = 0 No linear relationship between x and y: The value of Y does not depend on x. (None of the variation in y is explained by variation in x) y x R 2 = 0

Did you say 2?Wrong. Try again. So what?

Important Note: The correlation is not given directly in this software package. You need to look in two places for it. Taking the square root of the “R squared” (coefficient of determination) is not enough. You must look at the sign of the slope too. Positive slope is a positive r-value. Negative slope is a negative r-value.

So here you should note that the slope is positive. The correlation will be positive too. Since R 2 is 0.482, r will be

Coefficient of Determination = (0.694) 2 =0.4816

With the linear regression model, 48.2% of the variability in airline fares is accounted for by the variation in distance of the flight.

There is an increase of 7.86 cents for every additional mile. There is an increase of $7.86 for every additional 100 miles.

The model predicts a flight of zero miles will cost $ The airline may have built in an initial cost to pay for some of its expenses.

8. Using those estimates, draw the line on the scatterplot.