Section 6.2 Prediction.

Slides:



Advertisements
Similar presentations
AP Statistics Section 3.2 C Coefficient of Determination
Advertisements

Lesson 10: Linear Regression and Correlation
Kin 304 Regression Linear Regression Least Sum of Squares
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
AP Statistics Chapter 3 Practice Problems
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Linear Regression and Correlation
Correlation and Linear Regression
McGraw-Hill/Irwin Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 13 Linear Regression and Correlation.
Linear Regression and Correlation
Chapter 6 & 7 Linear Regression & Correlation
AP STATISTICS LESSON 3 – 3 LEAST – SQUARES REGRESSION.
Topic 10 - Linear Regression Least squares principle - pages 301 – – 309 Hypothesis tests/confidence intervals/prediction intervals for regression.
Regression Regression relationship = trend + scatter
Copyright ©2011 Nelson Education Limited Linear Regression and Correlation CHAPTER 12.
Section 4.2: Least-Squares Regression Goal: Fit a straight line to a set of points as a way to describe the relationship between the X and Y variables.
Relationships If we are doing a study which involves more than one variable, how can we tell if there is a relationship between two (or more) of the.
Section 3.2C. The regression line can be found using the calculator Put the data in L1 and L2. Press Stat – Calc - #8 (or 4) - enter To get the correlation.
Chapter 9: Correlation and Regression Analysis. Correlation Correlation is a numerical way to measure the strength and direction of a linear association.
CHAPTER 5 CORRELATION & LINEAR REGRESSION. GOAL : Understand and interpret the terms dependent variable and independent variable. Draw a scatter diagram.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 3 Association: Contingency, Correlation, and Regression Section 3.3 Predicting the Outcome.
Residuals Recall that the vertical distances from the points to the least-squares regression line are as small as possible.  Because those vertical distances.
Linear Regression and Correlation Chapter GOALS 1. Understand and interpret the terms dependent and independent variable. 2. Calculate and interpret.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin Linear Regression and Correlation Chapter 13.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
Describing Relationships. Least-Squares Regression  A method for finding a line that summarizes the relationship between two variables Only in a specific.
Correlation and Linear Regression Chapter 13 McGraw-Hill/Irwin Copyright © 2012 by The McGraw-Hill Companies, Inc. All rights reserved.
The simple linear regression model and parameter estimation
CHAPTER 3 Describing Relationships
Chapter 4.2 Notes LSRL.
Sections Review.
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
(Residuals and
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
Regression Analysis PhD Course.
Prepared by Lee Revere and John Large
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Least-Squares Regression
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
The Weather Turbulence
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Least-Squares Regression
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Warmup A study was done comparing the number of registered automatic weapons (in thousands) along with the murder rate (in murders per 100,000) for 8.
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Linear Regression and Correlation
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Presentation transcript:

Section 6.2 Prediction

Understanding Prediction 3 Key Principles of Prediction Prediction is based on fitting some “model” to a set of data. We are using a straight line model for our data. Prediction works best when the model fits the data closely. Prediction outside the range of data is risky. Imagine if we used a child growth chart to predict height at age 40!

The square of the regression, or r2, is a numerical quantity that tells us how well the least squares line predicts values of the response variable, y. r2 in regression The square of the correlation, r2, is the fraction of the variation in the values of y that is explained by the least-squares regression of y on x. The idea is that when there is a linear relationship, some of the variation in y is accounted for the fact that, as x changes, it pulls y along the regression line with it.

The coefficient of determination (denoted by r2) is a key output of regression analysis. It is interpreted as the proportion of the variance in the dependent variable that is predictable from the independent variable. The coefficient of determination is the square of the correlation (r) between predicted y scores and actual y scores; thus, it ranges from 0 to 1. With linear regression, the coefficient of determination is also equal to the square of the correlation between x and y scores. An r2 of 0 means that the dependent variable cannot be predicted from the independent variable. An r2 of 1 means the dependent variable can be predicted without error from the independent variable. An r2 between 0 and 1 indicates the extent to which the dependent variable is predictable. An r2 of 0.10 means that 10 percent of the variance in Y is predictable from X; an r2 of 0.20 means that 20 percent is predictable; and so on.

An r2 of 0 means that the dependent variable cannot be predicted from the independent variable. An r2 of 1 means the dependent variable can be predicted without error from the independent variable. An r2 between 0 and 1 indicates the extent to which the dependent variable is predictable. An r2 of 0.10 means that 10 percent of the variance in Y is predictable from X; an r2 of 0.20 means that 20 percent is predictable; and so on.

Enter the data into your calculator to find r and r2. Example: The table below lists data on body weight and backpack weight for 8 students. Body Weight 120 187 109 103 131 165 158 116 Backpack Weight 26 30 24 29 35 31 28 Enter the data into your calculator to find r and r2.

Because r2 = 0.632 for these data, about 63% of the observed variation in backpack weight can be explained by the straight line pattern What about the other 37%? It is due to other factors, such as physical fitness and motivation.

Exercises True or False: “When r = 0.7, this means that y can be predicted from x for 70% of the individuals in the sample.” False. It means that 49% of the variation in y can be explained by x. It says nothing about the individuals in the sample.

2. ) True or False: “When r2 = 0 2.) True or False: “When r2 = 0.7, this means that y can be predicted from x for 70% of the individuals in the sample.” False. While this statement used the correct numerical measurement for prediction, r2, it is still interpreting it incorrectly.

A researcher uses a regression equation to predict home heating bills (dollar cost), based on home size (square feet). The correlation between predicted bills and home size is 0.70. What is the correct interpretation of this finding? 70% of the variability in home heating bills can be explained by home size. (B) 49% of the variability in home heating bills can be explained by home size. (C) For each added square foot of home size, heating bills increased by 70 cents. (D) For each added square foot of home size, heating bills increased by 49 cents. (E) None of the above.

Complete the following exercises in the text: Page 377/6.29, 6.30, 6.31, 6.33