LINEAR REGRESSIONS: Cricket example About lines Line as a model:

Slides:



Advertisements
Similar presentations
Kin 304 Regression Linear Regression Least Sum of Squares
Advertisements

Stat 1301 More on Regression. Outline of Lecture 1. Regression Effect and Regression Fallacy 2. Regression Line as Least Squares Line 3. Extrapolation.
Linear Regression.  The following is a scatterplot of total fat versus protein for 30 items on the Burger King menu:  The model won’t be perfect, regardless.
1 Functions and Applications
Linear Regression Copyright © 2010, 2007, 2004 Pearson Education, Inc.
Some terminology When the relation between variables are expressed in this manner, we call the relevant equation(s) mathematical models The intercept and.
Regression What is regression to the mean?
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 8 Linear Regression © 2010 Pearson Education 1.
CHAPTER 8: LINEAR REGRESSION
Math 3680 Lecture #19 Correlation and Regression.
Correlations and scatterplots -- Optical illusion ? -- Finding the marginal distributions in the scatterplots (shoe size vs. hours of TV) Regressions --
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
LSRL Least Squares Regression Line
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Regression and Correlation
Optical illusion ? Correlation ( r or R or  ) -- One-number summary of the strength of a relationship -- How to recognize -- How to compute Regressions.
Business Statistics - QBM117 Least squares regression.
LINEAR REGRESSIONS: About lines Line as a model: Understanding the slope Predicted values Residuals How to pick a line? Least squares criterion “Point.
REGRESSION Predict future scores on Y based on measured scores on X Predictions are based on a correlation from a sample where both X and Y were measured.
1 Relationships We have examined how to measure relationships between two categorical variables (chi-square) one categorical variable and one measurement.
Chapters 10 and 11: Using Regression to Predict Math 1680.
Simple Linear Regression 1. 2 I want to start this section with a story. Imagine we take everyone in the class and line them up from shortest to tallest.
Correlation and regression 1: Correlation Coefficient
Biostatistics Unit 9 – Regression and Correlation.
1 FORECASTING Regression Analysis Aslı Sencer Graduate Program in Business Information Systems.
Statistics & Econometrics Statistics & Econometrics Statistics & Econometrics Statistics & Econometrics Statistics & Econometrics Statistics & Econometrics.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
Regression Regression relationship = trend + scatter
Objective: Understanding and using linear regression Answer the following questions: (c) If one house is larger in size than another, do you think it affects.
Relationships If we are doing a study which involves more than one variable, how can we tell if there is a relationship between two (or more) of the.
11/23/2015Slide 1 Using a combination of tables and plots from SPSS plus spreadsheets from Excel, we will show the linkage between correlation and linear.
Chapter 8 Linear Regression. Slide 8- 2 Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the.
STA291 Statistical Methods Lecture LINEar Association o r measures “closeness” of data to the “best” line. What line is that? And best in what terms.
CHAPTER 5 Regression BPS - 5TH ED.CHAPTER 5 1. PREDICTION VIA REGRESSION LINE NUMBER OF NEW BIRDS AND PERCENT RETURNING BPS - 5TH ED.CHAPTER 5 2.
 Find the Least Squares Regression Line and interpret its slope, y-intercept, and the coefficients of correlation and determination  Justify the regression.
7-3 Line of Best Fit Objectives
Lecture 5 Chapter 4. Relationships: Regression Student version.
Least Squares Regression Remember y = mx + b? It’s time for an upgrade… A regression line is a line that describes how a response variable y changes as.
9.2 Linear Regression Key Concepts: –Residuals –Least Squares Criterion –Regression Line –Using a Regression Equation to Make Predictions.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 3 Association: Contingency, Correlation, and Regression Section 3.3 Predicting the Outcome.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
LEAST-SQUARES REGRESSION 3.2 Least Squares Regression Line and Residuals.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
LESSON 6: REGRESSION 2/21/12 EDUC 502: Introduction to Statistics.
Linear Regression Chapter 8. Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Describing the Relation between Two Variables 4.
Chapters 8 Linear Regression. Correlation and Regression Correlation = linear relationship between two variables. Summarize relationship with line. Called.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Statistics 8 Linear Regression. Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King.
Linear Regression 1 Sociology 5811 Lecture 19 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
The simple linear regression model and parameter estimation
1 Functions and Applications
Unit 4 LSRL.
Regression and Correlation
Statistics 200 Lecture #5 Tuesday, September 6, 2016
Chapter 5 LSRL.
LSRL Least Squares Regression Line
Regression Fallacy.
Chapter 8 Part 2 Linear Regression
3 4 Chapter Describing the Relation between Two Variables
Least-Squares Regression
Chapter 8 Part 1 Linear Regression
Chapter 5 LSRL.
Correlation and Regression
Least-Squares Regression
Algebra Review The equation of a straight line y = mx + b
A medical researcher wishes to determine how the dosage (in mg) of a drug affects the heart rate of the patient. Find the correlation coefficient & interpret.
Lesson 2.2 Linear Regression.
Ch 9.
Presentation transcript:

LINEAR REGRESSIONS: Cricket example About lines Line as a model: Understanding the slope Predicted values Residuals How to pick a line? Least squares criterion “Point of Means” Slope is really just correlation Variance and what happens to it Variance of original y, predicted y, residuals “R squared” Residual plots More examples

When x goes up 1 unit, y goes up b units. About Lines y = mx + b y = a + bx slope = b y intercept a Slope When x goes up 1 unit, y goes up b units.

What can we do with the line? Understand the relationship ( Study the slope! ) Predict values of Y based on values of X Interpolation Extrapolation

Extrapolation Interpolation: Using a model to estimate Y for an X value within the range on which the model was based. Extrapolation: Estimating based on an X value outside the range. Interpolation Good, Extrapolation Bad.

How good is the line? RESIDUALS: For each actual point, we can compute Actual Y Predicted Y Residual ( = actual minus predicted ) Small residuals  GOOD No pattern in residuals  GOOD

How to choose the line Theory: Least Squares Criterion Choose the line with the smallest SUM OF SQUARES of the residuals. Or: Make the STANDARD DEVIATION of the residuals be as small as possible. Practice: Line of means Slope is really just r --- When X goes up one SD, Y goes up r SDs Really, b = r sy / sx

Linear models and non-linear models Model A: Model B: y = a + bx + error y = a x1/2 + error Model B has smaller errors. Is it a better model?

aa opas asl poasie ;aaslkf 4-9043578 y = 453209)_(*_n &*^(*LKH l;j;)(*&)(*& + error This model has even smaller errors. In fact, zero errors. Tradeoff: Small errors vs. complexity. (We’ll only consider linear models.)

The “Regression” Effect A preschool program attempts to boost children’s reading scores. Children are given a pre-test and a post-test. Pre-test: mean score ≈ 100, SD ≈ 10 Post-test: mean score ≈ 100, SD ≈ 10 The program seems to have no effect.

A closer look at the data shows a surprising result: Children who were below average on the pre-test tended to gain about 5-10 points on the post-test Children who were above average on the pre-test tended to lose about 5-10 points on the post-test.

A closer look at the data shows a surprising result: Children who were below average on the pre-test tended to gain about 5-10 points on the post-test Children who were above average on the pre-test tended to lose about 5-10 points on the post-test. Maybe we should provide the program only for children whose pre-test scores are below average?

Fact: In most test–retest and analogous situations, the bottom group on the first test will on average tend to improve, while the top group on the first test will on average tend to do worse. Other examples: • Students who score high on the midterm tend on average to score high on the final, but not as high. • An athlete who has a good rookie year tends to slump in his or her second year. (“Sophomore jinx”, "Sports Illustrated Jinx") • Tall fathers tend to have sons who are tall, but not as tall. (Galton’s original example!)

It works the other way, too: • Students who score high on the final tend to have scored high on the midterm, but not as high. • Tall sons tend to have fathers who are tall, but not as tall. • Students who did well on the post-test showed improvements, on average, of 5-10 points, while students who did poorly on the post-test dropped an average of 5-10 points.

Students can do well on the pretest… -- because they are good readers, or -- because they get lucky. The good readers, on average, do exactly as well on the post-test. The lucky group, on average, score lower. Students can get unlucky, too, but fewer of that group are among the high-scorers on the pre-test. So the top group on the pre-test, on average, tends to score a little lower on the post-test.