Regression & Prediction

Slides:



Advertisements
Similar presentations
Lesson 10: Linear Regression and Correlation
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
Statistical Techniques I EXST7005 Simple Linear Regression.
R Squared. r = r = -.79 y = x y = x if x = 15, y = ? y = (15) y = if x = 6, y = ? y = (6)
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Overview Correlation Regression -Definition
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
Chapter 10 Simple Regression.
REGRESSION What is Regression? What is the Regression Equation? What is the Least-Squares Solution? How is Regression Based on Correlation? What are the.
Statistics Psych 231: Research Methods in Psychology.
Business 205. Review Chi-Square Preview Regression.
Basic Statistical Concepts Part II Psych 231: Research Methods in Psychology.
REGRESSION Predict future scores on Y based on measured scores on X Predictions are based on a correlation from a sample where both X and Y were measured.
Bivariate Regression  Assumptions  Each variable is interval/ratio  There is linear (straight line) relationship between the variables.  Normal distribution.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Chapter 2 – Simple Linear Regression - How. Here is a perfect scenario of what we want reality to look like for simple linear regression. Our two variables.
Relationships Among Variables
McGraw-Hill/Irwin Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 13 Linear Regression and Correlation.
Exploring relationships between variables Ch. 10 Scatterplots, Associations, and Correlations Ch. 10 Scatterplots, Associations, and Correlations.
Chapter 12 Correlation and Regression Part III: Additional Hypothesis Tests Renee R. Ha, Ph.D. James C. Ha, Ph.D Integrative Statistics for the Social.
Linear Regression and Correlation
Stat13-lecture 25 regression (continued, SE, t and chi-square) Simple linear regression model: Y=  0 +  1 X +  Assumption :  is normal with mean 0.
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
1 FORECASTING Regression Analysis Aslı Sencer Graduate Program in Business Information Systems.
Simple Linear Regression One reason for assessing correlation is to identify a variable that could be used to predict another variable If that is your.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Ch4 Describing Relationships Between Variables. Section 4.1: Fitting a Line by Least Squares Often we want to fit a straight line to data. For example.
Linear Regression Least Squares Method: an introduction.
Warsaw Summer School 2015, OSU Study Abroad Program Regression.
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
Revision: Pivot Table 1. Histogram 2. Trends 3. Linear 4. Exponential
Chapter 6 (cont.) Difference Estimation. Recall the Regression Estimation Procedure 2.
Correlation and Regression: The Need to Knows Correlation is a statistical technique: tells you if scores on variable X are related to scores on variable.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
LESSON 6: REGRESSION 2/21/12 EDUC 502: Introduction to Statistics.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
STATISTICS 12.0 Correlation and Linear Regression “Correlation and Linear Regression -”Causal Forecasting Method.
Regression Analysis Deterministic model No chance of an error in calculating y for a given x Probabilistic model chance of an error First order linear.
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
OLS Regression What is it? Closely allied with correlation – interested in the strength of the linear relationship between two variables One variable is.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Correlation and Linear Regression
Theme 6. Linear regression
Regression and Correlation
Regression Analysis AGEC 784.
Correlation & Regression
Practice. Practice Practice Practice Practice r = X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4 (4) 72.
Reasoning in Psychology Using Statistics
The Lease Squares Line Finite 1.3.
Kin 304 Regression Linear Regression Least Sum of Squares
Ordinary Least Squares (OLS) Regression
Chapter 5 STATISTICS (PART 4).
BPK 304W Regression Linear Regression Least Sum of Squares
Correlation and Simple Linear Regression
Least-Squares Regression
Simple Linear Regression
Correlation and Simple Linear Regression
The Pearson Correlation
Introduction to Regression
Lesson 2.2 Linear Regression.
Simple Linear Regression and Correlation
Simple Linear Regression
R = R Squared
Introduction to Regression
Lesson 2.2 Linear Regression.
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Regression & Prediction 4/15/2019 Regression & Prediction

4/15/2019 Linear Regression Finding the best fitting straight line for a set of data This line is represented by the equation Y = bX + a, where a & b are fixed constants

Least Squares Regression 4/15/2019 Least Squares Regression Most commonly used regression line Makes the sum of the squared errors as small as possible Regression Line Equation Y=bX+a Y(hat) is the predicted value of Y given a certain X b is the slope a is the y-intercept

4/15/2019 Slope (b) Indicates by how much Y will change when X is increased by one point

Relationship of “b” to “r” 4/15/2019 Relationship of “b” to “r” When “r” is positive, “b” is positive When “r” is zero, “b” is zero When “r” is negative, “b” is negative

4/15/2019 Y-Intercept (a) a = MY - bMX Value of Y when X is equal to 0

How Good is Our Prediction - Standard Error of Estimate 4/15/2019 How Good is Our Prediction - Standard Error of Estimate Provides a measure of how accurately the regression equation predicts the Y values Gives a measure of the standard distance between a regression line and the actual data points Analogous to standard deviation.

Standard Error of Estimate - Formula 4/15/2019 Standard Error of Estimate - Formula

Another Formula for Standard Error of Estimate 4/15/2019 Another Formula for Standard Error of Estimate SSerror=(1-r2)SSY By using this formula for SSerror the standard error of estimate can also be computed as

4/15/2019 Example Exam 1 (X) Final Grades (Y) 62 74 73 93 88 68 82 79 85 91 77 72 94 96 65 61 92 98 95 A professor claims that the scores on the 1st exam provide an excellent indication of how students will perform throughout the term

Exam 1 (x) Final Grades (Y) X2 Y2 XY 62 74 3844 5476 4588 73 93 5329 8649 6789 88 68 7744 4624 5984 82 79 6724 6241 6478 85 91 7225 8281 7735 77 72 5929 5184 5544 94 96 8836 9216 9024 65 61 4225 3721 3965 92 8464 8372 6068 7905 98 95 9604 9025 9310 4/15/2019

4/15/2019 Calculate SP

4/15/2019 Calculate SSx And the SSx

4/15/2019 Calculate SSY And the SSy

4/15/2019 Calculate r

Calculate Regression Equation 4/15/2019 Calculate Regression Equation Y’ = bX+a a = MY - bMX

Standard Error of Estimate 4/15/2019 Exam 1 (x) Final Grades (Y) Predicted Final (Y-Y’) (Y-Y')2 62 74 70.35 3.65 13.3225 73 93 77.61 -15.39 236.8521 88 68 87.51 -19.51 380.6401 82 79 83.55 -4.55 20.7025 85 91 85.53 5.47 29.9209 77 72 80.25 -8.25 68.0625 94 96 91.47 4.53 20.5209 65 61 72.33 -11.33 128.3689 92 89.49 2.51 6.3001 78.27 3.73 13.9129 7.47 55.8009 98 95 94.11 0.89 0.7921 We know now that our predictions for y based on x will not be perfect, however we can calculate approximately how far off our predictions will be. That is, we need is is called the standard error of the estimate. Our regression equation is: Y’ = 0.66X+29.43 We can do this by finding the predicted value for each X value and then subtracting Y-hat from Y and squaring the deviation, and then dividing it by its degree of freedom to obtain a measure of variance and taking the square root

Standard Error of Estimate

Multiple Regression: Some Additional Points 4/15/2019 Multiple Regression: Some Additional Points If b1>b2, does that mean that X1 is a better predictor than X2?