Prediction/Regression

Slides:



Advertisements
Similar presentations
13- 1 Chapter Thirteen McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc., All Rights Reserved.
Advertisements

 Coefficient of Determination Section 4.3 Alan Craig
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
© McGraw-Hill Higher Education. All Rights Reserved. Chapter 2F Statistical Tools in Evaluation.
Simple Linear Regression 1. Correlation indicates the magnitude and direction of the linear relationship between two variables. Linear Regression: variable.
Cal State Northridge  320 Andrew Ainsworth PhD Regression.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Statistics for the Social Sciences
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Correlation and Regression Analysis
Ch 11: Correlations (pt. 2) and Ch 12: Regression (pt.1) Apr. 15, 2008.
Reminders  HW2 due today  Exam 1 next Tues (9/27) – Ch 1-5 –3 sections: Short answers (concepts, definitions) Calculations (you’ll be given the formulas)
Ch 11: Correlations (pt. 2) and Ch 12: Regression (pt.1) Nov. 13, 2014.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
Lecture 5: Simple Linear Regression
PSY 307 – Statistics for the Behavioral Sciences Chapter 7 – Regression.
Review Regression and Pearson’s R SPSS Demo
Relationships Among Variables
Example of Simple and Multiple Regression
Aron, Aron, & Coups, Statistics for the Behavioral and Social Sciences: A Brief Course (3e), © 2005 Prentice Hall Chapter 3 Correlation and Prediction.
Introduction to Linear Regression and Correlation Analysis
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
Chapter 6 & 7 Linear Regression & Correlation
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
Completing the ANOVA From the Summary Statistics.
Simple Linear Regression One reason for assessing correlation is to identify a variable that could be used to predict another variable If that is your.
Section 4.2 Least Squares Regression. Finding Linear Equation that Relates x and y values together Based on Two Points (Algebra) 1.Pick two data points.
© 2014 by Pearson Higher Education, Inc Upper Saddle River, New Jersey All Rights Reserved HLTH 300 Biostatistics for Public Health Practice, Raul.
Chapter 11 Correlation Pt 1: Nov. 12, Correlation Association between scores on two variables –e.g., age and coordination skills in children, price.
Soc 3306a Lecture 9: Multivariate 2 More on Multiple Regression: Building a Model and Interpreting Coefficients.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.2 Extending the Correlation and R-Squared for Multiple.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Part IV Significantly Different Using Inferential Statistics Chapter 15 Using Linear Regression Predicting Who’ll Win the Super Bowl.
Part IV Significantly Different: Using Inferential Statistics
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 4 Section 2 – Slide 1 of 20 Chapter 4 Section 2 Least-Squares Regression.
Copyright ©2011 Nelson Education Limited Linear Regression and Correlation CHAPTER 12.
Regression Lesson 11. The General Linear Model n Relationship b/n predictor & outcome variables form straight line l Correlation, regression, t-tests,
Chapter 4 Prediction. Predictor and Criterion Variables  Predictor variable (X)  Criterion variable (Y)
Welcome to MM570 Psychological Statistics Unit 4 Seminar Dr. Srabasti Dutta.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
11/23/2015Slide 1 Using a combination of tables and plots from SPSS plus spreadsheets from Excel, we will show the linkage between correlation and linear.
Chapter 3 Correlation.  Association between scores on two variables –e.g., age and coordination skills in children, price and quality.
Correlation MEASURING ASSOCIATION Establishing a degree of association between two or more variables gets at the central objective of the scientific enterprise.
Chapter 4 Summary Scatter diagrams of data pairs (x, y) are useful in helping us determine visually if there is any relation between x and y values and,
Midterm Review Ch 7-8. Requests for Help by Chapter.
Scatter Diagrams scatter plot scatter diagram A scatter plot is a graph that may be used to represent the relationship between two variables. Also referred.
CHAPTER 5 CORRELATION & LINEAR REGRESSION. GOAL : Understand and interpret the terms dependent variable and independent variable. Draw a scatter diagram.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Statistics for Psychology CHAPTER SIXTH EDITION Statistics for Psychology, Sixth Edition Arthur Aron | Elliot J. Coups | Elaine N. Aron Copyright © 2013.
Regression. Outline of Today’s Discussion 1.Coefficient of Determination 2.Regression Analysis: Introduction 3.Regression Analysis: SPSS 4.Regression.
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
Welcome to MM570 Psychological Statistics Unit 4 Seminar Dr. Bob Lockwood.
© The McGraw-Hill Companies, Inc., Chapter 10 Correlation and Regression.
Topics, Summer 2008 Day 1. Introduction Day 2. Samples and populations Day 3. Evaluating relationships Scatterplots and correlation Day 4. Regression and.
Prediction/Regression
Chapter 13 Multiple Regression
Reasoning in Psychology Using Statistics
Multiple Regression.
Chapter 15 Linear Regression
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
Example 1 5. Use SPSS output ANOVAb Model Sum of Squares df
Simple Linear Regression
Prediction/Regression
Prediction/Regression
Prediction/Regression
Introduction to Regression
Presentation transcript:

Prediction/Regression Chapter 12 Prediction/Regression Part 2: Apr. 17, 2008

Drawing the Regression Line Draw and label the axes for a scatter diagram Figure predicted value on criterion for a low value on predictor variable You can randomly choose what value to plug in.. Maybe x=1, so Ŷ = .688 + .92(1) = 1.61 Repeat step 2. with a high value on predictor, maybe x=6, so Ŷ = .688 + .92(6) = 6.21 Draw a line passing through the two marks (1, 1.61) and (6, 6.21) Hint: you can also use (Mx, My) to save time as one of your 2 points. Reg line always passes through the means of x and y.

Regression Error Now that you have a regression line or equation built from one sample, you can find predicted y scores using a new sample of x scores (Sample 2)… Then, assume that you later collect data on Sample 2’s actual y scores You can compare the accuracy of predicted ŷ to the actual y scores for Sample 2 Sometimes you’ll overestimate, sometimes underestimate…this is ERROR. Can we get a measure of error? How much is OK?

Error in regression Proportionate Reduction in Error (PRE) Actual score minus the predicted score Proportionate Reduction in Error (PRE) Squared error using prediction (reg) model = SSError =  (y - ŷ)2 Compare this to amount of error w/o this prediction (reg) model. If no other model, best guess would be the mean. Total squared error when predicting

Error and Proportionate Reduction in Error Formula for proportionate reduction in error: compares reg model to mean baseline (predicting everyone’s y score will be at the mean) Want reg model to be much better than mean(baseline) – that would indicate fewer prediction errors So you want PRE to be large…

Find prediction using reg model: Reg model was ŷ = .688 + .92(x) Use mean model to find error (y-My)2 for each person & sum up that column  SStot Find prediction using reg model: plug in x values into reg model to get ŷ Find (y-ŷ)2 for each person, sum up that column  SSerror Find PRE

Proportionate reduction in error = r2 If our reg model no better than mean, SSerror = SStotal, so (0/ SStot) = 0. Using this regression model, we reduce error over the mean model by 0%….not good prediction. If reg model has 0 error (perfect), SStot-0/SStot = 1, or 100% reduction of error. Proportionate reduction in error = r2 aka “Proportion of variance in y accounted for by x”, ranges between 0-100%.

Computing Error Sum of the squared residuals = SSerror X Y 6 6 1 2 5 6 6 6 .688 + .92(6) = 6.2 1 2 .688 + .92(1) = 1.6 5 6 .688 + .92(5) = 5.3 3 4 .688 + .92(3) = 3.45 3 2 .688 + .92(3) = 3.45 mean 3.6 4.0

Computing SSerror Sum of the squared residuals = SSerror X Y 6 6 6.2 6 6 6.2 -0.20 0.04 1 2 1.6 0.40 0.16 5 6 5.3 0.70 0.49 3 4 3.45 0.55 0.30 3 2 3.45 -1.45 2.10 mean 3.6 4.0 0.00 3.09 SSERROR

Computing SStotal = (y-My)2 X Y My y-My (y-My)2 6 6 4 2 4 1 2 4 -2 4 5 6 4 2 4 3 4 4 0 0 3 2 4 -2 4 mean 3.6 4.0 Σ=0 Σ=16 (SStotal)

PRE PRE = 16 - 3.09 16 = .807 We have 80.7% proportionate reduction in error from using our regression model as opposed to the mean baseline model So we’re doing much better using our regression as opposed to just predicting the mean for everyone…

SPSS Reg Example Analyze Regression  Linear Note that terms used in SPSS are “Independent Variable”…this is x (predictor) “Dependent Variable”…this is y (criterion) Class handout of output – what to look for: “Model Summary” section - shows R2 ANOVA section – 1st line gives ‘sig value’, if < .05  signif This tests the significance of the R2 (is the whole regression equation significant or not? If yes  it does predict y) Coefficients section – 1st line gives ‘constant’ = a Other lines give ‘standardized coefficients’ = b or beta for each predictor For each predictor, there is also a significance test (if ‘sig’ if < .05, that predictor is significantly different from 0 and does predict y) If it is significant, you’d want to interpret the beta (like a correlation)