Regression 10/29.

Slides:



Advertisements
Similar presentations
Test of (µ 1 – µ 2 ),  1 =  2, Populations Normal Test Statistic and df = n 1 + n 2 – 2 2– )1– 2 ( 2 1 )1– 1 ( 2 where ] 2 – 1 [–
Advertisements

Managerial Economics in a Global Economy
Bivariate Analyses.
Regression What is regression to the mean?
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
PSY 307 – Statistics for the Behavioral Sciences
Intro to Statistics for the Behavioral Sciences PSYC 1900
Ch. 14: The Multiple Regression Model building
Chapter 9 - Lecture 2 Computing the analysis of variance for simple experiments (single factor, unrelated groups experiments).
Basic Statistical Concepts Part II Psych 231: Research Methods in Psychology.
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
Relationships Among Variables
Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.
Correlation and Regression
Quantitative Business Analysis for Decision Making Multiple Linear RegressionAnalysis.
Example of Simple and Multiple Regression
Lecture 16 Correlation and Coefficient of Correlation
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, November 19 Chi-Squared Test of Independence.
The Chi-Square Distribution 1. The student will be able to  Perform a Goodness of Fit hypothesis test  Perform a Test of Independence hypothesis test.
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Chapter 15 Correlation and Regression
Simple Linear Regression One reason for assessing correlation is to identify a variable that could be used to predict another variable If that is your.
1 1 Slide Simple Linear Regression Coefficient of Determination Chapter 14 BA 303 – Spring 2011.
POSC 202A: Lecture 12/10 Announcements: “Lab” Tomorrow; Final ed out tomorrow or Friday. I will make it due Wed, 5pm. Aren’t I tender? Lecture: Substantive.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
Production Planning and Control. A correlation is a relationship between two variables. The data can be represented by the ordered pairs (x, y) where.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Environmental Modeling Basic Testing Methods - Statistics III.
Essentials of Business Statistics: Communicating with Numbers By Sanjiv Jaggia and Alison Kelly Copyright © 2014 by McGraw-Hill Higher Education. All rights.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Analysis of Variance 11/6. Comparing Several Groups Do the group means differ? Naive approach – Independent-samples t-tests of all pairs – Each test doesn't.
Chapter 13 Understanding research results: statistical inference.
Stats Methods at IC Lecture 3: Regression.
Multiple Regression.
Chapter 14 Introduction to Multiple Regression
Chapter 20 Linear and Multiple Regression
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
REGRESSION G&W p
Reasoning in Psychology Using Statistics
Review Guess the correlation
Regression 11/6.
Regression.
Basics of Group Analysis
Simple Linear Regression
ANOVA 11/3.
Analysis of Variance.
Regression model with multiple predictors
POSC 202A: Lecture Lecture: Substantive Significance, Relationship between Variables 1.
BPK 304W Correlation.
Correlation and Simple Linear Regression
Multiple Regression – Part II
CHAPTER 29: Multiple Regression*
Multiple Regression.
Correlation and Simple Linear Regression
Statistics for the Social Sciences
Introduction to Regression
Reasoning in Psychology Using Statistics
Simple Linear Regression and Correlation
Reasoning in Psychology Using Statistics
Reasoning in Psychology Using Statistics
Regression & Correlation (1)
Introduction to Regression
Introduction to Regression
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Regression 10/29

Prediction Correlation can tell how to predict one variable from another What about multiple predictor variables? Explaining Y using X1, X2, … Xm Income based on age, education Memory based on pre-test, IQ Social behavior based on personality dimensions Regression Finds best combination of predictors to explain outcome variable Determines unique contribution of each predictor Can test whether each predictor has reliable influence

Linear Prediction One predictor Multiple predictors Intercept: b0 Draw a line through data Multiple predictors Each predictor has linear effect Effects of different predictors simply add together Intercept: b0 Value of Y when all Xs are zero Regression coefficients: bi Influence of Xi Sign tells direction; magnitude tells strength Can be any value (not standardized like correlation)

Example Predict income from education and intelligence Y = $k/yr X1 = years past HS X2 = IQ Regression equation: 4-year college, IQ = 120 -70 + 10*4 + 1*120 = 90  $90k/yr 2-year college, IQ = 90 -70 + 10*2 + 1*90 = 40  $40k/yr

Finding Regression Coefficients Goal: regression equation with best fit to data Minimize error between Y and Math solution Matrix algebra Treats X as a matrix Rows for subjects, columns for variables Transposes, matrix inverse, and other fun Practical solution Use a computer R: linear model function: lm(Y~X1+X2+X3)

Regression vs. Correlation One predictor: regression = correlation Multiple predictors Predictors affect each other Each coefficient depends on what other predictors are included Example: 3rd-variable problem Math achievement and verbal achievement are correlated Math helps with English? Add IQ as a predictor: verbal ~ math + IQ bmath will be near 0 No effect of math once IQ is accounted for Regression finds best combination of predictors to explain outcome Best values of b1, b2, etc. taken together Each coefficient shows contribution of predictor beyond effects of others Allows identification of most important predictors

Explained Variability Sum of squares Same as mean squared error, but without dividing by df Easier because not complicated by degrees of freedom Total sum of squares Variability in Y SSY = S(Y – MY)2 Residual sum of squares Deviation between predicted and actual scores SSresidual = S(Y – )2

Explained Variability SSY SSresidual SSregression SSregression Variability explained by regression Difference between total and residual sums of squares SSregression = SSY - SSresidual R2 ("R-squared") Fraction of variability explained by the regression SSregression/SSY Measures how well the predictors (Xs) can predict or explain the outcome (Y) Same as r2 from correlation when only one predictor

Hypothesis Testing with Regression Does the regression explain meaningful variance in the outcome? Null hypothesis All regression coefficients equal zero in population Regression only “explaining” random error in sample Approach Find how much variance the predictors explain: SSregression Compare to amount expected by chance Ratio gives F statistic, which we can use to get p-value If F is larger than expected, regression explains more variance than expected by chance Reject null hypothesis if F is large enough

Hypothesis Testing with Regression SSY SSresidual SSregression Likelihood function for F is an F distribution Comes from ratio of two chi-square variables Two df values, from numerator and denominator p-value is probability of a result greater than F

Degrees of Freedom SSY dfY = n - 1 dfresidual = n – m – 1 SSregression dfregression = m SSY dfY = n - 1 dfresidual = n – m – 1

Testing Individual Predictors Does predictor variable have any effect on outcome? Does it provide any information beyond other predictors? Null hypothesis: bi = 0 in population Other bs might be nonzero Compute standard error for bi Uses SSresidual (MSE) to estimate t statistic: Get p-value from t distribution with dfresidual Can do one-tailed test if predicted direction of effect Sign of bi