Review Guess the correlation. -2.0 -0.9 -0.1 0.1 0.9.

Slides:



Advertisements
Similar presentations
Managerial Economics in a Global Economy
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
Bivariate Analyses.
Chapter 10 Curve Fitting and Regression Analysis
Regression What is regression to the mean?
Intro to Statistics for the Behavioral Sciences PSYC 1900
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Ch. 14: The Multiple Regression Model building
Chapter 9 - Lecture 2 Computing the analysis of variance for simple experiments (single factor, unrelated groups experiments).
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
Simple Linear Regression Analysis
Relationships Among Variables
Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.
Example of Simple and Multiple Regression
The Chi-Square Distribution 1. The student will be able to  Perform a Goodness of Fit hypothesis test  Perform a Test of Independence hypothesis test.
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Chapter 15 Correlation and Regression
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Simple Linear Regression One reason for assessing correlation is to identify a variable that could be used to predict another variable If that is your.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
Production Planning and Control. A correlation is a relationship between two variables. The data can be represented by the ordered pairs (x, y) where.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Chapter 13 Understanding research results: statistical inference.
Stats Methods at IC Lecture 3: Regression.
Multiple Regression.
Correlation.
Chapter 14 Introduction to Multiple Regression
Chapter 20 Linear and Multiple Regression
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
REGRESSION G&W p
Regression 11/6.
Regression 10/29.
Correlation 10/27.
Kin 304 Regression Linear Regression Least Sum of Squares
Inference for Regression
Regression.
Basics of Group Analysis
Simple Linear Regression
ANOVA 11/3.
Analysis of Variance.
Review. Review Statistics Needed Need to find the best place to draw the regression line on a scatter plot Need to quantify the cluster.
BPK 304W Regression Linear Regression Least Sum of Squares
POSC 202A: Lecture Lecture: Substantive Significance, Relationship between Variables 1.
BPK 304W Correlation.
Correlation and Simple Linear Regression
Multiple Regression – Part II
Reasoning in Psychology Using Statistics
Correlation and Regression
CHAPTER 29: Multiple Regression*
Multiple Regression.
Correlation and Simple Linear Regression
Statistics for the Social Sciences
I. Statistical Tests: Why do we use them? What do they involve?
Introduction to Regression
Reasoning in Psychology Using Statistics
Simple Linear Regression and Correlation
Reasoning in Psychology Using Statistics
Reasoning in Psychology Using Statistics
Reasoning in Psychology Using Statistics
Regression & Correlation (1)
Introduction to Regression
Introduction to Regression
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Review Guess the correlation. -2.0 -0.9 -0.1 0.1 0.9

Review Calculate the correlation between X and Y. zX = [ -1.0 -0.3 -0.2 -0.6 -1.6] zY = [ -0.6 -1.3 -0.6 -0.0 -1.3] -1.21 -.30 -.24 -.01

Review Which statement is true? If the correlation is zero, then the variables are independent If the correlation is nonzero, then the variables are independent If two variables are independent, then their correlation is zero If two variables have a nonlinear relationship, then their correlation is zero

Regression 11/5

Regression Correlation can tell how to predict one variable from another What about multiple predictor variables? Explaining Y using X1, X2, … Xm Income based on age, education Memory based on pre-test, IQ Social behavior based on personality dimensions Regression Finds best combination of predictors to explain outcome variable Determines unique contribution of each predictor Can test whether each predictor has reliable influence

Linear Prediction One predictor Multiple predictors Intercept: b0 Draw a line through data Multiple predictors Each predictor has linear effect Effects of different predictors simply add together Intercept: b0 Value of Y when all Xs are zero Regression coefficients: bi Influence of Xi Sign tells direction; magnitude tells strength Can be any value (not standardized like correlation)

Example Predict income from education and intelligence Y = $k/yr X1 = years past HS X2 = IQ Regression equation: 4-year college, IQ = 120 -70 + 10*4 + 1*120 = 90  $90k/yr 2-year college, IQ = 90 -70 + 10*2 + 1*90 = 40  $40k/yr

Finding Regression Coefficients Goal: regression equation with best fit to data Minimize squared error between Y and Math solution Matrix algebra Combines Xs into a matrix Rows for subjects, columns for variables Transposes, matrix inverse, derivatives and other fun Practical solution Use a computer R: linear model function, lm(Y~X1+X2+X3)

Regression vs. Correlation One predictor: regression = correlation Plus commitment about which variable predicts the other Multiple predictors Predictors affect each other Each coefficient depends on what other predictors are included Example: 3rd-variable problem Math achievement and verbal achievement both depend on overall IQ Math helps with English? Add IQ as a predictor: verbal ~ math + IQ bmath will be near 0 No effect of math once IQ is accounted for Regression finds best combination of predictors to explain outcome Best values of b1, b2, etc. taken together Each coefficient shows contribution of predictor beyond effects of others Allows identification of most important predictors

Explained Variability Total sum of squares Variability in Y SSY = S(Y – MY)2 Residual sum of squares Deviation between predicted and actual scores SSresidual = S(Y – )2

Explained Variability SSY SSresidual SSregression SSregression Variability explained by regression Difference between total and residual sums of squares SSregression = SSY - SSresidual R2 ("R-squared") Fraction of variability explained by the regression SSregression/SSY Measures how well the predictors (Xs) can predict or explain the outcome (Y) Same as r2 from correlation when only one predictor

Hypothesis Testing with Regression Does the regression explain meaningful variance in the outcome? Null hypothesis All regression coefficients equal zero in population Regression only “explaining” random error in sample Approach Find how much variance the predictors explain: SSregression Compare to amount expected by chance Ratio gives F statistic, which we can use for hypothesis testing If F is larger than expected by H0, regression explains more variance than expected by chance Reject null hypothesis if F is large enough

Hypothesis Testing with Regression SSY SSresidual SSregression Likelihood function for F is an F distribution Comes from ratio of two chi-square variables Two df values, from numerator and denominator P(F5,10) F a p Fcrit

Degrees of Freedom SSY dfY = n - 1 dfresidual = n – m – 1 SSregression dfregression = m SSY dfY = n - 1 dfresidual = n – m – 1

Testing Individual Predictors Does predictor variable have any effect on outcome? Does it provide any information beyond other predictors? Null hypothesis: bi = 0 in population Other bs might be nonzero Compute standard error for bi Uses MSresidual to estimate t statistic: Get tcrit or p-value from t distribution with dfresidual Can do one-tailed test if predicted direction of effect Sign of bi

Review Predicting income from education and IQ: A high-school graduate with IQ = 125 considers getting a 4-year bachelor’s degree. What is her expected increase in income? $4k $40k $95k $165k

Review A study investigating personality differences in speech rate measures 100 subjects on 5 personality dimensions. These dimensions are used to predict words per minute (WPM) in a timed speech. The total variability in the outcome variable is SSWPM = 12000. The residual variability after accounting for the personality predictors is SSresidual = 9000. What is R2 for the regression? 0.06 0.25 0.75 30 3000

Review A study investigating personality differences in speech rate measures 100 subjects on 5 personality dimensions. These dimensions are used to predict words per minute (WPM) in a timed speech. The total variability in the outcome variable is SSWPM = 12000. The residual variability after accounting for the personality predictors is SSresidual = 9000. Calculate the F statistic for testing whether personality predicts speech rate. (dfregression = 5, dfresidual = 94) 0.02 0.33 4.70 6.27