Multiple Regression. Outline Purpose and logic : page 3 Purpose and logic : page 3 Parameters estimation : page 9 Parameters estimation : page 9 R-square.

Slides:



Advertisements
Similar presentations
Analysis of Variance (ANOVA). Hypothesis H 0 :  i =  G H 1 :  i | (  i   G ) Logic S 2 within = error variability S 2 between = error variability.
Advertisements

Hypothesis Testing Steps in Hypothesis Testing:
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Chapter 10 Curve Fitting and Regression Analysis
Linear regression models
Ch11 Curve Fitting Dr. Deshi Ye
Introduction to Regression Analysis
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
Multiple regression analysis
Statistics 350 Lecture 16. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
Linear Regression and Correlation
Correlation Patterns. Correlation Coefficient A statistical measure of the covariation or association between two variables. Are dollar sales.
Chapter 11 Multiple Regression.
Quantitative Business Analysis for Decision Making Simple Linear Regression.
Ch. 14: The Multiple Regression Model building
Analysis of Variance & Multivariate Analysis of Variance
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
Simple Linear Regression and Correlation
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression Analysis
Relationships Among Variables
Correlation & Regression
Example of Simple and Multiple Regression
Introduction to Linear Regression and Correlation Analysis
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
Chapter 13: Inference in Regression
Correlation and Linear Regression
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
Anthony Greene1 Correlation The Association Between Variables.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Warsaw Summer School 2015, OSU Study Abroad Program Regression.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1Spring 02 First Derivatives x y x y x y dy/dx = 0 dy/dx > 0dy/dx < 0.
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Multiple Linear Regression Partial Regression Coefficients.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
VI. Regression Analysis A. Simple Linear Regression 1. Scatter Plots Regression analysis is best taught via an example. Pencil lead is a ceramic material.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Chapter Thirteen Copyright © 2006 John Wiley & Sons, Inc. Bivariate Correlation and Regression.
Environmental Modeling Basic Testing Methods - Statistics III.
–The shortest distance is the one that crosses at 90° the vector u Statistical Inference on correlation and regression.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin Linear Regression and Correlation Chapter 13.
Simple linear regression and correlation Regression analysis is the process of constructing a mathematical model or function that can be used to predict.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
REGRESSION G&W p
Chapter 4. Inference about Process Quality
Review of Chapter 3 where Multiple Linear Regression Model:
LESSON 24: INFERENCES USING REGRESSION
Simple Linear Regression
Model Comparison: some basic concepts
Statistical Inference about Regression
Correlation and Regression
Simple Linear Regression
Simple Linear Regression
3.2. SIMPLE LINEAR REGRESSION
Statistical Inference for the Mean: t-test
Presentation transcript:

Multiple Regression

Outline Purpose and logic : page 3 Purpose and logic : page 3 Parameters estimation : page 9 Parameters estimation : page 9 R-square : page 13 R-square : page 13 Hypothesis testing : page 17 Hypothesis testing : page 17 Partial and semi-partial regression : page 21 Partial and semi-partial regression : page 21 Confidence intervals : page 26 Confidence intervals : page 26

Multiple Regression There are more than one predictor There are more than one predictor Example with 2 predictors Example with 2 predictors

Multiple Regression Because we have two predictors, it is possible to illustrate the situation using a 3-dimensional scatter plot Because we have two predictors, it is possible to illustrate the situation using a 3-dimensional scatter plot

Multiple Regression The different relations can be illustrated by a bivariate scatter plots. The different relations can be illustrated by a bivariate scatter plots. x1x1 x1x1 x2x2 y x2x2 y

Multiple Regression As well as the bivariate correlations (see SSCP). As well as the bivariate correlations (see SSCP).

Multiple Regression From the regression line to the regression hyperplane From the regression line to the regression hyperplane

Multiple Regression Graphics Graphics It is not possible to illustrates vectors in 5 dimensions. However, the computations will be the same.

Multiple Regression How the regression coefficients can be obtained ? Universal formula, it does not change whatever the number of predictors (it is a special case of multivariate regression).

Multiple Regression The b 0 coefficient can also be directly obtained if we include the unity vector 1 as a variable.

Multiple Regression How the regression coefficients can be obtained ?

Standardized Regression Coefficients It allow to measure the « importance » of the predictors, since they all have a variability of 1 and a mean of 0. Therefore, an increase of one unit by z 1, will increase 0.74 standard deviation by y Z. ^ Or

R-Square Like in simple regression, in multiple regression we will use the R-square coefficient (R 2 ); also named coefficient of determination. This R 2 have the same interpretation as the one in simple regression: percentage of explained variance given by all the predictors. Sum of Squares and Cross Product Matrix (SSCP)

R-Square Recall: By dividing the SSCP matrix by the corresponding degrees of freedom, we get the variance-covariance matrix. We can also get the bivariate correlations

R-Square In addition, the SSCP matrix can be partitioned in function of the independent (predictors) and dependent variables (criterion). Scp Spp Spc Scc

R-Square R 2 is obtained by: R 2 adj is an unbiased estimation of the population variability given a sample Scp Spp Spc Scc Number of participantsNumber of predictors (independent variables)

Hypothesis testing The hypothesis is that the R-square between the predictors and the criterion is null. In other words, we are trying to know if the X and y variables are linearly independent. If we reject that hypothesis, then the two populations are not independent, there is a linear relation between the two.

Hypothesis testing Using confidence intervals. Using observed statistics.

Hypothesis testing The hypothesis is that the R-Square between the predictors and the criterion is null. In other words, we are trying to know if the X and y variables are linearly independent. If we reject that hypothesis, then the two populations are not independent, there is a linear relation between the two. Because the F obs >F crit ( >19.00), we reject H 0 and we therefore accept H 1. The 2 populations are linearly dependant. Explained variabilityUnexplained variability

ANOVA Table The hypothesis is that the R-Square between the predictors and the criterion is null. Because F(2,2)= , p.<0.05, we reject H 0 and we accept H 1. The 2 populations are linearly dependant. =

Partial and semi-partial correlations The idea is to put forward the effect of one predictor on the dependant variable by controlling the effects of the other predictors. Squared semi partial correlation coefficient It is the total variability (R 2 ) minus the total variability excluding the studied predictor from the data set. It is the portion of variability that is unique to this particular predictor Squared partial correlation coefficient It is the variability proportion associated with one predictor but not with the others. In other words, it is the unestimated variability by the other predictors that is estimated by the predictor studied.

Partial and semi-partial correlations a b c e Y x1x1 x2x2

Example 39 % of the y variability is explained solely by the first predictor. 9% of the y variability is explained solely by the second predictor. 90% of the unexplained y variability by the second predictor is explained by the first predictor. 67% of the unexplained y variability by the first predictor is explained by the second predictor. x1x1 x2x2

Partial and semi-partial correlations Signification testing x1x1 x2x2 The various parameters (pr i, b i, B i ) are directly related from the portion of explained variability by the semi partial coefficient. Therefore, if this last coefficient is statistically significant, then all the other parameters will be.

Standard errors associated with the regression parameters Standard error associated with the regression coefficients Standard errors associated with the standardized regression coefficients

Confidance intervals associated with the regression parameters Confidence intervals associated with the regression coefficients Confidence intervals associated with the standardized regression coefficients