Sequential sums of squares … or … extra sums of squares.

Slides:



Advertisements
Similar presentations
Qualitative predictor variables
Advertisements

Multicollinearity.
More on understanding variance inflation factors (VIFk)
BA 275 Quantitative Business Methods
Topic 15: General Linear Tests and Extra Sum of Squares.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
Note 14 of 5E Statistics with Economics and Business Applications Chapter 12 Multiple Regression Analysis A brief exposition.
Statistics 350 Lecture 23. Today Today: Exam next day Good Chapter 7 questions: 7.1, 7.2, 7.3, 7.28, 7.29.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Model selection Stepwise regression. Statement of problem A common problem is that there is a large set of candidate predictor variables. (Note: The examples.
Hypothesis tests for slopes in multiple linear regression model Using the general linear test and sequential sums of squares.
A (second-order) multiple regression model with interaction terms.
Simple linear regression Linear regression with one predictor variable.
Linear Lack of Fit (LOF) Test An F test for checking whether a linear regression function is inadequate in describing the trend in the data.
M23- Residuals & Minitab 1  Department of ISM, University of Alabama, ResidualsResiduals A continuation of regression analysis.
Variable selection and model building Part II. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
Econ 3790: Business and Economics Statistics
1 1 Slide © 2003 Thomson/South-Western Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple Coefficient of Determination.
Econ 3790: Business and Economics Statistics Instructor: Yogesh Uppal
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Chapter 11 Linear Regression Straight Lines, Least-Squares and More Chapter 11A Can you pick out the straight lines and find the least-square?
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
An alternative approach to testing for a linear association The Analysis of Variance (ANOVA) Table.
Chap 14-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics.
Copyright ©2011 Nelson Education Limited Linear Regression and Correlation CHAPTER 12.
Solutions to Tutorial 5 Problems Source Sum of Squares df Mean Square F-test Regression Residual Total ANOVA Table Variable.
Chapter 13 Multiple Regression
Lack of Fit (LOF) Test A formal F test for checking whether a specific type of regression function adequately fits the data.
Multiple regression. Example: Brain and body size predictive of intelligence? Sample of n = 38 college students Response (Y): intelligence based on the.
Inference with computer printouts. Coefficie nts Standard Errort StatP-value Lower 95% Upper 95% Intercept
Overview of our study of the multiple linear regression model Regression models with more than one slope parameter.
The general linear test approach to regression analysis.
Regression. Height Weight How much would an adult female weigh if she were 5 feet tall? She could weigh varying amounts – in other words, there is a distribution.
EXTRA SUMS OF SQUARES  Extra sum squares  Decomposition of SSR  Usage of Extra sum of Squares  Coefficient of partial determination.
Regression through the origin
Chapter 12 Simple Linear Regression.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Variable selection and model building Part I. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Multicollinearity. Multicollinearity (or intercorrelation) exists when at least some of the predictor variables are correlated among themselves. In observational.
Chapter 9 Minitab Recipe Cards. Contingency tests Enter the data from Example 9.1 in C1, C2 and C3.
Faculty of Arts Atkinson College MATH 3330 M W 2003 Welcome Seventh Lecture for MATH 3330 M Professor G.E. Denzel.
Interaction regression models. What is an additive model? A regression model with p-1 predictor variables contains additive effects if the response function.
Chapter 26: Inference for Slope. Height Weight How much would an adult female weigh if she were 5 feet tall? She could weigh varying amounts – in other.
Simple linear regression. What is simple linear regression? A way of evaluating the relationship between two continuous variables. One variable is regarded.
Simple linear regression. What is simple linear regression? A way of evaluating the relationship between two continuous variables. One variable is regarded.
Analysis of variance approach to regression analysis … an (alternative) approach to testing for a linear association.
Model selection and model building. Model selection Selection of predictor variables.
Announcements There’s an in class exam one week from today (4/30). It will not include ANOVA or regression. On Thursday, I will list covered material and.
Multiple Regression.
Chapter 20 Linear and Multiple Regression
Least Square Regression
Least Square Regression
John Loucks St. Edward’s University . SLIDES . BY.
Simple Linear Regression
Regression model with multiple predictors
Relationship with one independent variable
9/19/2018 ST3131, Lecture 6.
Multiple Regression II
Solutions for Tutorial 3
Multiple Regression.
Multiple Regression II
Chapter 7 Minitab Recipe Cards
Multiple Regression Chapter 14.
Relationship with one independent variable
SIMPLE LINEAR REGRESSION
Chapter Fourteen McGraw-Hill/Irwin
Essentials of Statistics for Business and Economics (8e)
Introduction to Regression
Presentation transcript:

Sequential sums of squares … or … extra sums of squares

Sequential sums of squares: what are they? The reduction in the error sum of squares when one or more predictor variables are added to the regression model. Or, the increase in the regression sum of squares when one or more predictor variables are added to the regression model.

Sequential sums of squares: why? They can be used to test whether one slope parameter is 0. They can be used to test whether a subset (more than two, but less than all) of the slope parameters are 0.

Example: Brain and body size predictive of intelligence? Sample of n = 38 college students Response (Y): intelligence based on the PIQ (performance) scores from the (revised) Wechsler Adult Intelligence Scale. Predictor (X 1 ): Brain size based on MRI scans (given as count/10,000) Predictor (X 2 ): Height in inches Predictor (X 3 ): Weight in pounds

OUTPUT #1 The regression equation is PIQ = MRI Predictor Coef SE Coef T P Constant MRI Analysis of Variance Source DF SS MS F P Regression Error Total

OUTPUT #2 The regression equation is PIQ = MRI Height Predictor Coef SE Coef T P Constant MRI Height Analysis of Variance Source DF SS MS F P Regression Residual Total Source DF Seq SS MRI Height

OUTPUT #3 The regression equation is PIQ = MRI Height Weight Predictor Coef SE Coef T P Constant MRI Height Weight Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS MRI Height Weight 1 0.0

Sequential sums of squares: definition using SSE notation SSR(X 2 |X 1 ) = SSE(X 1 ) - SSE(X 1,X 2 ) In general, you subtract the error sum of squares due to all of the predictors both left and right of the bar from the error sum of squares due to the predictor to the right of the bar. SSR(X 2,X 3 |X 1 ) = SSE(X 1 ) - SSE(X 1,X 2,X 3 )

Sequential sums of squares: definition using SSR notation SSR(X 2 |X 1 ) = SSR(X 1,X 2 ) – SSR(X 1 ) In general, you subtract the regression sum of squares due to the predictor to the right of the bar from the regression sum of squares due to all of the predictors both left and right of the bar. SSR(X 2,X 3 |X 1 ) = SSR(X 1,X 2,X 3 )-SSR(X 1 )

Decomposition of regression sum of squares In multiple regression, there is more than one way to decompose the regression sum of squares. For example:

OUTPUT #2 The regression equation is PIQ = MRI Height Predictor Coef SE Coef T P Constant MRI Height Analysis of Variance Source DF SS MS F P Regression Residual Total Source DF Seq SS MRI Height

OUTPUT #4 The regression equation is PIQ = Height MRI Predictor Coef SE Coef T P Constant Height MRI Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS Height MRI

Decomposition of SSR: how?

Even more ways to decompose SSR when 3 or more predictors

Degrees of freedom and regression mean squares A sequential sum of squares involving one extra predictor variable has one degree of freedom associated with it: A sequential sum of squares involving two extra predictor variables has two degrees of freedom associated with it:

Sequential sums of squares in Minitab The SSR is automatically decomposed into one-degree-of-freedom sequential sums of squares, in the order in which the predictor variables are entered into the model. To get sequential sum of squares involving two or more predictor variables, sum the appropriate one-degree-of-freedom sequential sums of squares.

OUTPUT #3 The regression equation is PIQ = MRI Height Weight Predictor Coef SE Coef T P Constant MRI Height Weight Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS MRI Height Weight 1 0.0

OUTPUT #5 The regression equation is PIQ = Height Weight MRI Predictor Coef SE Coef T P Constant Height Weight MRI Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS Height Weight MRI

Testing one slope β 1 = β MRI is 0 Predictor Coef SE Coef T P Constant Height Weight MRI Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS Height Weight MRI

Testing one slope β 2 = β HT is 0 Predictor Coef SE Coef T P Constant MRI Weight Height Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS MRI Weight Height

Testing one slope β 3 = β WT is 0 Predictor Coef SE Coef T P Constant MRI Height Weight Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS MRI Height Weight 1 0.0

Testing one slope β k is 0: why it works? Full model:Reduced model:

Testing one slope β k is 0: why it works? (cont’d) The general linear test statistic: becomes:

Testing whether β 2 = β 3 = 0 Full model:Reduced model:

Testing whether β 2 = β 3 = 0 (cont’d) The general linear test statistic: becomes:

OUTPUT #3 The regression equation is PIQ = MRI Height Weight Predictor Coef SE Coef T P Constant MRI Height Weight Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS MRI Height Weight 1 0.0

Cumulative Distribution Function F distribution with 2 DF in numerator and 34 DF in denominator x P( X <= x ) P-value is:

Getting P-value for F-statistic in Minitab Select Calc >> Probability Distributions >> F… Select Cumulative Probability. Use default noncentrality parameter of 0. Type in numerator DF and denominator DF. Select Input constant. Type in F-statistic. Answer appears in session window. P-value is 1 minus the number that appears.

Test whether β 1 = β 3 = 0 Analysis of Variance Source DF SS MS F P Regression Error Total Source DF Seq SS Height Weight MRI