Assumptions of Multiple Regression 1. Form of Relationship: –linear vs nonlinear –Main effects vs interaction effects 2. All relevant variables present.

Slides:



Advertisements
Similar presentations
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Advertisements

Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Prediction, Correlation, and Lack of Fit in Regression (§11. 4, 11
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
Simple Linear Regression 1. Correlation indicates the magnitude and direction of the linear relationship between two variables. Linear Regression: variable.
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
Multiple Linear Regression Model
Statistics for the Social Sciences
Statistics for Managers Using Microsoft® Excel 5th Edition
BA 555 Practical Business Analysis
Additional Topics in Regression Analysis
Chapter Topics Types of Regression Models
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
Chapter 11 Multiple Regression.
Topic 3: Regression.
Regression Diagnostics Checking Assumptions and Data.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Correlation and Regression Analysis
Slide 1 Testing Multivariate Assumptions The multivariate statistical techniques which we will cover in this class require one or more the following assumptions.
Relationships Among Variables
Checking Regression Model Assumptions NBA 2013/14 Player Heights and Weights.
Quantitative Business Analysis for Decision Making Multiple Linear RegressionAnalysis.
Regression and Correlation Methods Judy Zhong Ph.D.
ANCOVA Lecture 9 Andrew Ainsworth. What is ANCOVA?
Inference for regression - Simple linear regression
Simple Linear Regression Models
Quantitative Methods Heteroskedasticity.
The Examination of Residuals. The residuals are defined as the n differences : where is an observation and is the corresponding fitted value obtained.
1 Statistical Distribution Fitting Dr. Jason Merrick.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Regression Analysis Week 8 DIAGNOSTIC AND REMEDIAL MEASURES Residuals The main purpose examining residuals Diagnostic for Residuals Test involving residuals.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
September 18-19, 2006 – Denver, Colorado Sponsored by the U.S. Department of Housing and Urban Development Conducting and interpreting multivariate analyses.
1 Regression Analysis The contents in this chapter are from Chapters of the textbook. The cntry15.sav data will be used. The data collected 15 countries’
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Model Building and Model Diagnostics Chapter 15.
Residual Analysis Purposes –Examine Functional Form (Linear vs. Non- Linear Model) –Evaluate Violations of Assumptions Graphical Analysis of Residuals.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
4 basic analytical tasks in statistics: 1)Comparing scores across groups  look for differences in means 2)Cross-tabulating categoric variables  look.
ANCOVA.
Maths Study Centre CB Open 11am – 5pm Semester Weekdays
Simple Statistical Designs One Dependent Variable.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
Chapter 11 REGRESSION Multiple Regression  Uses  Explanation  Prediction.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Chapter 15 Multiple Regression Model Building
BINARY LOGISTIC REGRESSION
Correlation and Simple Linear Regression
12 Inferential Analysis.
(Residuals and
Correlation and Simple Linear Regression
CHAPTER 29: Multiple Regression*
Correlation and Simple Linear Regression
Undergraduated Econometrics
12 Inferential Analysis.
Simple Linear Regression
Product moment correlation
Regression Assumptions
Regression Assumptions
Diagnostics and Remedial Measures
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Assumptions of Multiple Regression 1. Form of Relationship: –linear vs nonlinear –Main effects vs interaction effects 2. All relevant variables present –Perfect reliability of predictors 3. Homoscedasticity (homogeneity) of errors of prediction 4. -Independence of errors of prediction 5. Normality of errors of prediction (residuals)

REMEDIES 1 Linear vs. Nonlinear modeling: –Linearity means for our purposes that each term in a model is added to another, and that each term has only one parameter as part of it –Eg. y = b 1 x 1 + b 2 x 2 + b 0 + e is linear because each variable (x 1, x 2, (intercept), e ) has only one parameter with it: b 1, b 2, b 0, 1 –An alternative definition is that x 1 and all the other variables are actually best modeled by some transformation, such as x 2 1

FORM OF RELATIONSHIP Linearity- use lowess line to provide evidence

Relationship problems in MR Counseling Linearity: transform using power functions: –If X is curvilinear in relation to Y, change X to X 2, Y = b 1 X 2 + e –If there is a possible interaction effect of two predictors, X1 and X2, create a new variable equal to the product (chapter 7) –Transform Y using log(Y) or SQRT(Y) in SPSS: Compute

REMEDIES 1 x 2 1 is still linear in the sense of the first definition, but when we plot it we see a different prediction line: x21x21 x1x1 y y b linear b quadratic data y x x

Main Effects and Interactions Main effects: effect for one predictor consistent across all values of second predictor Standard regression model: y = b 1 x 1 + b 2 x 2 + b 0 Interaction: effect additional to main effects Defined as product of two variables: x 1* x 2 new predictor variable

2. The Missing Variable ALL models can change seriously if we are missing an important variable Z –b-weights will usually change if Z is correlated with the predictors already in the model –Standard error of estimation and standard errors of b-weights will be reduced if Z is related to Y and not the other predictors –All above will change depending on the combination of relationships

Relevant Variables Theory, theory, theory Test additional variables for change in R- square, change in b- and beta weights of remaining predictors, change in residual characteristics.

Omitted relevant variable Effect of omitted predictor on beta weights, R-square, path coefficients Added value plots- examine error of new predictor from other predictors graphed against errors of prediction of original predictors

Measurement error in IVs Validity is attenuated (correlation between predictor and outcome) if the predictor is measured unreliably Disattenuated correlation between predictor and dependent variable is estimated by dividing by the square root of the reliability of the predictor (and of the dependent variable if we want a construct estimate of the correlation)

Measurement error in IVs Structural Equation Modeling (SEM) correctly estimates standard errors of parameters under ML estimation, assumes large samples (at least 100) OLS does not correctly correct the disattenuated correlation standard error for the b-weight

Omitted relevant variable LLR Smoother Residua of Social Stressl Residual in Attitude to School A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A Error of Initial Prediction by Error of Social Stress Prediction Predictors: Sensation Seeking, Locus of Control, Self Reliance

Omitted relevant variable Predictors: Sensation Seeking, Locus of Control, Self Reliance Linear Regression Residual of Social Stress Residual of Attitude to School A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A Unstandardized Residual = * RES_2 R-Square = 0.01 Error of Initial Prediction by Error of Social Stress Prediction

Measurement Error Requires Structural Equation Model Approach to include measurement error in predictors –Requires independent estimate of reliability of each predictor appropriate for the sample T2 T1 e1 e2 T3 √1-Rel. T2 √1-Rel. T1 e3 β 31 β 32

MRA ASSUMPTIONS-3 Homoscedasticity- errors of the prediction hyperplane are assumed homogeneous across the hyperplane (think of normal distributions of errors sticking out above and below the hyperplane: y x1 x2 Predictor values for x 1 and x 2

3. Nonhomogeneity of Variance If variances are unequal due to known groupings or clustering, each group’s variance can be estimated separately in SEM to fit correctly a regression model If variances are changing linearly for a predictor (or set of predictors): –Weighted least squares can be used –Nonlinear modeling (Hierarchical linear model) of the variance can be conducted using SEM

Homoscedasticity of residuals: SPSS Regression: SAVE: Unstandardized Predicted Values, Unstandardized Residuals SPSS Graph: Interactive: Scatterplot: Fit: Smoother LLR Smoother Unstandardized Predicted Value Unstandardized Residual A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A AA A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A Error of Initial Prediction by Error of Social Stress Prediction Linear Regression Unstandardized Predicted Value Unstandardized Residual A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A AA A A A A AA A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A AA A A A A A A A = * PRE_1 R-Square = 0.00 Error of Initial Prediction by Error of Social Stress Prediction

Error bars suggest increasing variance with age, poor estimates for age 11, 19

Nonhomogeneous Variance Correction Levene’s test, Brown-Forsythe correction for ANOVA Not comparable correction for regression; divide predictor into 5 subgroups based on distribution, use ANOVA Use Weighted Least Squares (see text)

Dependent Errors Clustering- data occur in groups (classrooms, clinics, etc.) so that errors are more similar within cluster than between cluster Autocorrelation- data are correlated within person over time (or between clustered pairs, triplets etc.)

Dependent Errors Clustering becomes a category variable in the analysis through dummy coding; separate group models can be constructed through SEM, or the category variable can be included in the path model (which is better depends on homogeneous variance or not)

Dependent Errors Autocorrelation is computed for time data by correlating pairs of data; each pair is the data at one time point with its successor. Eg. (x 1, x 2 ), (x 2, x 3 ), (x 3, x 4 ), etc. The regression of y on itself r uses the autocorrelation y t – r y t-1 = y* t The new dependent variable y* t is now modeled in a regression

Dependent Errors Means correlation among errors in sets of the data (eg. Some siblings in a group of adolescent sample of BASC) –Likely problem only if a meaningful fraction of the total sample size Difficult to determine, separate from random pairings of errors Other data (cluster information) needed, such as age for BASC Anxiety

Dependent Errors Time-related or longitudinal data may have autocorrelation (correlation over time of residuals); Durbin-Watson test gives omnibus test; ARMA model given below: Time1Time2Time3 e2e3...    

MRA ASSUMPTIONS-5 NORMALITY OF RESIDUALS:Violation of this assumption is not a problem –unless SKEWNESS is severe ( >± 4 or 5, maybe even larger) and –KURTOSIS is severe (> 3)Combinations of skewness and kurtosis at the edge of these values may cause problems Effects of violation: –Type I error rates increase, sometimes greatly –Estimates can be biased

Normality of residuals Residual of Predicted Attitude to School Count

Q-Q Plot Plots the quantiles of a variable's distribution against the quantiles of any of a number of test distributions. Probability plots are generally used to determine whether the distribution of a variable matches a given distribution. If the selected variable matches the test distribution, the points cluster around a straight line. Available test distributions include beta, chi-square, exponential, gamma, half-normal, Laplace, Logistic, Lognormal, normal, pareto, Student's t, Weibull, and uniform. Depending on the distribution selected, you can specify degrees of freedom and other parameters. You obtain probability plots for transformed values. Transformation options include natural log, standardize values, difference, and seasonally difference. You can specify the method for calculating expected distributions, and for resolving "ties," or multiple observations with the same value.

Q-Q plot: SPSS GRAPH: Q-Q Observed Value e t r m l V l Normal Q-Q Plot of ATTITUDE TO SCHOOL Residual of prediction

P-P plots Plots a variable's cumulative proportions against the cumulative proportions of any of a number of test distributions. Probability plots are generally used to determine whether the distribution of a variable matches a given distribution. If the selected variable matches the test distribution, the points cluster around a straight line. Available test distributions include beta, chi-square, exponential, gamma, half-normal, Laplace, Logistic, Lognormal, normal, pareto, Student's t, Weibull, and uniform. Depending on the distribution selected, you can specify degrees of freedom and other parameters. You obtain probability plots for transformed values. Transformation options include natural log, standardize values, difference, and seasonally difference. You can specify the method for calculating expected distributions, and for resolving "ties," or multiple observations with the same value.

P-P Plot: SPSS: Graph: P-P Residual of prediction