Checking Assumptions Primary Assumptions Secondary Assumptions

Slides:



Advertisements
Similar presentations
Multiple Regression W&W, Chapter 13, 15(3-4). Introduction Multiple regression is an extension of bivariate regression to take into account more than.
Advertisements

Forecasting Using the Simple Linear Regression Model and Correlation
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
Psychology 202b Advanced Psychological Statistics, II February 3, 2011.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: What it Is and How it Works. Overview What is Bivariate Linear Regression? The Regression Equation How It’s Based on r Assumptions.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 14 Using Multivariate Design and Analysis.
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
Lecture 6: Multiple Regression
Chapter Topics Types of Regression Models
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
Predictive Analysis in Marketing Research
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
Business Statistics - QBM117 Statistical inference for regression.
Multiple Regression Research Methods and Statistics.
Chapter 7 Forecasting with Simple Regression
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Multiple Regression Dr. Andy Field.
Simple Linear Regression Analysis
Correlation & Regression
Regression and Correlation Methods Judy Zhong Ph.D.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
STA302/ week 111 Multicollinearity Multicollinearity occurs when explanatory variables are highly correlated, in which case, it is difficult or impossible.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Chapter 12 Examining Relationships in Quantitative Research Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Maths Study Centre CB Open 11am – 5pm Semester Weekdays
Chapter 16 Data Analysis: Testing for Associations.
REGRESSION DIAGNOSTICS Fall 2013 Dec 12/13. WHY REGRESSION DIAGNOSTICS? The validity of a regression model is based on a set of assumptions. Violation.
Correlation and Regression: The Need to Knows Correlation is a statistical technique: tells you if scores on variable X are related to scores on variable.
Multivariate Data Analysis Chapter 2 – Examining Your Data
Multiple regression.
Ch14: Linear Least Squares 14.1: INTRO: Fitting a pth-order polynomial will require finding (p+1) coefficients from the data. Thus, a straight line (p=1)
Multiple Regression Analysis Regression analysis with two or more independent variables. Leads to an improvement.
Venn diagram shows (R 2 ) the amount of variance in Y that is explained by X. Unexplained Variance in Y. (1-R 2 ) =.36, 36% R 2 =.64 (64%)
Multiple Linear Regression An introduction, some assumptions, and then model reduction 1.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
Multiple Regression Scott Hudson January 24, 2011.
Regression and Correlation of Data Summary
Inference for Least Squares Lines
Modeling in R Sanna Härkönen.
Lecture 10 Regression Analysis
Bivariate & Multivariate Regression Analysis
Correlation, Bivariate Regression, and Multiple Regression
Multiple Regression Prof. Andy Field.
AP Statistics Chapter 14 Section 1.
The Correlation Coefficient (r)
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Multiple Regression.
Chapter 12: Regression Diagnostics
BIVARIATE REGRESSION AND CORRELATION
Lecture 14 Review of Lecture 13 What we’ll talk about today?
CHAPTER 29: Multiple Regression*
Multiple Regression A curvilinear relationship between one variable and the values of two or more other independent variables. Y = intercept + (slope1.
Residuals The residuals are estimate of the error
No notecard for this quiz!!
I271b Quantitative Methods
Solution 9 1. a) From the matrix plot, 1) The assumption about linearity seems ok; 2).The assumption about measurement errors can not be checked at this.
Regression is the Most Used and Most Abused Technique in Statistics
Simple Linear Regression
Chapter 13 Additional Topics in Regression Analysis
Regression and Correlation of Data
Multicollinearity Multicollinearity occurs when explanatory variables are highly correlated, in which case, it is difficult or impossible to measure their.
The Correlation Coefficient (r)
Presentation transcript:

Checking Assumptions Primary Assumptions Secondary Assumptions Adequate variance in each IV (do your predictor variables vary?) Absence of influential cases, e.g., outliers (do all cases exert approximately the same influence on the regression coefficients?) Linearity – does a straight line fit the data points reasonably well? Secondary Assumptions Constant Error Variance (homoscedasticity) Normally distributed error variance In cases that involve moderate violations of these assumptions, the regression analysis is weaken, not invalidated (Tabachnick & Fidell, 1996).

Outliers A case that differs substantially from the main trend of the data Affect the values of the estimated regression coefficients and cause bias Changes in the slope and intercept Detect outliers by examining frequency distributions for individual variables or assessing residuals for multivariate outliers.

r = .62 M = 18.1 Outlier r = .95 M = 19.3

Influential Case – This case is exerting a massive influence on the data, but the outlier produces a very small standardized residual.

Multicollinearity Strong correlation between two or more independent variables in the regression model. Avoid putting two variable that are essentially the same into the equation. Perfect collinearity makes it impossible to obtain unique estimates of the regression coefficients. If two IVs are perfectly correlated then the value of b for each variable is interchangeable Low levels of collinearity pose little threat to regression models As collinearity increases so does the SE of the b-coefficients, Increased standard errors in turn means that coefficients for some independent variables may be found not to be significantly different from 0. In other words, by overinflating the standard errors, multicollinearity makes some variables statistically insignificant when they should be significant. So examine the bivariate correlations between the IVs, and look for “big” values