Some issues in multivariate regression

Slides:



Advertisements
Similar presentations
Functional Form and Dynamic Models
Advertisements

Multiple Regression.
Autocorrelation and Heteroskedasticity
Properties of Least Squares Regression Coefficients
Econometric Modeling Through EViews and EXCEL
Multiple Regression Analysis
Multiple Regression W&W, Chapter 13, 15(3-4). Introduction Multiple regression is an extension of bivariate regression to take into account more than.
The Simple Regression Model
Structural Equation Modeling
CHAPTER 3: TWO VARIABLE REGRESSION MODEL: THE PROBLEM OF ESTIMATION
NOTATION & ASSUMPTIONS 2 Y i =  1 +  2 X 2i +  3 X 3i + U i Zero mean value of U i No serial correlation Homoscedasticity Zero covariance between U.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
1 Regression Models & Loss Reserve Variability Prakash Narayan Ph.D., ACAS 2001 Casualty Loss Reserve Seminar.
12-1 Multiple Linear Regression Models Introduction Many applications of regression analysis involve situations in which there are more than.
Specifying an Econometric Equation and Specification Error
Chapter 13 Additional Topics in Regression Analysis
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 7. Specification and Data Problems.
GRA 6020 Multivariate Statistics The regression model OLS Regression Ulf H. Olsson Professor of Statistics.
Specific to General Modelling The traditional approach to econometrics modelling was as follows: 1.Start with an equation based on economic theory. 2.Estimate.
Chapter 4 Multiple Regression.
Multicollinearity Omitted Variables Bias is a problem when the omitted variable is an explanator of Y and correlated with X1 Including the omitted variable.
Chapter 15 Panel Data Analysis.
Basic Mathematics for Portfolio Management. Statistics Variables x, y, z Constants a, b Observations {x n, y n |n=1,…N} Mean.
Dealing with Heteroscedasticity In some cases an appropriate scaling of the data is the best way to deal with heteroscedasticity. For example, in the model.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Multivariate Regression Analysis Estimation. Why Multiple Regression? Suppose you want to estimate the effect of x1 on y, but you know the following:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 3) Slideshow: prediction Original citation: Dougherty, C. (2012) EC220 - Introduction.
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
1 PREDICTION In the previous sequence, we saw how to predict the price of a good or asset given the composition of its characteristics. In this sequence,
2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)
Specification Error I.
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Chap 14-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 14 Additional Topics in Regression Analysis Statistics for Business.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Six.
1Spring 02 First Derivatives x y x y x y dy/dx = 0 dy/dx > 0dy/dx < 0.
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
May 2004 Prof. Himayatullah 1 Basic Econometrics Chapter 7 MULTIPLE REGRESSION ANALYSIS: The Problem of Estimation.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Problems with the Durbin-Watson test
Chapter 22: Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
Correlation. Up Until Now T Tests, Anova: Categories Predicting a Continuous Dependent Variable Correlation: Very different way of thinking about variables.
5. Consistency We cannot always achieve unbiasedness of estimators. -For example, σhat is not an unbiased estimator of σ -It is only consistent -Where.
Lesson 14 - R Chapter 14 Review. Objectives Summarize the chapter Define the vocabulary used Complete all objectives Successfully answer any of the review.
Multiple Regression Analysis Regression analysis with two or more independent variables. Leads to an improvement.
MULTICOLLINEARITY: WHAT HAPPENS IF THE REGRESSORS ARE CORRELATED?
The Instrumental Variables Estimator The instrumental variables (IV) estimator is an alternative to Ordinary Least Squares (OLS) which generates consistent.
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Heteroscedasticity Heteroscedasticity is present if the variance of the error term is not a constant. This is most commonly a problem when dealing with.
Basic Estimation Techniques
Micro Economics in a Global Economy
Regression Diagnostics
Further Inference in the Multiple Regression Model
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
Fundamentals of regression analysis 2
STOCHASTIC REGRESSORS AND THE METHOD OF INSTRUMENTAL VARIABLES
Chapter 15 Panel Data Analysis.
Basic Estimation Techniques
Managerial Economics in a Global Economy
Chapter 6: MULTIPLE REGRESSION ANALYSIS
The regression model in matrix form
Goodness of Fit The sum of squared deviations from the mean of a variable can be decomposed as follows: TSS = ESS + RSS This decomposition can be used.
Econometrics Chengyaun yin School of Mathematics SHUFE.
Tutorial 1: Misspecification
Econometrics Analysis
Chapter 13 Additional Topics in Regression Analysis
Tutorial 6 SEG rd Oct..
Multicollinearity What does it mean? A high degree of correlation amongst the explanatory variables What are its consequences? It may be difficult to separate.
Testing Causal Hypotheses
Presentation transcript:

Some issues in multivariate regression 1. Increasing the number of right-hand side variables will always increase the R-squared for the regression. 2. If the right-hand side variables are highly correlated with each other then the standard errors of the OLS coefficients will become large. 3. If we omit relevant variables then the OLS estimates will be biased. 4. If we include irrelevant variables the OLS estimates will be unbiased but inefficient.

Adjusted R-squared Adding extra regressors to the RHS of an equation will always increase R-squared. Therefore it does not make sense to use R-squared as a basis for choosing a set of regressors. An alternative is the adjusted R-squared or R-bar squared. This penalises the loss of degrees of freedom when we add extra regressors.

The regression model in mean deviation form Suppose we have: We can redefine the variables as: This will make the algebra considerably easier in our analysis of the multivariate model.

Multicollinearity Consider the regression model The variance covariance matrix of the regression coefficients can be derived as: where

Now consider the D term. This can be written: where is the correlation coefficient for the two X variables. It follows that that closer the correlation between the X variables, the closer D will be to zero and the larger the variance-covariance matrix for the OLS coefficients.

If the x variables in a regression equation are highly correlated then this can result in estimates with a high variance. This is is known as the problem of multicollinearity. The usual signs of multicollinearity are: For individual variables the standard errors are large and the t-ratios are low. For the regression as a whole the F-statistic is highly significant. The R2 will tend to be high. Some degree of multicollinearity is almost always present in econometric models.

Omitted Variables Estimated Model True Model OLS is therefore biased

Inclusion of Irrelevant Variables Estimated Model True Model In this case OLS is unbiased but inefficient