5. Spatial regression models 5.1 Basic types of spatial regression models There are two basic types of spatial regression models which can be chosen subject.

Slides:



Advertisements
Similar presentations
Cointegration and Error Correction Models
Advertisements

Functional Form and Dynamic Models
Autocorrelation and Heteroskedasticity
Regression Analysis.
Multiple Regression Analysis
The Multiple Regression Model.
Chapter 12 Simple Linear Regression
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Econ 140 Lecture 81 Classical Regression II Lecture 8.
Chapter 12 Simple Linear Regression
Chapter 13 Multiple Regression
Chapter 13 Additional Topics in Regression Analysis
Chapter 10 Simple Regression.
Chapter 12 Simple Regression
Additional Topics in Regression Analysis
The Simple Regression Model
1 Pertemuan 13 Uji Koefisien Korelasi dan Regresi Matakuliah: A0392 – Statistik Ekonomi Tahun: 2006.
Chapter Topics Types of Regression Models
Chapter 11 Multiple Regression.
Economics 20 - Prof. Anderson
Topic 3: Regression.
Inferences About Process Quality
SIMPLE LINEAR REGRESSION
Spatial Methods in Econometrics Daniela Gumprecht Department for Statistics and Mathematics, University of Economics and Business Administration, Vienna.
Simple Linear Regression and Correlation
Introduction to Regression Analysis, Chapter 13,
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
ECON 6012 Cost Benefit Analysis Memorial University of Newfoundland
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Regression Method.
Ms. Khatijahhusna Abd Rani School of Electrical System Engineering Sem II 2014/2015.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
CHAPTER 14 MULTIPLE REGRESSION
INTRODUCTORY LINEAR REGRESSION SIMPLE LINEAR REGRESSION - Curve fitting - Inferences about estimated parameter - Adequacy of the models - Linear.
1 1 Slide Simple Linear Regression Coefficient of Determination Chapter 14 BA 303 – Spring 2011.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Chap 14-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 14 Additional Topics in Regression Analysis Statistics for Business.
Ordinary Least Squares Estimation: A Primer Projectseminar Migration and the Labour Market, Meeting May 24, 2012 The linear regression model 1. A brief.
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
VI. Regression Analysis A. Simple Linear Regression 1. Scatter Plots Regression analysis is best taught via an example. Pencil lead is a ceramic material.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Problems with the Durbin-Watson test
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
Chapter 15 Panel Data Models Walter R. Paczkowski Rutgers University.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Regression Analysis: A statistical procedure used to find relations among a set of variables B. Klinkenberg G
The Probit Model Alexander Spermann University of Freiburg SS 2008.
Heteroscedasticity Heteroscedasticity is present if the variance of the error term is not a constant. This is most commonly a problem when dealing with.
The Probit Model Alexander Spermann University of Freiburg SoSe 2009
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
REGRESSION DIAGNOSTIC IV: MODEL SPECIFICATION ERRORS
Simultaneous equation system
Fundamentals of regression analysis
Chapter 11 Simple Regression
Serial Correlation and Heteroskedasticity in Time Series Regressions
Spatial Econometric Analysis
Serial Correlation and Heteroscedasticity in
Simple Linear Regression
Tutorial 1: Misspecification
Chapter 7: The Normality Assumption and Inference with OLS
Autocorrelation MS management.
Financial Econometrics Fin. 505
Serial Correlation and Heteroscedasticity in
Introduction to Regression
Presentation transcript:

5. Spatial regression models 5.1 Basic types of spatial regression models There are two basic types of spatial regression models which can be chosen subject to the results of the LM tests in the standard regression model: - the spatial cross-regressive model, - the spatial lag model, - the spatial error model. Spatial cross-regressive Model Substantive spatial dependence can be captured by a spatial lags in the explora- tory variables X 2, X 3, …, X k or the endogenous variable Y. In the former case, the spatial lag variables Wx 2, Wx 3, …, Wx k will be incorporated into the standard regression model as additional regressors. We term the regression model with spa- tially lagged exogenous regressors spatial cross-regressive model. Substantive spatial interaction can occur in different applications. Output growth of a region may Not only depend on own region’s initial income but as well on income in adjacent re- regions. In this case, spillover effects are restricted to neighbourhood regions. Such a restriction may especially hold for spillovers of tacit knowledge which is expected to be exchanged within local areas. Parameter estimation in the cross-regressive model can be performed as in the standard regression model by OLS. This results from the fact that spatial lag vari- ables share the properties with the original regressors, which are assumed to be non-stochastic.

Spatial lag model The spatial lag model captures as well substantial spatial dependencies like external effects or spatial interactions. It assumes that such dependencies mani- fests in the spatial lag Wy of the dependent variable Y. Regional growth may be fostered by growth in neighbourhood regions by flows of goods for example. In this case, spillover effects are not restricted to adjacent regions but propagated over the entire regional system. In accordance to the the time-series analogue the pure spatial lag model is also termed spatial autoregressive (SAR) model. In applications the model also in- corporates a set of explanatory variables X 1, X 2, …, X k. This extension is ex- pressed by the term mixed regressive, spatial autoregressive model. In all instances OLS estimation will produce biased and inconsistent parameter esti- mates, we introduce the method of instruments (IV method) and the method of maximum likelihood (ML method) as adequate estimation methods for that type of model. Because only the spatial lag Wy is relevant for the choice of an alternative estimation method to OLS, the term spatial lag mo-del is often kept in cases where the model is extended by exogeneous X-variables.

Spatial error model The spatial error model is applicable when spatial autocorrelation occurs as nuisance resulting form misspecification or inadequate delineation of spatial units. Unmodelled interaction among regions are restricted to the error terms. In convergence studies, the convergence rate will be properly assessed by sta- dard estimation methods. However, a random shock occurring in a specific region Is not restricted to that region and its neighbourhood but will diffuse across the entire regional system. Spatial dependence in for of nuisance entails that the disturbances ε i are no longer independently identically distributed (i.i.d.), but follow an autoregressive (AR) or moving average (MA) process. In analogy to the Markov process in time-series analysis, the disturbance term is assumed to folllow a first order autoregressive (AR) process in the spatial error model. In consideration of the explanation of the dependent variable Y by a set of exogenous variables X 1, X 2, …, X k, the spatial error model serves as an abbreviation for a linear regression model with a spa- tial autoregressive disturbance. In contrast to substantial dependence, spatial dependence in form of nuisance does not entail inconsistency of OLS estimated regression coefficients. However, As their standard errors are biased, significance tests based on OLS estimation can be misleading. In order to allow for valid inference, other estimation principles

must adopted. We outline maximum likelihood (ML) estimation in the spatial error model. An alternative is provided by Kelejian and Prucha (1999) in form of a General Moment (GM) estimator which is not dealt with in our introductory course. Specification tests The presence of spatial error autocorrelation can be assessed by the Moran test applied on the residuals of the standard regression model (see section 4.2). This omnibus test does, however, not point to a basic spatial model. No specific spatial test is available for the spatial cross-regressive model. Florax and Fol- mer (1992) suggest to apply the well-known F-test for linear restrictions on the regression coefficients to identify spatial autocorrelation due to omitted lagged exogenous variables. This test requires the estimation of both the restricted and the unrestricted regression model. For the two other type of spatial models Lagrange Multiplier (LM) tests that are tailured for the special spatial settings are available: - the LM lag test and - the LM error test.

Both tests have to be performed after estimating the the standard regression mo-del (see section 4.2). If the test statistic LM(lag) turns out to be significant, but LM(error) is insignificant, spatial autocorrelation appears to be substantial, which means that the spatial lag model is viewed to be appropriate. In the converse case the spatial error model will be a sensible choice for analysing the relation-ship between Y and the X-variables. When both test statistics, LM(lag) and LM(error) prove to be significant, the test statistic with the higher significance, i.e. the lower p-value, will guide the choice of the spatial regres- sion model. Ac-cording to this rule, in case of p[LM(lag)] < p[LM(err)] the spatial lag model will be chosen, whereas for p[LM(err)] < p[LM(lag)] the spatial error model will be the preferable spatial setting.

5.2 The spatial cross-regressive model The spatial cross-regressive model presumes that the exploratory variables X 2, X 3, …, X k as well as their spatial lags LX 2, …, LX k influence a geo-referenced dependent variable Y. In this approach Y is not only affected by values variables take in the same region but also the take in neighbouring regions: (5.1) y is an nx1 vector of the endogenous variable Y, x j an nx1 vector of the exoge- nous variable X j (where x 1 is a vector of ones for the intercept), W an nxn spatial weight matrix and ε an nxn vector of disturbances. The parameters β j j=1, 2,…,k denote theregression coefficients of the exogenous variables X 1, X 2, …, X k and the parameters γ j the regression coefficients of the exogneous spatial lags Wx 2, …, Wx k. The disturbances ε i are assumed to meet the standard assumptions for a linear regression model (see section 4.1): expection of zero, constant variance σ² and absence of autocorrelation. For statistical inferences like significance tests we additionally assume the disturbances to be normally distributed.

A more compact form of the spatial cross-regressive model reads (5.2) nxk matrix with observations of the k explanatory variables: kx1 vector of regression coefficients of exognous variables: (k-)x1 vector of regression coefficients of lagged exognous variables: (n-1)xk matrix with observations of the k-1 lagged explanatory variables:. OLS estimator of the Regression coefficients:

- F test on omitted spatially lagged exogenous variables Test of the spatially lagged variables for a relevant subset S of X-variables one at a time Null hypothesis H 0 : γ j = 0 for subset X j ε S SSR c : Constrained residual sum of squares from a regression in which H 0 holds i.e. a regression of Y on the original exogenous variables X 1, X 2, …, X k SSR u : Unconstrained residual sum of squares from a regression of Y on the original exogenous variables X 1, X 2, …, X k and the spatially lagged exo- genous variable LX j Test statistic: (5.3) F follows an F distribution with 1 and n-k degrees of freedom. Testing decision: F > F(1;n-k-1;1-α) => Reject H 0 or p Reject H 0

Generalisation of the F test: Instead of testing the spatially lagged variables of the subset S one at a time, they can also be tested simultanously Null hypothesis H 0 : All regression coefficients γ j of the X-variables of the subset S are equal to zero Test statistic: (5.4) q: number of X-variables in the subset S (when all X-variables without the vec- tor of one are included in S, q is equal to k – 1) F follows an F distribution with q and n-k degrees of freedom. Testing decision: F > F(q;n-k-q;1-α) => Reject H 0 or p Reject H 0

Example: In order to illustrate the spatial cross-regressive model, we refer to the example of 5 regions for which data are available on output growth (X) and productivity growth (Y): Region12345 Output growth (X) Productivity growth (Y) The cross-regressive model presumes that productivity growth in region i does Not only depend on own region‘s output growth but as well on output growth in adjacent regions (5.5) with x i1 =1 for all i and x i2 = x i. When LX captures technological externalities from neighbouring regions γ 2 is expected to take a positive sign. When output growth X can be treated as as exogenous variable, the spatial lag variable LX is as well exogneous, because the spatial weights are determined a priori. Thus, the spatial regression model (5.5) can be estimated by OLS.

Vector of the endogenous variable y: Standardized weights matrix W: Matrix of spatially lagged endogenous variables X* : Matrix of original x-variables X: Observation matrix [X X*]:

Matrix product [X X*]‘[X X*]: Inverse ([X X*]‘[X X*]) -1 : Matrix product [X X*]‘y: OLS estimator of the regression coefficients:

Vector of fitted values : Vector of residuals e:

Residual variance : Standard error of regression (SER):

i  Coefficient of determination Working table ( ) SST = , SSE = , SSR = SST – SSE = = or

Critical value (α=0.05, two-sided test): t(3,0.975) = Test statistic: Test of significance of regression coefficients - for β 1 (H 0 : β 1 = 0) OLS estimator for β 1 : - for β 2 (H 0 : β 2 = 0) OLS estimator for β 2 : Test statistic: Testing decision: (| t 2 |=5.018) > [t(3;0.975)=3.182] => Reject H 0 Critical value (α=0.05, two-sided test): t(3,0.975) = Testing decision: (| t 1 |=0.144) Accept H 0

Critical value (α=0.05, two-sided test): t(3,0.975) = Test statistic: Test of significance of regression coefficients - for γ 2 (H 0 : γ 2 = 0) OLS estimator for γ 2 : Testing decision: (| t 1 |=1.560) Accept H 0

F test for the regression as a whole Null hypothesis H 0 : β 2 = γ 2 =0 Constrained residual sum of squares: SSR c = SST = Unconstrained residual sum of squares: SSR u = SSR = Test statistic: or (The difference of both computations of F are only due to rounding errors.) Critical value(α=0.05): F(1;2;0.95) = 18.5 Testing decision: (F=49.364) > [F(1;2;0.95)=18.5] => Reject H 0

F test on omitted spatially lagged exogenous variables Null hypothesis H 0 : γ 2 =0 Constrained residual sum of squares (regression of Y on X 1 (=constant) and X 2 ): SSR c = SST = (see sect. 4.1  standard regression model) Unconstrained residual sum of squares (regression of Y on X 1 (constant), X 2 and LX 2 ): SSR u = SSR = Test statistic: Critical value(α=0.05): F(1;2;0.95) = 18.5 Testing decision: (F=2.409) Accept H 0