Econometrics III Evgeniya Anatolievna Kolomak, Professor.

Slides:



Advertisements
Similar presentations
SJS SDI_21 Design of Statistical Investigations Stephen Senn 2 Background Stats.
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Multiple Regression Analysis
Classical Linear Regression Model
The Simple Regression Model
CHAPTER 3: TWO VARIABLE REGRESSION MODEL: THE PROBLEM OF ESTIMATION
Part 12: Asymptotics for the Regression Model 12-1/39 Econometrics I Professor William Greene Stern School of Business Department of Economics.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Linear regression models
Ch11 Curve Fitting Dr. Deshi Ye
Definition  Regression Model  Regression Equation Y i =  0 +  1 X i ^ Given a collection of paired data, the regression equation algebraically describes.
The General Linear Model. The Simple Linear Model Linear Regression.
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
The Simple Linear Regression Model: Specification and Estimation
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
CHAPTER 3 ECONOMETRICS x x x x x Chapter 2: Estimating the parameters of a linear regression model. Y i = b 1 + b 2 X i + e i Using OLS Chapter 3: Testing.
Simple Linear Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
The Simple Regression Model
2. Point and interval estimation Introduction Properties of estimators Finite sample size Asymptotic properties Construction methods Method of moments.
Chapter Topics Types of Regression Models
Chapter 11 Multiple Regression.
Simple Linear Regression Analysis
Topic4 Ordinary Least Squares. Suppose that X is a non-random variable Y is a random variable that is affected by X in a linear fashion and by the random.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression Analysis
Lecture 10A: Matrix Algebra. Matrices: An array of elements Vectors Column vector Row vector Square matrix Dimensionality of a matrix: r x c (rows x columns)
Regression and Correlation Methods Judy Zhong Ph.D.
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Chapter 11 Simple Regression
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
LECTURE 2. GENERALIZED LINEAR ECONOMETRIC MODEL AND METHODS OF ITS CONSTRUCTION.
Introduction to Linear Regression
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin The Two-Variable Model: Hypothesis Testing chapter seven.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Chapter 8: Simple Linear Regression Yang Zhenlin.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
Chap 5 The Multiple Regression Model
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Lecturer: Ing. Martina Hanová, PhD..  How do we evaluate a model?  How do we know if the model we are using is good?  assumptions relate to the (population)
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
THE LINEAR REGRESSION MODEL: AN OVERVIEW
Evgeniya Anatolievna Kolomak, Professor
ECONOMETRICS DR. DEEPTI.
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
6-1 Introduction To Empirical Models
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Simple Linear Regression
J.-F. Pâris University of Houston
Simple Linear Regression
Linear Regression Summer School IFPRI
Ch3 The Two-Variable Regression Model
Introduction to Regression
Regression Models - Introduction
Presentation transcript:

Econometrics III Evgeniya Anatolievna Kolomak, Professor

Theme 1. The classical multiple linear regression

The classical multiple linear regression. Model specification Multiple linear regression model: y i =f(x i1, x i2,…, x iK )+ε i y i =β 1 x i1 +β 2 x i2 +…+β K x iK +ε i, i=1,…,n Where y – dependent (explained) variable or regressand; x 1,x 2,…, x K - independent (explanatory) variables or regressors; ε i - random disturbance (errors of the variables).

The classical multiple linear regression. Model specification We assume that each observation (y i, x i1, x i2,…,x iK ), i=1,…,n is generated by the process described by y i =β 1 x i1 +β 2 x i2 +…+β k x iK + ε i deterministic part random part

The classical multiple linear regression. Assumptions of the model 1.Linear functional form 2.Identifiability of the model parameters 3.Expected value of the disturbance given observed information 4.Variance and covariance of the disturbances given observed information 5.Nature of the independent variables 6.Probability distribution of the stochastic part of the model

Assumptions of the model. 1. Linear functional form Let y - column vector (nx1) of the observations y 1,y 2,..,y n x k – column vector (nx1) of the observations on variable x k ; X – matrix (nxK) assemble x k ; ε - column vector (nx1) containing the n disturbances. The model y i =β 1 x i1 +β 2 x i2 +…+β k x iK +ε i can be written y= x 1 β 1 +…..+x k β k +ε y=Xβ+ε Our primary interest is estimation and inference about β

Assumptions of the model. 2. Identifiability of the model parameters Identification condition: X is n x K matrix with rank K The columns of X are linearly independent and there are at least K observations.

Assumptions of the model. 3. Expected value of the disturbance ε i The disturbance conditioned on observations X is assumed to have expected value 0 at every observation E[ε i │X]=0 for all i=1,…,n The assumption implies E[y│X]=Xβ

Assumptions of the model. 4. Variance and covariance of the disturbances Constant variance or homoscedasticity Var[ε i │X]=σ 2 for all i=1,…,n Absence of correlation across observations or absence of autocorrelation Cov[ε i, ε j │X]=0 for all i≠j Summarizing E[εε T │X]= σ 2 I I – identity matrix

Assumptions of the model. 5. Non-stochastic regressors X is known n x K matrix of constants X is non-stochastic matrix X and ε are uncorrelated

Assumptions of the model. 6. Probability distribution of ε The disturbances ε i are normally distributed ε│X ~ N[0, σ 2 I] Normality enables to obtain several exact statistical results and to construct test statistics.

Assumptions of the classical regression model. Summary NumberAssumption A1y=Xβ+ε A2X is n x K matrix with rank K A3E[ε│X]=0 A4E[εε T │X]= σ 2 I A5X is non-stochastic matrix A6ε│X ~ N[0, σ 2 I]

Least squares regression The population quantities are β and ε i The sample estimations are b and e i For any value of b we estimate e i = y i - x i T b y i = x i T b + e i Vector b is chosen so that the fitted line x i T b is close to the data points. The fitting criterion is least squares.

Least squares regression The solving problem is: Minimize b e T e = (y – Xb) T (y – Xb) b satisfies the least squares normal equations: X T Xb=X T y If the inverse of X T X exists, which follows from the full rank assumption, then b=(X T X) -1 X T y

Least squares regression Algebraic properties of the least squares Let the first column of X is 1s, then 1. The least squares residuals sum to zero. Σ i e i =0 2. The regression passes through the point of means of the data y=x T b 3. The mean of fitted values from the regression equals the mean of the actual values ̂ y=Xb

Partitioned regression Suppose the regression involves two set of variables X 1 and X 2. However we are interested in estimates for β 2 only. The normal equations are: We first solve for b 1 b 1 is the coefficient vector in the regression of (y-x 2 T b 2 ) on x 1.

Partitioned regression Given the solution for b 1 we solve the second equation I – identity matrix and Matrix M 1 ∙is: 1) symmetric ( ) and 2) idempotent ( ). Proof: 1. 2.

Partitioned regression Properties of matrixes M 1 and P M 1 ∙y – vector of residuals in the regression of y on x 1 M 1 ∙x 2 – matrix of residuals of the regressions of x 2 on x 1 Let x 2 * =M 1 ∙x 2 and y * =M 1 ∙y. Then

Partitioned regression Theorem Frisch-Waugh. The sub-vector b 2 is the set of coefficients obtained when the residuals from a regression of y on X 1 are regressed on the set of residuals obtained when each column of X 2 is regressed on X 1. The algorithm is as follows: 1.To regress y on X 1 and to estimate residuals e y. 2.To regress each column of X 2 on X 1 and estimate residuals e x 2. 3.To regress e y on e x 2 and estimate coefficients b 2 and residuals e.

Goodness of fit and the analysis of variance The total variation in y: TSS=Σ i=1 n (y i - y) 2 In terms of regression equation: y i =x i T b+e i y i - y = (x i - x) T b + e i regression part error part For the full set of observations: Total sum of squares = regression sum of squares + error sum of squares TSS=RSS+ESS

Goodness of fit and the analysis of variance Coefficient of determination:

Goodness of fit and the analysis of variance SourceDegrees of freedom Mean Square RegressionK-1 (assuming a constant term) Residualsn-k Totaln-1 Coefficient of determination

Statistical properties of the Least Squares Estimator in finite sample Gaus-Markov Theorem. In the classical liner regression model y=Xβ+ε, the least squares estimator b is the minimum variance linear unbiased estimator of β

Statistical properties of the Least Squares Estimator in finite sample If the disturbances ε i are normally distributed ε│X ~ N[0, σ 2 I] then b│X ~ N[β, σ 2 (X T X) -1 ] b k │X ~ N[β k, σ 2 (X T X) -1 kk ] Rao-Blackwell Theorem. In the classical liner regression model y=Xβ+ε with normally distributed disturbance the least squares estimator b has the minimum variance of all unbiased estimators.

Statistical properties of the Least Squares Estimator in finite sample If we wish to test hypotheses about β or to construct confidence intervals, we need an estimate of the covariance matrix Var[b]= σ 2 (X T X) -1 Since e i is an estimate of ε i, estimation of σ 2 is And Est. Var[b]= s 2 (X T X) -1

Statistical properties of the Least Squares Estimator in finite sample