Chengyuan Yin School of Mathematics

Slides:



Advertisements
Similar presentations
Christopher Dougherty EC220 - Introduction to econometrics (chapter 8) Slideshow: model b: properties of the regression coefficients Original citation:
Advertisements

Session 2 – Introduction: Inference
Distributions of sampling statistics Chapter 6 Sample mean & sample variance.
Part 11: Asymptotic Distribution Theory 11-1/72 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Part 12: Asymptotics for the Regression Model 12-1/39 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: asymptotic properties of estimators: plims and consistency Original.
Matrix Algebra Matrix algebra is a means of expressing large numbers of calculations made upon ordered sets of numbers. Often referred to as Linear Algebra.
Matrix Algebra Matrix algebra is a means of expressing large numbers of calculations made upon ordered sets of numbers. Often referred to as Linear Algebra.
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
ASYMPTOTIC PROPERTIES OF ESTIMATORS: PLIMS AND CONSISTENCY
DISTRIBUTION OF THE SAMPLE MEAN
Section 6.1 Let X 1, X 2, …, X n be a random sample from a distribution described by p.m.f./p.d.f. f(x ;  ) where the value of  is unknown; then  is.
Generalized Regression Model Based on Greene’s Note 15 (Chapter 8)
Least Squares Asymptotics Convergence of Estimators: Review Least Squares Assumptions Least Squares Estimator Asymptotic Distribution Hypothesis Testing.
Probability theory 2011 Convergence concepts in probability theory  Definitions and relations between convergence concepts  Sufficient conditions for.
Topic4 Ordinary Least Squares. Suppose that X is a non-random variable Y is a random variable that is affected by X in a linear fashion and by the random.
Cross-sectional:Observations on individuals, households, enterprises, countries, etc at one moment in time (Chapters 1–10, Models A and B). 1 During this.
CHAPTER 11: Sampling Distributions
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
CHAPTER 11: Sampling Distributions
Limits and the Law of Large Numbers Lecture XIII.
All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.
Statistical Techniques I EXST7005 Distribution of Sample Means.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Chapter 10 – Sampling Distributions Math 22 Introductory Statistics.
Chapter 5.6 From DeGroot & Schervish. Uniform Distribution.
CHAPTER 11: Sampling Distributions ESSENTIAL STATISTICS Second Edition David S. Moore, William I. Notz, and Michael A. Fligner Lecture Presentation.
Stat 1510: Sampling Distributions
405 ECONOMETRICS Domodar N. Gujarati Prof. M. El-Sakka
Generalised method of moments approach to testing the CAPM Nimesh Mistry Filipp Levin.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
1 We will now look at the properties of the OLS regression estimators with the assumptions of Model B. We will do this within the context of the simple.
Sampling Distributions Chapter 18. Sampling Distributions A parameter is a measure of the population. This value is typically unknown. (µ, σ, and now.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Review of Statistics.  Estimation of the Population Mean  Hypothesis Testing  Confidence Intervals  Comparing Means from Different Populations  Scatterplots.
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
INSTRUMENTAL VARIABLES 1 Suppose that you have a model in which Y is determined by X but you have reason to believe that Assumption B.7 is invalid and.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Sampling Distributions Chapter 18. Sampling Distributions A parameter is a number that describes the population. In statistical practice, the value of.
Econometrics III Evgeniya Anatolievna Kolomak, Professor.
Large Sample Distribution Theory
Large Sample Theory EC 532 Burak Saltoğlu.
Standard Errors Beside reporting a value of a point estimate we should consider some indication of its precision. For this we usually quote standard error.
Chapter 7. Classification and Prediction
Vera Tabakova, East Carolina University
Sampling Distributions and Estimation
Sample Mean Distributions
Evgeniya Anatolievna Kolomak, Professor
Fundamentals of regression analysis
Sampling Distributions
Parameter, Statistic and Random Samples
Sampling Distribution of Means: Basic Theorems
Large Sample Theory EC 532 Burak Saltoğlu.
CONCEPTS OF ESTIMATION
The Multivariate Normal Distribution, Part 2
Chengyuan Yin School of Mathematics
Econometrics Chengyaun yin School of Mathematics SHUFE.
CHAPTER 15 SUMMARY Chapter Specifics
Chengyuan Yin School of Mathematics
Econometrics Chengyuan Yin School of Mathematics.
Charles University Charles University STAKAN III
Lecturer Dr. Veronika Alhanaqtah
Econometrics I Professor William Greene Stern School of Business
Chapter 5 Properties of a Random Sample
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Econometrics I Professor William Greene Stern School of Business
Applied Statistics and Probability for Engineers
Presentation transcript:

Chengyuan Yin School of Mathematics Econometrics Chengyuan Yin School of Mathematics

11. Asymptotic Distribution Theory Econometrics 11. Asymptotic Distribution Theory

Preliminary This and our class presentation will be a moderately detailed sketch of these results. More complete presentations appear in Chapter 4 of your text. Please read this chapter thoroughly. We will develop the results that we need as we proceed. Also, (I believe) that this topic is the most difficult conceptually in this course, so do feel free to ask questions in class.

Asymptotics: Setting Most modeling situations involve stochastic regressors, nonlinear models or nonlinear estimation techniques. The number of exact statistical results, such as expected value or true distribution, that can be obtained in these cases is very low. We rely, instead, on approximate results that are based on what we know about the behavior of certain statistics in large samples. Example from basic statistics: What can we say about 1/ We know a lot about . What do we know about its reciprocal?

Convergence Definitions, kinds of convergence as n grows large: 1. To a constant; example, the sample mean, 2. To a random variable; example, a t statistic with n -1 degrees of freedom

Convergence to a Constant Sequences and limits. Sequence of constants, indexed by n (n(n+1)/2 + 3n + 5) Ordinary limit: --------------------------  ? (n2 + 2n + 1) (The use of the “leading term”) Convergence of a random variable. What does it mean for a random variable to converge to a constant? Convergence of the variance to zero. The random variable converges to something that is not random.

Convergence Results Convergence of a sequence of random variables to a constant - convergence in mean square: Mean converges to a constant, variance converges to zero. (Far from the most general, but definitely sufficient for our purposes.) A convergence theorem for sample moments. Sample moments converge in probability to their population counterparts. Generally the form of The Law of Large Numbers. (Many forms; see Appendix D in your text.) Note the great generality of the preceding result. (1/n)Σig(zi) converges to E[g(zi)].

Probability Limit

Mean Square Convergence

Probability Limits and Expecations What is the difference between E[xn] and plim xn?

Consistency of an Estimator If the random variable in question, xn is an estimator (such as the mean), and if plim xn = θ Then xn is a consistent estimator of θ. Estimators can be inconsistent for two reasons: (1) They are consistent for something other than the thing that interests us. (2) They do not converge to constants. They are not consistent estimators of anything. We will study examples of both.

The Slutsky Theorem Assumptions: If xn is a random variable such that plim xn = θ. For now, we assume θ is a constant. g(.) is a continuous function with continuous derivatives. g(.) is not a function of n. Conclusion: Then plim[g(xn)] = g[plim(xn)] assuming g[plim(xn)] exists. (VVIR!) Works for probability limits. Does not work for expectations.

Slutsky Corollaries

Slutsky Results for Matrices Functions of matrices are continuous functions of the elements of the matrices. Therefore, If plimAn = A and plimBn = B (element by element), then Plim(An-1) = [plim An]-1 = A-1 and plim(AnBn) = plimAnplim Bn = AB

Limiting Distributions Convergence to a kind of random variable instead of to a constant xn is a random sequence with cdf Fn(xn). If plim xn = θ (a constant), then Fn(xn) becomes a point. But, Fn may converge to a specific random variable. The distribution of that random variable is the limiting distribution of xn.

Limiting Distribution

A Slutsky Theorem for Random Variables (Continuous Mapping)

An Extension of the Slutsky Theorem

Application of the Slutsky Theorem

Central Limit Theorems Central Limit Theorems describe the large sample behavior of random variables that involve sums of variables. “Tendency toward normality.” Generality: When you find sums of random variables, the CLT shows up eventually. The CLT does not state that means of samples have normal distributions.

A Central Limit Theorem

Lindberg-Levy vs. Lindeberg-Feller Lindeberg-Levy assumes random sampling – observations have the same mean and same variance. Lindeberg-Feller allows variances to differ across observations, with some necessary assumptions about how they vary. Most econometric estimators require Lindeberg-Feller.

Order of a Sequence Order of a sequence ‘Little oh’ o(.). Sequence hn is o(n) (order less than n) iff n- hn  0. Example: hn = n1.4 is o(n1.5) since n-1.5 hn = 1 /n.1  0. ‘Big oh’ O(.). Sequence hn is O(n) iff n- hn  a finite nonzero constant. Example 1: hn = (n2 + 2n + 1) is O(n2). Example 2: ixi2 is usually O(n1) since this is nthe mean of xi2 and the mean of xi2 generally converges to E[xi2], a finite constant. What if the sequence is a random variable? The order is in terms of the variance. Example: What is the order of the sequence in random sampling? Var[ ] = σ2/n which is O(1/n)

Asymptotic Distribution An asymptotic distribution is a finite sample approximation to the true distribution of a random variable that is good for large samples, but not necessarily for small samples. Stabilizing transformation to obtain a limiting distribution. Multiply random variable xn by some power, a, of n such that the limiting distribution of naxn has a finite, nonzero variance. Example, has a limiting variance of zero, since the variance is σ2/n. But, the variance of √n is σ2. However, this does not stabilize the distribution because E[ ] = √ nμ. The stabilizing transformation would be

Asymptotic Distribution Obtaining an asymptotic distribution from a limiting distribution Obtain the limiting distribution via a stabilizing transformation Assume the limiting distribution applies reasonably well in finite samples Invert the stabilizing transformation to obtain the asymptotic distribution Asymptotic normality of a distribution.

Asymptotic Efficiency Comparison of asymptotic variances How to compare consistent estimators? If both converge to constants, both variances go to zero. Example: Random sampling from the normal distribution, Sample mean is asymptotically normal[μ,σ2/n] Median is asymptotically normal [μ,(π/2)σ2/n] Mean is asymptotically more efficient

The Delta Method The delta method (combines most of these concepts) Nonlinear transformation of a random variable: f(xn) such that plim xn =  but n (xn - ) is asymptotically normally distributed. What is the asymptotic behavior of f(xn)? Taylor series approximation: f(xn)  f() + f() (xn - ) By Slutsky theorem, plim f(xn) = f() n[f(xn) - f()]  f() [n (xn - )] Large sample behaviors of the LHS and RHS sides are the same (generally - requires f(.) to be nicely behaved. RHS is a constant times something familiar. Large sample variance is [f()]2 times large sample Var[n (xn - )] Return to asymptotic variance of xn Gives us the VIR for the asymptotic distribution of a function.

Delta Method

Delta Method - Applications

Krinsky and Robb vs. the Delta Method

Sample Experiment-Two Functions of the Sample Mean: 1/mean and exp(mean) Proc=DltaMthd(Seed) $ Calc ; Ran(Seed) $ Sample ; 1 - 1000 $ Create ; x = Rnn(3,5) $ Calc ; List ; xbar = Xbr(x) $ Calc ; AsyVar = Var(X)/n ; AsySD = Sqr(AsyVar) $ Calc ; f1=1/xbar ; df1dx = -f1^2 ; f2 = exp(xbar) ; df2dx = f2 $ Calc ; deltaV1 = df1dx^2 * AsyVar ; deltaV2 = df2dx^2 * AsyVar $ Create ; RS = Rnn(xbar,AsySd) ; rf1 = 1/rs ; rf2 = exp(rs) $ Calc ; List ; DeltaV1 ; KRV1 = Var(RF1) ; DeltaV2 ; KRV2 = Var(RF2) $ EndProc$ Exec ; Proc=DltaMthd(542395)$ (Set seed for 1000 random draws)

Experiment Results --> Exec ; Proc=DltaMthd(542395)$ XBAR = 3.1290685989406820 DELTAV1 = .00024928790325646860 KRV1 = .00027015447289268300 DELTAV2 = 12.480615225137060 KRV2 = 14.215304185032530 --> Exec ; Proc=DltaMthd(731295)$ XBAR = 2.8973760339840560 DELTAV1 = .00035210787074436620 KRV1 = .00035308170393997490 DELTAV2 = 8.1531237478298080 KRV2 = 8.3146354826729940

Delta Method – More than One Parameter

More than One Function and More than One Coefficient

Application: CES Function

Application: CES Function

Application: CES Function Using Christensen and Greene Electricity data Create ;x1=1 ;x2=logfuel ; x3=logcap ; x4=-.5*(logfuel-logcap)^2$ Regress ;lhs=logq;rh1=x1,x2,x3,x4$ Calc ;b1=b(1);b2=b(2);b3=b(3);b4=b(4) $ Calc ;gamma=exp(b1) ; delta=b2/(b2+b3) ; nu=b2+b3 ; rho=b4*(b2+b3)/(b2*b3)$ Calc ;g11=exp(b1) ;g12=0 ;g13=0 ;g14=0 ;g21=0 ;g22=b3/(b2+b3)^2 ;g23=-b2/(b2+b3)^2 ;g24=0 ;g31=0 ;g32=1 ;g33=1 ;g34=0 ;g41=0 ;g42=-b3*b4/(b2^2*b3) ;g43=-b2*b4/(b2*b3^2) ;g44=(b2+b3)/(b2*b3)$ Matrix ; g=[g11,g12,g13,g14/g21,g22,g23,g24/g31,g32,g33,g34/g41,g42,g43,g44]$ Matrix ; VDelta=G*VARB*G' $ Matrix ; theta=[gamma/delta/nu/rho] ; Stat(theta,vdelta)$

Estimated CES Function

Asymptotics for Least Squares Looking Ahead… Assumptions: Convergence of XX/n (doesn’t require nonstochastic or stochastic X). Convergence of X’/n to 0. Sufficient for consistency. Assumptions: Convergence of (1/n)X’ to a normal vector, gives asymptotic normality What about asymptotic efficiency?