Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-1 Basic Mathematical tools Today, we will review some basic mathematical tools. Then we.

Slides:



Advertisements
Similar presentations
Chapter 3 Properties of Random Variables
Advertisements

The Simple Regression Model
The Multiple Regression Model.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 18 Sampling Distribution Models.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
Introduction and Overview
Objectives (BPS chapter 24)
Regression with One Explanator
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 6: Interpreting Regression Results Logarithms (Chapter 4.5) Standard Errors (Chapter.
1 A REVIEW OF QUME 232  The Statistical Analysis of Economic (and related) Data.
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 4: Mathematical Tools for Econometrics Statistical Appendix (Chapter 3.1–3.2)
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 3: Monte Carlo Simulations (Chapter 2.8–2.10)
The Simple Linear Regression Model: Specification and Estimation
Suggestions for Best Individual Performance
Copyright 2004 David J. Lilja1 What Do All of These Means Mean? Indices of central tendency Sample mean Median Mode Other means Arithmetic Harmonic Geometric.
Chapter 10 Simple Regression.
Copyright (c) Bani K. Mallick1 STAT 651 Lecture #18.
Evaluating Hypotheses
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 5: Regression with One Explanator (Chapter 3.1–3.5, 3.7 Chapter 4.1–4.4)
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Created by Tom Wegleitner, Centreville, Virginia Section 3-1.
Chapter 11 Multiple Regression.
The Simple Regression Model
Business and Economics 7th Edition
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
1 Psych 5500/6500 Statistics and Parameters Fall, 2008.
Correlation and Regression
Today: Central Tendency & Dispersion
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Section 10-3 Regression.
1 Tendencia central y dispersión de una distribución.
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Sociology 5811: Lecture 7: Samples, Populations, The Sampling Distribution Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
Correlation.
Chapter 3 – Descriptive Statistics
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Created by Tom Wegleitner, Centreville, Virginia Section 3-1 Review and.
Copyright © 2010 Pearson Education, Inc. Slide
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 16 Random Variables.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 3 Descriptive Statistics: Numerical Methods.
Topic 5 Statistical inference: point and interval estimate
Probabilistic and Statistical Techniques 1 Lecture 24 Eng. Ismail Zakaria El Daour 2010.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Lecture 3 A Brief Review of Some Important Statistical Concepts.
What is the MPC?. Learning Objectives 1.Use linear regression to establish the relationship between two variables 2.Show that the line is the line of.
Measures of Variability In addition to knowing where the center of the distribution is, it is often helpful to know the degree to which individual values.
Chapter 3 Descriptive Statistics: Numerical Methods Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Describing Behavior Chapter 4. Data Analysis Two basic types  Descriptive Summarizes and describes the nature and properties of the data  Inferential.
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved.
Copyright © 2009 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Statistical analysis Outline that error bars are a graphical representation of the variability of data. The knowledge that any individual measurement.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
Simple regression model: Y =  1 +  2 X + u 1 We have seen that the regression coefficients b 1 and b 2 are random variables. They provide point estimates.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 18 Sampling Distribution Models.
Lecture 7: What is Regression Analysis? BUEC 333 Summer 2009 Simon Woodcock.
Chapter 3 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 Chapter 3: Measures of Central Tendency and Variability Imagine that a researcher.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Section 7-1 Review and Preview.
Central Tendency & Dispersion
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 12 Analyzing the Association Between Quantitative Variables: Regression Analysis Section.
Agresti/Franklin Statistics, 1 of 88 Chapter 11 Analyzing Association Between Quantitative Variables: Regression Analysis Learn…. To use regression analysis.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Describing Samples Based on Chapter 3 of Gotelli & Ellison (2004) and Chapter 4 of D. Heath (1995). An Introduction to Experimental Design and Statistics.
Descriptive Statistics for one variable. Statistics has two major chapters: Descriptive Statistics Inferential statistics.
STATISTICS People sometimes use statistics to describe the results of an experiment or an investigation. This process is referred to as data analysis or.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
The Simple Regression Model
Simple Linear Regression - Introduction
Regression Models - Introduction
Presentation transcript:

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-1 Basic Mathematical tools Today, we will review some basic mathematical tools. Then we will use these tools to derive general properties of estimators. We can also use these tools to derive the “best” estimators.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-2 Summations The  symbol is a shorthand notation for discussing sums of numbers. It works just like the + sign you learned about in elementary school.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-3 Algebra of Summations

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-4 Summations: A Useful Trick

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-5 Double Summations The “Secret” to Double Summations: keep a close eye on the subscripts.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-6 Descriptive Statistics How can we summarize a collection of numbers? – Mean: the arithmetic average. The mean is highly sensitive to a few large values (outliers). – Median: the midpoint of the data. The median is the number above which lie half the observed numbers and below which lie the other half. The median is not sensitive to outliers.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-7 Descriptive Statistics (cont.) – Mode: the most frequently occurring value. – Variance: the mean squared deviation of a number from its own mean. The variance is a measure of the “spread” of the data. – Standard deviation: the square root of the variance. The standard deviation provides a measure of a typical deviation from the mean.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-8 Descriptive Statistics (cont.) – Covariance: the covariance of two sets of numbers, X and Y, measures how much the two sets tend to “move together.” If Cov(X,Y)  0, then if X is above its mean, we would expect that Y would also be above its mean.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved. 4-9 Descriptive Statistics (cont.) – Correlation Coefficient: the correlation coefficient between X and Y “norms” the covariance by the standard deviations of X and Y. You can think of this adjustment as a unit correction. The correlation coefficient will always fall between -1 and 1.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved A Quick Example

Copyright © 2006 Pearson Addison-Wesley. All rights reserved A Quick Example (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved A Quick Example (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Populations and Samples Two uses for statistics: – Describe a set of numbers – Draw inferences from a set of numbers we observe to a larger population The population is the underlying structure which we wish to study. Surveyors might want to relate 6000 randomly selected voters to all the voters in the United States. Macroeconomists might want to relate data about unemployment and inflation from 1958–2004 to the underlying process linking unemployment and inflation, to predict future realizations.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Populations and Samples (cont.) We cannot observe the entire population. Instead, we observe a sample drawn from the population of interest. In the Monte Carlo demonstration from last time, an individual dataset was the sample and the Data Generating Process described the population.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Populations and Samples (cont.) The descriptive statistics we use to describe data can also describe populations. What is the mean income in the United States? What is the variance of mortality rates across countries? What is the covariance between gender and income?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Populations and Samples (cont.) In a sample, we know exactly the mean, variance, covariance, etc. We can calculate the sample statistics directly. We must infer the statistics for the underlying population. Means in populations are also called expectations.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Populations and Samples (cont.) If the true mean income in the United States is , then we expect a simple random sample to have sample mean . In practice, any given sample will also include some “sampling noise.” We will observe not , but  + . If we have drawn our sample correctly, then on average the sampling error over many samples will be 0. We write this as E (  ) = 0

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectations Expectations are means over all possible samples (think “super” Monte Carlo). Means are sums. Therefore, expectations follow the same algebraic rules as sums. See the Statistics Appendix for a formal definition of Expectations.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Algebra of Expectations k is a constant. E(k) = k E(kY) = kE(Y) E(k+Y) = k + E(Y) E(Y+X) = E(Y) + E(X) E(  Y i ) =  E(Y i ), where each Y i is a random variable.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Law of Iterated Expectations The expected value of the expected value of Y conditional on X is the expected value of Y. If we take expectations separately for each subpopulation (each value of X ), and then take the expectation of this expectation, we get back the expectation for the whole population.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Variances Population variances are also expectations.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Algebra of Variances One value of independent observations is that Cov ( Y i,Y j ) = 0, killing all the cross-terms in the variance of the sum.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding Consider the Data Generating Process (DGP): Y i =  +  i E(  i ) = 0 Var(  i ) =   2 Question: What are: E(Y i ) ? Var(Y i ) ?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding (cont.) Y i =  +  i E(  i ) = 0 Var(  i ) =   2 Note:  is a constant (we don’t observe it, but it’s still a constant). E(Y i ) = E(  +  i ) = E(  +  i ) =  +0 =  Var(Y i ) = Var (  +  i ) = Var(  i ) =   2 We will make (very) frequent use of this sort of calculation.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Algebra of Covariances Population Covariances are also expectations.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Algebra of Covariances (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Data Generating Processes In this class, we are trying to draw inferences from a sample (our data) back to an underlying process. We begin by making very concrete assumptions about the underlying process. These assumptions are called the “Data Generating Process.”

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Data Generating Process (cont.) The Data Generating Process is our model of how the world really works. The DGP includes some parameter (usually called something like  ) that we would like to know. We don’t usually want to describe the data, we want to make predictions about what will happen next.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Data Generating Process (cont.) What does the DGP assume? – The relation between Y, X, and  – The mean, variance, and covariances of  – Cross-sample properties of X

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Our Baseline DGP: Gauss–Markov Our benchmark DGP: Gauss–Markov X ’s fixed across samples We will refer to this DGP (very) frequently.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Data Generating Process Are the Gauss–Markov Assumptions true? “Models are to be used, not believed.” Where does the DGP come from? – Economic intuition – Inspection of the data – Convenience Later in the class, we will learn more complicated DGP’s. Every DGP involves dramatic simplification of reality.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Data Generating Process (cont.) It is very appropriate to be skeptical of the Gauss–Markov Assumptions. We will relax these assumptions as the class goes on. You want to be aware of the limitations of your assumptions so you can: – Fix them – Qualify your analysis when you can’t fix them

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Data Generating Process (cont.) Making precise assumptions lets you identify precisely what aspects of the model are responsible for what components of your results. Making precise assumptions provides a clear basis for arguing about the limits of your analysis.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved A Strategy for Inference The DGP tells us the assumed relationships between the data we observe and the underlying process of interest. Using the assumptions of the DGP and the algebra of expectations, variances, and covariances, we can derive key properties of our estimators, and search for estimators with desirable properties.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved An Example:  g 1

Copyright © 2006 Pearson Addison-Wesley. All rights reserved An Example:  g 1 (cont.)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Linear Estimators (Chapter 3.1)  g 1 is unbiased. Can we generalize? We will focus on linear estimators Linear estimator: a weighted sum of the Y ’s Note: we will use hats to denote estimated quantities

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Linear Estimators (cont.) Linear estimator: Example:  g 1 is a linear estimator.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding Linear Estimator: Question: Can you find weights w i to write  g 2 and  g 4 as linear estimators?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Checking Understanding (cont.) All of our “best guesses” are linear estimators!

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectation of Linear Estimators (Chapter 3.2)

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectation of Linear Estimator (cont.)

A linear estimator is unbiased if  w i X i = 1 Are  g 2,  g 3, and  g 4 unbiased?

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Expectation of Linear Estimator (cont.) Similar calculations hold for  g 3. All 4 of our “best guesses” are unbiased.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review In a sample, we know exactly the mean, variance, covariance, etc. We can calculate the sample statistics directly. We must infer the statistics for the underlying population. Means in populations are also called expectations.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review (cont.) Algebra of Expectations  k is a constant.  E(k) = k  E(kY) = kE(Y)  E(k+Y) = k + E(Y)  E(Y+X) = E(Y) + E(X)  E(  Y i ) =  E(Y i ), where each Y i is a random variable.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Review (cont.) Algebra of Variances

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Our Baseline DGP: Gauss–Markov Our benchmark DGP: Gauss–Markov Y =  X +  E(  i ) = 0 Var(  i ) =   2 Cov(  i,  j ) = 0, for i ≠ j X ’s fixed across samples We will refer to this DGP (very) frequently.

Copyright © 2006 Pearson Addison-Wesley. All rights reserved Linear Estimators We will focus on linear estimators. Linear estimator: a weighted sum of the Y ’s

Expectation of Linear Estimators