So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.

Slides:



Advertisements
Similar presentations
Multiple Regression.
Advertisements

Things to do in Lecture 1 Outline basic concepts of causality
Properties of Least Squares Regression Coefficients
Kin 304 Regression Linear Regression Least Sum of Squares
The Simple Regression Model
Linear Regression.  The following is a scatterplot of total fat versus protein for 30 items on the Burger King menu:  The model won’t be perfect, regardless.
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Regression Greg C Elvers.
Correlation and Regression
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
Chapter 8 Linear Regression © 2010 Pearson Education 1.
CHAPTER 8: LINEAR REGRESSION
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
Part 1 Cross Sectional Data
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
Lecture 3 Cameron Kaplan
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
The Simple Linear Regression Model: Specification and Estimation
Econ 140 Lecture 121 Prediction and Fit Lecture 12.
The Simple Regression Model
Business Statistics - QBM117 Least squares regression.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Basic Statistical Concepts Part II Psych 231: Research Methods in Psychology.
Chap 3-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 3 Describing Data: Numerical Statistics for Business and Economics.
Lorelei Howard and Nick Wright MfD 2008
Z Scores & Correlation Greg C Elvers.
Relationships Among Variables
Ordinary Least Squares
Chapter 8: Bivariate Regression and Correlation
Lecture 3-2 Summarizing Relationships among variables ©
Introduction to Linear Regression and Correlation Analysis
STATISTICS: BASICS Aswath Damodaran 1. 2 The role of statistics Aswath Damodaran 2  When you are given lots of data, and especially when that data is.
Relationships between Variables. Two variables are related if they move together in some way Relationship between two variables can be strong, weak or.
Chapter 4-5: Analytical Solutions to OLS
Statistics and Quantitative Analysis U4320 Segment 8 Prof. Sharyn O’Halloran.
1 Chapter 10 Correlation and Regression 10.2 Correlation 10.3 Regression.
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
2.4 Units of Measurement and Functional Form -Two important econometric issues are: 1) Changing measurement -When does scaling variables have an effect.
Objective: Understanding and using linear regression Answer the following questions: (c) If one house is larger in size than another, do you think it affects.
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Stat 13, Tue 5/29/ Drawing the reg. line. 2. Making predictions. 3. Interpreting b and r. 4. RMS residual. 5. r Residual plots. Final exam.
STA291 Statistical Methods Lecture LINEar Association o r measures “closeness” of data to the “best” line. What line is that? And best in what terms.
Correlation – Recap Correlation provides an estimate of how well change in ‘ x ’ causes change in ‘ y ’. The relationship has a magnitude (the r value)
6. Simple Regression and OLS Estimation Chapter 6 will expand on concepts introduced in Chapter 5 to cover the following: 1) Estimating parameters using.
Residuals Recall that the vertical distances from the points to the least-squares regression line are as small as possible.  Because those vertical distances.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
Essentials of Business Statistics: Communicating with Numbers By Sanjiv Jaggia and Alison Kelly Copyright © 2014 by McGraw-Hill Higher Education. All rights.
Linear Regression Chapter 8. Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Regression Analysis: A statistical procedure used to find relations among a set of variables B. Klinkenberg G
Introduction Dispersion 1 Central Tendency alone does not explain the observations fully as it does reveal the degree of spread or variability of individual.
BUSINESS MATHEMATICS & STATISTICS. Module 6 Correlation ( Lecture 28-29) Line Fitting ( Lectures 30-31) Time Series and Exponential Smoothing ( Lectures.
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
Nvmrao Three useful results: GOODNESS OF FIT This sequence explains measures of goodness of fit in regression analysis. It is convenient to start by demonstrating.
Stats Methods at IC Lecture 3: Regression.
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
The simple linear regression model and parameter estimation
Chapter 4 Basic Estimation Techniques
Regression Analysis AGEC 784.
Ch. 2: The Simple Regression Model
The Simple Regression Model
The Least-Squares Regression Line
…Don’t be afraid of others, because they are bigger than you
Ch. 2: The Simple Regression Model
No notecard for this quiz!!
Presentation transcript:

So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive meaning about the influence of the variable X on the size of the slope, since it shows that: i) the greater the covariance between X and Y, the larger the (absolute value of) the slope ii) the smaller the variance of X, the larger the (absolute value of) the slope

It is equally important to be able to interpret the effect of an estimated regression coefficient Given OLS essentially passes a straight line through the data, then given Y = b0 + b1X So the OLS estimate of the slope will give an estimate of the unit change in the dependent variable y following a unit change in the level of the explanatory variable (so you need to be aware of the units of measurement of your variables in order to be able to interpret what the OLS coefficient is telling you)

PROPERTIES OF OLS Using the fact that for any individual observation, i, the ols residual is the difference between the actual and predicted value Sub. in So that Summing over all N observations in the data set and dividing by N 6

So the mean value of the OLS residuals is zero Since the sum of any series divided by the sample size gives the mean, can write and since So the mean value of the OLS residuals is zero (as any residual should be, since random and unpredictable by definition) 6

The 2nd useful property of OLS is that the mean of the OLS predicted values equals the mean of the actual values in the data (so OLS predicts average behaviour in the data set – another useful property)

So the mean value of the OLS residuals is zero Since the sum of any series divided by the sample size gives the mean, can write and since So the mean value of the OLS residuals is zero (as any residual should be, since random and unpredictable by definition) 6

Proof: summing Dividing by N We know from above that so

Story so far…… Looked at idea behind OLS – fit a straight line thru the data that gives the “best fit” Do this by choosing a value for the intercept and the slope of the straight line that will minimise the sum of squared residuals OLS formula to do this given by

Things to do in Lecture 3 Useful algebra results underlying OLS Invent a summary measure of how well regression fits the data (R2) Look at contribution of individual variables in explaining outcomes, (Bias, Efficiency)

Useful Algebraic aspects of OLS and since So the mean value of the OLS residuals is zero (as any residual should be, since random and unpredictable by definition) 6

OLS predicts average behaviour Proof: summing Dividing by N We know from above that so OLS predicts average behaviour

The 3rd useful result is that the covariance between the fitted values of Y and the residuals must be zero. 22

GOODNESS OF FIT Now we know how to summarise the relationships in the data using the OLS method, we next need a summary measure of “how well” the estimated OLS line fits the data Think of the dispersion of all possible y values (the variation in Y) being represented by a circle And similarly the dispersion in the range of x values Y X

Little overlap in values so X not explain much of variation in Y The more the circles overlap the more the variation in the X data explains the variation in y Y Y X X Little overlap in values so X not explain much of variation in Y Large overlap in values so X variable explains much of variation in Y

Using the rules on covariances (see problem set 0) we know that To derive a statistical measure which does much the same thing remember that Using the rules on covariances (see problem set 0) we know that So the variation in the variable of interest, var(Y), is explained by either the variation in the variables included in the OLS model, or by variation in the residual 29

So we use the ratio As a measure of how well the model fits the data. (R2 is also known as the coefficient of determination) So R2 measures the % of variation in the dependent variable explained by the model. If the model explains all the variation in y then the ratio equals 1 So the closer the ratio is to one the better the fit.

It is more common however to use one further algebraic adjustment. Given (1) says that Can write this as The 1/n is common to both sides, so can cancel out and using the results that Then we have 29

This is called the Total Sum of Squares. The left side of the equation is the sum of the squared deviations of Y about its sample mean. This is called the Total Sum of Squares. The right hand side consists of the sum of squared deviations of the predictions around the sample mean (the Explained Sum of Squares) and the Residual Sum of Squares From this can have an alternative definition of the goodness of fit 29

Can see from above that it must hold that when ESS = 0, then R2 = 0 (and model explains none of the variation in the dependent variable) when ESS = TSS, then R2 = 1 (and model explains all of the variation in the dependent variable) In general the R2 lies between these two extremes. You will find that for cross-section data (ie samples of individuals, firms etc) the R2 are typically in the region of 0.2 for time-series data (ie samples of aggregate (whole economy) data measured at different points in time) the R2 are typically in the region of 0.9

Why? Cross section data typically are bigger and have a lot more variation while time series data sets are smaller and have less variation in the dependent variable (easier to fit a straight line if data don’t vary much) While we would like the R2 to be as high as possible you can only compare R2 in models with the SAME dependent (Y) variable, so can’t say one model is a better fit than another if dependent variables are different

Proof: We know the formula for the correlation coefficient GOODNESS OF FIT So the R2 measures the proportion of variance in the dependent variable explained by the model Another useful interpretation of the R2 is that it equals the square of the correlation coefficient between the actual and predicted values of Y Proof: We know the formula for the correlation coefficient Sub. In for (actual value = predicted value plus residual ) 43

Expand the covariance terms GOODNESS OF FIT Expand the covariance terms (since already proved And can always write any variance term as square root of the product 43

Thus the correlation coefficient is the square root of R2. GOODNESS OF FIT Cancelling terms so Thus the correlation coefficient is the square root of R2. Eg R2 =0.25 implies correlation coefficient between Y variable & X variable (or between Y and predicted values ) = √0.25 = 0.5 43