Lecture 3 Cameron Kaplan

Slides:



Advertisements
Similar presentations
Things to do in Lecture 1 Outline basic concepts of causality
Advertisements

Multiple Regression Analysis
CHAPTER 3: TWO VARIABLE REGRESSION MODEL: THE PROBLEM OF ESTIMATION
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
Introduction: The General Linear Model b b The General Linear Model is a phrase used to indicate a class of statistical models which include simple linear.
Definition  Regression Model  Regression Equation Y i =  0 +  1 X i ^ Given a collection of paired data, the regression equation algebraically describes.
Lecture 9 Today: Ch. 3: Multiple Regression Analysis Example with two independent variables Frisch-Waugh-Lovell theorem.
(c) 2007 IUPUI SPEA K300 (4392) Outline Least Squares Methods Estimation: Least Squares Interpretation of estimators Properties of OLS estimators Variance.
Bivariate Regression Analysis
Lecture 4 This week’s reading: Ch. 1 Today:
Simple Linear Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
1 Regression Analysis Regression used to estimate relationship between dependent variable (Y) and one or more independent variables (X). Consider the variable.
7/2/ Lecture 51 STATS 330: Lecture 5. 7/2/ Lecture 52 Tutorials  These will cover computing details  Held in basement floor tutorial lab,
Lecture 2 (Ch3) Multiple linear regression
Ch. 14: The Multiple Regression Model building
Business Statistics - QBM117 Least squares regression.
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Introduction to Regression Analysis, Chapter 13,
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
3. Multiple Regression Analysis: Estimation -Although bivariate linear regressions are sometimes useful, they are often unrealistic -SLR.4, that all factors.
Introduction to Linear Regression and Correlation Analysis
Linear Trend Lines = b 0 + b 1 X t Where is the dependent variable being forecasted X t is the independent variable being used to explain Y. In Linear.
7.1 Multiple Regression More than one explanatory/independent variable This makes a slight change to the interpretation of the coefficients This changes.
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
MULTIPLE REGRESSION WITH TWO EXPLANATORY VARIABLES: EXAMPLE 1 This sequence provides a geometrical interpretation of a multiple regression model with two.
Regression Maarten Buis Outline Recap Estimation Goodness of Fit Goodness of Fit versus Effect Size transformation of variables and effect.
Lecturer: Kem Reat, Viseth, PhD (Economics)
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Statistical Methods Statistical Methods Descriptive Inferential
Regression. Population Covariance and Correlation.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Six.
Warsaw Summer School 2015, OSU Study Abroad Program Regression.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Ordinary Least Squares Regression.
10B11PD311 Economics REGRESSION ANALYSIS. 10B11PD311 Economics Regression Techniques and Demand Estimation Some important questions before a firm are.
Chapter 5 Demand Estimation Managerial Economics: Economic Tools for Today’s Decision Makers, 4/e By Paul Keat and Philip Young.
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Four Ending Wednesday, September 19 (Assignment 4 which is included in this study guide.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin The Two-Variable Model: Hypothesis Testing chapter seven.
STA291 Statistical Methods Lecture LINEar Association o r measures “closeness” of data to the “best” line. What line is that? And best in what terms.
Chapter 2 Ordinary Least Squares Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Environmental Modeling Basic Testing Methods - Statistics III.
9.2 Linear Regression Key Concepts: –Residuals –Least Squares Criterion –Regression Line –Using a Regression Equation to Make Predictions.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 3 Association: Contingency, Correlation, and Regression Section 3.3 Predicting the Outcome.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Lecture 8: Ordinary Least Squares Estimation BUEC 333 Summer 2009 Simon Woodcock.
Regression Analysis: A statistical procedure used to find relations among a set of variables B. Klinkenberg G
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Multiple Independent Variables POLS 300 Butz. Multivariate Analysis Problem with bivariate analysis in nonexperimental designs: –Spuriousness and Causality.
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
Describing Bivariate Relationships. Bivariate Relationships When exploring/describing a bivariate (x,y) relationship: Determine the Explanatory and Response.
Linear Regression 1 Sociology 5811 Lecture 19 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
The simple linear regression model and parameter estimation
Chapter 14 Introduction to Multiple Regression
Part 5 - Chapter
Part 5 - Chapter 17.
Reasoning in Psychology Using Statistics
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
Part 5 - Chapter 17.
Linear regression Fitting a straight line to observations.
Goodness of Fit The sum of squared deviations from the mean of a variable can be decomposed as follows: TSS = ESS + RSS This decomposition can be used.
Simple Linear Regression
Least-Squares Regression
A medical researcher wishes to determine how the dosage (in mg) of a drug affects the heart rate of the patient. Find the correlation coefficient & interpret.
Presentation transcript:

Lecture 3 Cameron Kaplan Econ 488 Lecture 3 Cameron Kaplan

Announcements Midterm Date Change: Now October 22 Syllabus will be updated soon. Library Session: October 8

How Regression Works Estimate the slope of the line that passes through the origin given the following data

How Regression Works Estimate the slope of the line that passes through the origin given the following data

How Regression Works Try this one… Estimate the slope of the line that passes through the origin given the following data

How Regression Works Try this one… Estimate the slope of the line that passes through the origin given the following data

Answers 1. (3,1): Slope = 1/3 2. (4,2): Slope = ½ How did you get that? Slope = Y/X

How Regression Works Now suppose we have 2 values: Estimate the slope of the line that passes through the origin given the following data

Possible Estimators 1. Average of the two slopes.

Possible Estimators 2. Midpoint Estimator

Possible Estimators 2. Midpoint Estimator: Or = (1.5)/(3.5) = 3/7 Or = (1+2)/(3+4) = 3/7

Possible Estimators 3. Ordinary Least Squares (OLS) We want a line that is as close as possible to all of the points

Ordinary Least Squares We want to find a line that makes these residuals, e1and e2 as small as possible. e2 e1

Ordinary Least Squares Equation of the line: (pronounced: “y i hat is equal to beta-hat x i”) The underlying data generating process is: (notice there are no hats) Finally, the observed values of X and Y can be described by: e is the “residual”, which is actually observed.  is the “stochastic error term”, which is never observed

Ordinary Least Squares By equations (1) and (3), we can see that: So, we want to choose a line so that ei is as small as possible. But, ei can be negative or positive, so we can’t just minimize ei.

Ordinary Least Squares We could choose a that minimizes the absolute value of ei. That is, This is what is called the “Least Absolute Deviations” method. However, this is mathematically difficult, and there is another way that is better: Minimize ei2! ei2 is always positive, so we can minimize it.

Ordinary Least Squares Choose a that minimizes Remember, We want to minimize the sum of this. This is equivalent to:

Ordinary Least Squares Using calculus, the first order condition (FOC) for a minimum is that the first derivative is equal to zero. Take derivative with respect to Solve for :

Our Example What is the OLS slope estimate for our example? Y1=1, X1=3 So,

Possible Estimators Now we have 3 estimators: 1. Average of slopes to each point: OR =5/120.4167

Possible Estimator 2. Midpoint: = 3/7 0.4286 3. Ordinary Least Squares = 11/25 0.4444

Possible Estimator Which estimator is best? Let’s try an exercise.

OLS with an intercept term

Example Height and Shoe Size

Sum of Squares How much of the variation in the dependent variable is explained by the estimated regression equation? Total Sum of Squares (TSS) – How spread out are the y values in the sample? Explained Sum of Squares (ESS) – The sample variation in

Sum of Squares Residual Sum of Squares (RSS) – The sample variation in ei TSS= ESS+RSS Some of the variation in y can be explained by the regression, and some cannot If the RSS is small relative to the TSS, the equation is a good fit.

R-squared R-squared (or R2) is the proportion of the variation in Y that is explained by the regression. 0 ≤ R2 ≤ 1

R-squared

R-squared

R-squared

R-Squared

R-Squared

Multiple Regression Each coefficient is a partial regression coefficient β2 is the change in Y associated with a one unit increase in X2, holding the other X’s (i.e. X1, X3, X4, etc.) constant.

Multiple Regression Example Suppose we run this regression, and get: This means that, on average, a one year increase in education is associated with a $0.599 per hour increase in wages, holding experience and tenure constant.

Degrees of Freedom How many more observations do you have to have above the number of coefficients you are trying to estimate? Can you estimate the slope and intercept given just one point? You always need at least as many observations as the number of coefficients you are estimating. But having more is better. Extra observations are extra degrees of freedom. Degrees of Freedom = n-k-1

R-squared vs. Adjusted R-squared Whenever you add an extra variable, R2 will go up. Why? The extra variable will add at least some explanatory power to the regression. However, by adding another variable, you have an additional coefficient to estimate. Degrees of Freedom go down. So there is a benefit of adding an extra variable (R2 goes up) and a cost (d.f go down). Adjusted R2 adjusts the R2 to account for the loss in degrees of freedom.

Adjusted R-square Note that it is possible to get a negative adjusted R-squared