Part 17: Regression Residuals 17-1/38 Statistics and Data Analysis Professor William Greene Stern School of Business IOMS Department Department of Economics.

Slides:



Advertisements
Similar presentations
Managerial Economics in a Global Economy
Advertisements

Multiple Regression Analysis
Lesson 10: Linear Regression and Correlation
Section 10-3 Regression.
Kin 304 Regression Linear Regression Least Sum of Squares
Part 1: Simple Linear Model 1-1/301-1 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Econ 140 Lecture 81 Classical Regression II Lecture 8.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Simple Linear Regression. Start by exploring the data Construct a scatterplot  Does a linear relationship between variables exist?  Is the relationship.
Statistics Measures of Regression and Prediction Intervals.
 Coefficient of Determination Section 4.3 Alan Craig
Ch.6 Simple Linear Regression: Continued
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
Chapter 10 Curve Fitting and Regression Analysis
Linear regression models
Correlation and Regression
Simple Regression. Major Questions Given an economic model involving a relationship between two economic variables, how do we go about specifying the.
Ch11 Curve Fitting Dr. Deshi Ye
Correlation and regression
Sociology 601 Class 17: October 28, 2009 Review (linear regression) –new terms and concepts –assumptions –reading regression computer outputs Correlation.
Chapter 12 Simple Linear Regression
Chapter 10 Simple Regression.
9. SIMPLE LINEAR REGESSION AND CORRELATION
The Simple Regression Model
Statistical Inference and Regression Analysis: GB Professor William Greene Stern School of Business IOMS Department Department of Economics.
Chapter 11 Multiple Regression.
Simple Linear Regression Analysis
Quantitative Business Analysis for Decision Making Simple Linear Regression.
Part 19: Residuals and Outliers 19-1/27 Statistics and Data Analysis Professor William Greene Stern School of Business IOMS Department Department of Economics.
Part 18: Regression Modeling 18-1/44 Statistics and Data Analysis Professor William Greene Stern School of Business IOMS Department Department of Economics.
Part 5: Functional Form 5-1/36 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Part 7: Multiple Regression Analysis 7-1/54 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Chapter 12 Section 1 Inference for Linear Regression.
Simple Linear Regression Analysis
Part 3: Regression and Correlation 3-1/41 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Advantages of Multivariate Analysis Close resemblance to how the researcher thinks. Close resemblance to how the researcher thinks. Easy visualisation.
Measures of Regression and Prediction Intervals
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Introduction to Linear Regression and Correlation Analysis
Chapter 11 Simple Regression
Correlation and Linear Regression
1 FORECASTING Regression Analysis Aslı Sencer Graduate Program in Business Information Systems.
Measures of relationship Dr. Omar Al Jadaan. Agenda Correlation – Need – meaning, simple linear regression – analysis – prediction.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Ch4 Describing Relationships Between Variables. Section 4.1: Fitting a Line by Least Squares Often we want to fit a straight line to data. For example.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Part 2: Model and Inference 2-1/49 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
© 2001 Prentice-Hall, Inc.Chap 13-1 BA 201 Lecture 19 Measure of Variation in the Simple Linear Regression Model (Data)Data.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
The simple linear regression model and parameter estimation
Chapter 20 Linear and Multiple Regression
CHAPTER 3 Describing Relationships
Ch12.1 Simple Linear Regression
Relationship with one independent variable
Quantitative Methods Simple Regression.
Correlation and Simple Linear Regression
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
Correlation and Simple Linear Regression
Statistical Inference and Regression Analysis: GB
Relationship with one independent variable
Simple Linear Regression and Correlation
Econometrics I Professor William Greene Stern School of Business
Introduction to Regression
Presentation transcript:

Part 17: Regression Residuals 17-1/38 Statistics and Data Analysis Professor William Greene Stern School of Business IOMS Department Department of Economics

Part 17: Regression Residuals 17-2/38 Statistics and Data Analysis Part 17 – The Linear Regression Model

Part 17: Regression Residuals 17-3/38 Regression Modeling  Theory behind the regression model  Computing the regression statistics  Interpreting the results  Application: Statistical Cost Analysis

Part 17: Regression Residuals 17-4/38 A Linear Regression Predictor: Box Office = Buzz

Part 17: Regression Residuals 17-5/38 Data and Relationship  We suggested the relationship between box office sales and internet buzz is Box Office = Buzz  Box Office is not exactly equal to xBuzz  How do we reconcile the equation with the data?

Part 17: Regression Residuals 17-6/38 Modeling the Underlying Process  A model that explains the process that produces the data that we observe: Observed outcome = the sum of two parts (1) Explained: The regression line (2) Unexplained (noise): The remainder. Internet Buzz is not the only thing that explains Box Office, but it is the only variable in the equation.  Regression model The “model” is the statement that part (1) is the same process from one observation to the next.

Part 17: Regression Residuals 17-7/38 The Population Regression  THE model: (1) Explained: Explained Box Office = α + β Buzz (2) Unexplained: The rest is “noise, ε.” Random ε has certain characteristics  Model statement Box Office = α + β Buzz + ε Box Office is related to Buzz, but is not exactly equal to α + β Buzz

Part 17: Regression Residuals 17-8/38 The Data Include the Noise

Part 17: Regression Residuals 17-9/38 What explains the noise? What explains the variation in fuel bills?

Part 17: Regression Residuals 17-10/38 Noisy Data? What explains the variation in milk production other than number of cows?

Part 17: Regression Residuals 17-11/38 Assumptions  (Regression) The equation linking “Box Office” and “Buzz” is stable E[Box Office | Buzz] = α + β Buzz  Another sample of movies, say 2012, would obey the same fundamental relationship.

Part 17: Regression Residuals 17-12/38 Model Assumptions  y i = α + β x i + ε i α + β x i is the “regression function” ε i is the “disturbance. It is the unobserved random component  The Disturbance is Random Noise Mean zero. The regression is the mean of y i. ε i is the deviation from the regression. Variance σ 2.

Part 17: Regression Residuals 17-13/38 We will use the data to estimate  and β

Part 17: Regression Residuals 17-14/38 We also want to estimate  2 =√E[ε i 2 ] e=y-a-bBuzz

Part 17: Regression Residuals 17-15/38 Standard Deviation of the Residuals  Standard deviation of ε i = y i -α-βx i is σ  σ = √E[ε i 2 ] (Mean of ε i is zero)  Sample a and b estimate α and β  Residual e i = y i – a – bx i estimates ε i  Use √(1/N-2)Σe i 2 to estimate σ. Why N-2? Relates to the fact that two parameters (α,β) were estimated. Same reason N-1 was used to compute a sample variance.

Part 17: Regression Residuals 17-16/38 Residuals

Part 17: Regression Residuals 17-17/38 Summary: Regression Computations

Part 17: Regression Residuals 17-18/38 Using s e to identify outliers Remember the empirical rule, 95% of observations will lie within mean ± 2 standard deviations? We show (a+bx) ± 2s e below.) This point is 2.2 standard deviations from the regression. Only 3.2% of the 62 observations lie outside the bounds. (We will refine this later.)

Part 17: Regression Residuals 17-19/38

Part 17: Regression Residuals 17-20/38 Linear Regression Sample Regression Line

Part 17: Regression Residuals 17-21/38

Part 17: Regression Residuals 17-22/38

Part 17: Regression Residuals 17-23/38 Results to Report

Part 17: Regression Residuals 17-24/38 The Reported Results

Part 17: Regression Residuals 17-25/38 Estimated equation

Part 17: Regression Residuals 17-26/38 Estimated coefficients a and b

Part 17: Regression Residuals 17-27/38 S = s e = estimated std. deviation of ε

Part 17: Regression Residuals 17-28/38 Square of the sample correlation between x and y

Part 17: Regression Residuals 17-29/38 N-2 = degrees of freedom N-1 = sample size minus 1

Part 17: Regression Residuals 17-30/38 Sum of squared residuals, Σ i e i 2

Part 17: Regression Residuals 17-31/38 S 2 = s e 2

Part 17: Regression Residuals 17-32/38

Part 17: Regression Residuals 17-33/38

Part 17: Regression Residuals 17-34/38 The Model  Constructed to provide a framework for interpreting the observed data What is the meaning of the observed relationship (assuming there is one)  How it’s used Prediction: What reason is there to assume that we can use sample observations to predict outcomes? Testing relationships

Part 17: Regression Residuals 17-35/38 A Cost Model Electricity.mpj Total cost in $Million Output in Million KWH N = 123 American electric utilities Model: Cost = α + βKWH + ε

Part 17: Regression Residuals 17-36/38 Cost Relationship

Part 17: Regression Residuals 17-37/38 Sample Regression

Part 17: Regression Residuals 17-38/38 Interpreting the Model  Cost = Output + e  Cost is $Million, Output is Million KWH.  Fixed Cost = Cost when output = 0 Fixed Cost = $2.44Million  Marginal cost = Change in cost/change in output = * $Million/Million KWH = $/KWH = cents/KWH.

Part 17: Regression Residuals 17-39/38 Summary  Linear regression model Assumptions of the model Residuals and disturbances  Estimating the parameters of the model Regression parameters Disturbance standard deviation  Computation of the estimated model