Violations of Assumptions In Least Squares Regression

Slides:



Advertisements
Similar presentations
The Simple Linear Regression Model Specification and Estimation Hill et al Chs 3 and 4.
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
The Simple Regression Model
FACTORIAL ANOVA Overview of Factorial ANOVA Factorial Designs Types of Effects Assumptions Analyzing the Variance Regression Equation Fixed and Random.
Simple Linear Regression
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
9. SIMPLE LINEAR REGESSION AND CORRELATION
CHAPTER 3 ECONOMETRICS x x x x x Chapter 2: Estimating the parameters of a linear regression model. Y i = b 1 + b 2 X i + e i Using OLS Chapter 3: Testing.
Confidence intervals. Population mean Assumption: sample from normal distribution.
1 BA 275 Quantitative Business Methods Residual Analysis Multiple Linear Regression Adjusted R-squared Prediction Dummy Variables Agenda.
REGRESSION What is Regression? What is the Regression Equation? What is the Least-Squares Solution? How is Regression Based on Correlation? What are the.
REGRESSION MODEL ASSUMPTIONS. The Regression Model We have hypothesized that: y =  0 +  1 x +  | | + | | So far we focused on the regression part –
Chapter Topics Types of Regression Models
Regression Diagnostics - I
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
Simple Linear Regression Analysis
GRA 6020 Multivariate Statistics Regression examples Ulf H. Olsson Professor of Statistics.
1.The independent variables do not form a linearly dependent set--i.e. the explanatory variables are not perfectly correlated. 2.Homoscedasticity --the.
Violations of Assumptions In Least Squares Regression.
Quantitative Business Analysis for Decision Making Simple Linear Regression.
REGRESSION Predict future scores on Y based on measured scores on X Predictions are based on a correlation from a sample where both X and Y were measured.
Business Statistics - QBM117 Statistical inference for regression.
Basics of Sampling Theory P = { x 1, x 2, ……, x N } where P = population x 1, x 2, ……, x N are real numbers Assuming x is a random variable; Mean/Average.
Standard error of estimate & Confidence interval.
1 FORECASTING Regression Analysis Aslı Sencer Graduate Program in Business Information Systems.
Error Component Models Methods of Economic Investigation Lecture 8 1.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
1Spring 02 Problems in Regression Analysis Heteroscedasticity Violation of the constancy of the variance of the errors. Cross-sectional data Serial Correlation.
6.3 THE CENTRAL LIMIT THEOREM. DISTRIBUTION OF SAMPLE MEANS  A sampling distribution of sample means is a distribution using the means computed from.
ANOVA Assumptions 1.Normality (sampling distribution of the mean) 2.Homogeneity of Variance 3.Independence of Observations - reason for random assignment.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin The Two-Variable Model: Hypothesis Testing chapter seven.
1 G Lect 2w Review of expectations Conditional distributions Regression line Marginal and conditional distributions G Multiple Regression.
ECON 338/ENVR 305 CLICKER QUESTIONS Statistics – Question Set #8 (from Chapter 10)
The Simple Linear Regression Model: Specification and Estimation ECON 4550 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
1 Regression Review Population Vs. Sample Regression Line Residual and Standard Error of Regression Interpretation of intercept & slope T-test, F-test.
Economics 20 - Prof. Anderson1 Time Series Data y t =  0 +  1 x t  k x tk + u t 1. Basic Analysis.
Homogeneity of Variance Pooling the variances doesn’t make sense when we cannot assume all of the sample Variances are estimating the same value. For two.
Week 21 Statistical Assumptions for SLR  Recall, the simple linear regression model is Y i = β 0 + β 1 X i + ε i where i = 1, …, n.  The assumptions.
Which Class Did Better? Group 12A RESULTS(%) 0< x  2020< x  3030< x  5050< x  7070< x  100 FREQUENCY Group 12B RESULTS(%) 0< x  2020< x  3030
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
Regression Overview. Definition The simple linear regression model is given by the linear equation where is the y-intercept for the population data, is.
Inference about the slope parameter and correlation
The simple linear regression model and parameter estimation
Statistical Modelling
B&A ; and REGRESSION - ANCOVA B&A ; and
Kin 304 Regression Linear Regression Least Sum of Squares
ECONOMETRICS DR. DEEPTI.
Chapter 3: TWO-VARIABLE REGRESSION MODEL: The problem of Estimation
BPK 304W Regression Linear Regression Least Sum of Squares
BIVARIATE REGRESSION AND CORRELATION
Simple Linear Regression - Introduction
Slope of the regression line:
BA 275 Quantitative Business Methods
The regression model in matrix form
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Statistical Assumptions for SLR
Homogeneity of Variance
Chapter 4, Regression Diagnostics Detection of Model Violation
Simple Linear Regression
Simple Linear Regression
Regression Assumptions
Chapter 13 Additional Topics in Regression Analysis
Linear Regression Summer School IFPRI
Regression and Correlation of Data
Notes: Sample Means
Ch3 The Two-Variable Regression Model
Violations of Assumptions In Least Squares Regression
Regression Assumptions
Presentation transcript:

Violations of Assumptions In Least Squares Regression

Standard Assumptions in Regression Errors are Normally Distributed with mean 0 Errors have constant variance Errors are independent X is Measured without error

Example Xs and OLS Estimators “t” is used to imply time ordering

Non-Normality Errors (t3 errors)

Errors = 3*t3 E(et) = E(3*t3)=3E(t3)=3*0=0 V(et) = V(3*t3)=9V(t3)=9*(3/(3-2))=27 Based on 100,000 simulations, the 95% CI for b1 contained 10 in 95.31% of the samples. Average=9.999271, SD=0.3478 Average s2(b1) = 0.1216903

Non-Constant Error Variance (Heteroscedasticity) Mean: E(Y) = b0 + b1 X = 50 + 10X Standard Deviation: sX = 0.1(X+5) Distribution: Normal

Non-Constant Error Variance (Heteroscedasticity) Based on 100,000 simulations, the 95% CI for b1 contained 10 in 92.62% of the samples. Average=9.999724, SD=0.4667337 Average s2(b1) = 0.1813961 < 0.2184

Correlated Errors Example: s2= 9, r=0.5, n=22

Correlated Errors Based on 100,000 simulations, the 95% CI for b1 contained 10 in 84.46% of the samples. Average=9.99853, SD=0.3037228 Average s2(b1) = 0.0474835 < 0.092157

Measurement Error in X Z=True Value of Independent Variable (Unobserved) X=Observed Value of Independent Variable X=Z+U (Z can be fixed or random) Z, U independent (assumed) when Z is random V(U) is independent of Z when Z is fixed

Measurement Error in X Based on 100,000 simulations, the 95% CI for b1 contained 10 in 76.56% of the samples. Average=9.197172, SD=0.6112944 Average s2(b1) = 0.4290028 >> (0.20226)2