1 BUSI 6220 By Dr. Nick Evangelopoulos, © 2012 Brief overview of Linear Regression Models (Pre-MBA level)

Slides:



Advertisements
Similar presentations
Correlation and regression
Advertisements

Inference for Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Statistics for Managers Using Microsoft® Excel 5th Edition
Statistics for Managers Using Microsoft® Excel 5th Edition
Multiple Linear Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
Lecture 6: Multiple Regression
Chapter 11 Multiple Regression.
Multiple Linear Regression
Ch. 14: The Multiple Regression Model building
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Chapter 15: Model Building
Regression Model Building Setting: Possibly a large set of predictor variables (including interactions). Goal: Fit a parsimonious model that explains variation.
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
Lecture 5 Correlation and Regression
Correlation & Regression
Quantitative Business Analysis for Decision Making Multiple Linear RegressionAnalysis.
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Simple Linear Regression
Chapter 12 Multiple Regression and Model Building.
Lecture 12 Model Building BMTRY 701 Biostatistical Methods II.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Univariate Linear Regression Problem Model: Y=  0 +  1 X+  Test: H 0 : β 1 =0. Alternative: H 1 : β 1 >0. The distribution of Y is normal under both.
Maths Study Centre CB Open 11am – 5pm Semester Weekdays
©2006 Thomson/South-Western 1 Chapter 14 – Multiple Linear Regression Slides prepared by Jeff Heyl Lincoln University ©2006 Thomson/South-Western Concise.
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 2: Review of Multiple Regression (Ch. 4-5)
Simple Linear Regression (SLR)
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
Multiple Regression  Similar to simple regression, but with more than one independent variable R 2 has same interpretation R 2 has same interpretation.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Exam 2 Review. Data referenced throughout review An Educational Testing Service (ETS) research scientist used multiple regression analysis to model y,
More on regression Petter Mostad More on indicator variables If an independent variable is an indicator variable, cases where it is 1 will.
Maths Study Centre CB Open 11am – 5pm Semester Weekdays
Biostatistics Regression and Correlation Methods Class #10 April 4, 2000.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Stats Methods at IC Lecture 3: Regression.
Chapter 15 Multiple Regression Model Building
The simple linear regression model and parameter estimation
BINARY LOGISTIC REGRESSION
Chapter 15 Multiple Regression and Model Building
Statistical Modelling
Logistic Regression APKC – STATS AFAC (2016).
Correlation, Bivariate Regression, and Multiple Regression
Chapter 9 Multiple Linear Regression
Multiple Regression and Model Building
Multiple Regression.
Business Statistics, 4e by Ken Black
Chapter 15 – Multiple Linear Regression
Multiple Regression Models
DRQ 8 Dr. Capps AGEC points
Introduction to Logistic Regression
Multiple Regression Chapter 14.
Adequacy of Linear Regression Models
Regression III.
Adequacy of Linear Regression Models
Chapter 13 Additional Topics in Regression Analysis
Adequacy of Linear Regression Models
Adequacy of Linear Regression Models
Business Statistics, 4e by Ken Black
Presentation transcript:

1 BUSI 6220 By Dr. Nick Evangelopoulos, © 2012 Brief overview of Linear Regression Models (Pre-MBA level)

2 BUSI 6220 Simple Linear Regression Y = b 0 + b 1 X + residual Model with one predictor variable: X Y The fitted line is a straight line, since the model is of 1 st order: Ŷ = b 0 + b 1 X explained part of Y unexplained part of Y (error)

3 BUSI 6220 Quadratic Regression Y = b 0 + b 1 X + b 2 X 2 Quadratic Regression model: X Y

4 BUSI 6220 Polynomial Regression Y = b 0 + b 1 X + b 2 X 2 + b 3 X 3 3 rd -order Regression model: X Y

5 BUSI 6220 Indicator (Dummy) Variables Dummy, or indicator, variables allow for the inclusion of qualitative variables in the model For example: I1 =I1 =I1 =I1 = 1if female 0if male

6 BUSI 6220 Indicator (Dummy) Variables Y = b 0 + b 1 X + b 2 I Model with Indicator variable: Rewrite the model as: For I = 0, For I = 1, Y = b 0 + b 1 X Y = (b 0 + b 2 ) + b 1 X X Y

7 BUSI 6220 Indicator Variables with interaction Y = b 0 + b 1 X + b 2 I + b 3 XI Rewrite the model as: For I = 0, For I = 1, Y = b 0 + b 1 X Y = (b 0 + b 2 ) + (b 1 + b 3 ) X Model with Indicator variable: X Y

8 BUSI 6220 Hypothesis Test on the Slope of the Regression Line H o :  1 = 0 (X provides no information) H a :  1 ≠ 0 (X does provide information) Two-Tailed Test Test Statistic: t =t =t =t = b1b1sbsbb1b1sbsb 1 For large data sets, reject H o if |t| > 2

9 BUSI 6220 Model Assumptions and Residual Analysis Randomness Constant Variance Normal Distribution Residuals should have… Y Y - Y ^ Residual Plot ^

10 BUSI 6220 Residual Analysis Y - Y ^ Y Residual Plot Violation of the constant variance assumption How to fix it: Transformation ^

11 BUSI 6220 Residual Analysis Violation of the randomness assumption How to fix it: Add more predictor variables to explain patterns. In time series data, add lags of Y or X as predictors: Y t-1, X1 t-1, X1 t-2, X2 t-1, etc. Y - Y ^ Y Sequence Plot of Residuals ^

12 BUSI 6220 Residual Analysis Violation of the normality assumption How to fix it: Transformation (start with easy transformations, such as Log(Y), then continue with bucket transformations, etc.) Frequency Histogram of the residuals

13 BUSI 6220 Stepwise Procedures Procedures either choose or eliminate variables, one at a time, in an effort to avoid including variables with either no predictive ability or are highly correlated with other predictor variables  Forward selection Add one variable at a time until contribution is insignificant  Backward elimination Remove one variable at a time starting with the “worst” until R 2 drops significantly  Stepwise selection Forward regression with the ability to remove variables that become insignificant

14 BUSI 6220 Stepwise Regression Include X 3 Include X 6 Include X 2 Include X 5 Remove X 2 (When X 5 was inserted into the model X 2 became unnecessary) Include X 7 Remove X 7 - it is insignificant Stop Final model includes X 3, X 5 and X 6 An example implementation of the stepwise procedure:

15 BUSI 6220 Interpretation of Regression Coefficients: Linear Regression When Y is continuous (Linear Regression), the standard interpretation of a slope coefficient is as follows: The slope tells you the change in Y when a particular input X increases by one unit, while all other inputs are kept constant. Example: If a linear regression coefficient is equal to -2.5, that indicates that when the predictor increases, the target variable decreases (since the coefficient is negative ). Specifically, for each unit increase in the predictor variable, the target variable decreases by 2.5 units

16 BUSI 6220 Interpretation of Regression Coefficients: Logistic Regression When Y is binary (Logistic Regression), the standard interpretation of an odds ratio coefficient is as follows: The odds ratio coefficient is a multiplier on the odds ratio for the target event T (=probability for T / probability for non-T) when a particular input X increases by one unit, while all other inputs are kept constant. Example: If an odds ratio coefficient is equal to 1.05, that indicates that when the predictor increases, the probability for the target event also increases. Specifically, for each unit increase in the predictor variable, the odds ratio P(event/nonevent) gets multiplied by 1.05