ENGR 610 Applied Statistics Fall 2007 - Week 11 Marshall University CITE Jack Smith.

Slides:



Advertisements
Similar presentations
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Advertisements

Forecasting Using the Simple Linear Regression Model and Correlation
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Linear regression models
Chapter 12 Simple Linear Regression
Chapter 13 Multiple Regression
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
Chapter 12 Simple Regression
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 11 th Edition.
Chapter 12 Multiple Regression
1 Pertemuan 13 Uji Koefisien Korelasi dan Regresi Matakuliah: A0392 – Statistik Ekonomi Tahun: 2006.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 13-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Pengujian Parameter Koefisien Korelasi Pertemuan 04 Matakuliah: I0174 – Analisis Regresi Tahun: Ganjil 2007/2008.
Chapter Topics Types of Regression Models
Linear Regression Example Data
Ch. 14: The Multiple Regression Model building
Korelasi dalam Regresi Linear Sederhana Pertemuan 03 Matakuliah: I0174 – Analisis Regresi Tahun: Ganjil 2007/2008.
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
Pertemua 19 Regresi Linier
Simple Linear Regression. Chapter Topics Types of Regression Models Determining the Simple Linear Regression Equation Measures of Variation Assumptions.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 10 th Edition.
Chapter 7 Forecasting with Simple Regression
Introduction to Regression Analysis, Chapter 13,
Chapter 13 Simple Linear Regression
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
Chapter 8 Forecasting with Multiple Regression
Chapter 13 Simple Linear Regression
Regression and Correlation Methods Judy Zhong Ph.D.
Introduction to Linear Regression and Correlation Analysis
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 12-1 Chapter 12 Simple Linear Regression Statistics for Managers Using.
CPE 619 Simple Linear Regression Models Aleksandar Milenković The LaCASA Laboratory Electrical and Computer Engineering Department The University of Alabama.
Simple Linear Regression Models
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Chapter 14 Introduction to Multiple Regression
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Chapter 5: Regression Analysis Part 1: Simple Linear Regression.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Chap 13-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 12.
Simple Linear Regression (SLR)
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
Lecture 10: Correlation and Regression Model.
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 14-1 Chapter 14 Introduction to Multiple Regression Statistics for Managers using Microsoft.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Statistics for Managers Using Microsoft® Excel 5th Edition
Chapter 12 Simple Linear Regression.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 12-1 Chapter 12 Simple Linear Regression Statistics for Managers Using.
ENGR 610 Applied Statistics Fall Week 8 Marshall University CITE Jack Smith.
ENGR 610 Applied Statistics Fall Week 12 Marshall University CITE Jack Smith.
Conceptual Foundations © 2008 Pearson Education Australia Lecture slides for this course are based on teaching materials provided/referred by: (1) Statistics.
Chapter 13 Simple Linear Regression
Chapter 14 Introduction to Multiple Regression
Chapter 20 Linear and Multiple Regression
Statistics for Managers using Microsoft Excel 3rd Edition
Simple Linear Regression
Chapter 11 Simple Regression
Relationship with one independent variable
Chapter 13 Simple Linear Regression
PENGOLAHAN DAN PENYAJIAN
Relationship with one independent variable
Chapter 13 Simple Linear Regression
Presentation transcript:

ENGR 610 Applied Statistics Fall Week 11 Marshall University CITE Jack Smith

Overview for Today Review Simple Linear Regression, Ch 12 Go over problem Multiple Linear Regression, Ch 13 (1-5) Multiple explanatory variables Coefficient of multiple determination Adjusted R 2 Residue Analysis F-test t test and confidence interval for slope Partial F-tests for each individual contributions Coefficients of partial determination Homework assignment

Regression Modeling Analysis of variance to “fit” a predictive model for a response (dependent) variable to a set of one or more explanatory (independent) variables Minimize residual error w.r.t. linear coefficients Interpolative over relevant range - do not extrapolative Typically linear, but may be curvilinear or more complex (w.r.t. independent variables) Related to Correlation Analysis - measuring the strength of association between variables Regression is about variance in the response variable Correlation is about co-variance - symmetric

Types of Regression Models Based on Scatter Plots Y vs X Dependent vs independent Linear Models Positive, negative or no slope Zero or non-zero intercept Curvilinear Models Positive, negative or no “slope” Positive, negative or varied curvature May be U shaped, with extrema May be asymptotically or piece-wise linear May be polynomial, exponential, inverse,…

Least-Square Linear Regression Simple Linear Model (for population) Y i =  0 +  1 X i +  i X i = value of independent variable Y i = observed value of dependent variable  0 = Y-intercept (Y at X=0)  1 = slope (  Y/  X)  i = random error for observation i Y i ’ = b 0 + b 1 X i (predicted value) b 0 and b 1 are called regression coefficients e i = Y i - Y i ’ (residual) Minimize  e i 2 for sample with respect to b 0 and b 1

Partitioning of Variation Total variation Regression variation Random variation (Mean response) SST = SSR + SSE Coefficient of Determination r 2 = SSR/SST Standard Error of the Estimate

Partitioning of Variation - Graphically

Assumptions of Regression (and Correlation) Normality of error about regression line Homoscedasticity (equal variance) along X Independence of errors with respect to X No autocorrelation in time Analysis of residuals to test assumptions Histogram, Box-and-Whisker plots Normalcy plot Ordered plots (by X, by time,…) See figures on pp 584-5

t Test for Slope H 0 :  1 = 0 Critical t value based on chosen level of significance, , and n-2 degrees of freedom

F Test for Single Regression F = MSR / MSE Reject H 0 if F > F U ( ,1,n-2) [or p<  ] Note: t 2 ( ,n-2) = F U ( ,1,n-2) One-Way ANOVA Summary SourceDegrees of Freedom (df) Sum of Squares (SS) Mean Square (MS) (Variance) Fp-value Regression1SSRMSR = SSRMSR/ MSE Errorn-2SSEMSE = SSE/(n-2) Totaln-1SST

Confidence and Prediction Intervals Confidence Interval Estimate for the Slope Confidence Interval Estimate for the Mean Confidence Interval Estimate for Individual Response See Fig 12.16, p 592

Pitfalls Not testing assumptions of least-square regression by analyzing residuals, looking for Patterns Outliers Non-uniform distribution about mean See Figs , p Not being aware of alternatives to least-square regression when assumptions are violated Not knowing subject matter being modeled

Computing by Hand Slope Y-Intercept

Computing by Hand Measures of Variation

Coefficient of Correlation For a regression For a correlation Covariance Also called… Pearson’s product-moment correlation coefficient

t Test for Correlation H 0 :  = 0 Critical t value based on chosen level of significance, , and n-2 degrees of freedom Compared to F U ( ,1,n-2) = t 2 ( ,n-2) Or

Multiple Regression Linear model - multiple dependent variables Y i =  0 +  1 X 1i + … +  j X ji +  i X ji = value of independent variable Y i = observed value of dependent variable  0 = Y-intercept (Y at X=0)  j = slope (  Y/  X j )  i = random error for observation i Y i ’ = b 0 + b 1 X i + … + b j X ji (predicted value) The b j ’s are called the regression coefficients e i = Y i - Y i ’ (residual) Minimize  e i 2 for sample with respect to all b j

Partitioning of Variation Total variation Regression variation Random variation (Mean response) SST = SSR + SSE Coefficient of Multiple Determination R 2 Y.12..k = SSR/SST Standard Error of the Estimate

Adjusted R 2 To account for sample size (n) and number of dependent variables (k) for comparison purposes

Residual Analysis Plot residuals vs Y i ’ (predicted values) X 1, X 2,…,X k Time (for autocorrelation) Check for Patterns Outliers Non-uniform distribution about mean See Figs , p 597-8

F Test for Multiple Regression F = MSR / MSE Reject H 0 if F > F U ( ,k,n-k-1) [or p<  ] k = number of independent variables One-Way ANOVA Summary SourceDegrees of Freedom (df) Sum of Squares (SS) Mean Square (MS) (Variance) Fp-value RegressionkSSRMSR = SSR/kMSR/ MSE Errorn-k-1SSEMSE = SSE/(n-k-1) Totaln-1SST

Alternate F-Test Compared to F U ( ,k,n-k-1)

t Test for Slope H 0 :  j = 0 Critical t value based on chosen level of significance, , and n-k-1 degrees of freedom See output from PHStat

Confidence and Prediction Intervals Confidence Interval Estimate for the Slope Confidence Interval Estimate for the Mean and Prediction Interval Estimate for Individual Response Beyond the scope of this text

Partial F Tests Significance test for contribution from individual independent variable Measure of incremental improvement All others already taken into account F j = SSR(X j |{X i≠j }) / MSE SSR(X j |{X i≠j }) = SSR - SSR({X i≠j }) Reject H 0 if F j > F U ( ,1,n-k-1) [or p<  ] Note: t 2 ( ,n-k-1) = F U ( ,1,n-k-1)

Coefficients of Partial Determination See PHStat output in Fig 13.10, p 637

Homework Review “Multiple Regression”, Work through Appendix 13.1 Work and hand in Problem Read “Multiple Regression”, Quadratic model Dummy-variable model Using transformations Collinearity (VIF) Modeling building C p statistic and stepwise regression Preview problems