Lecture 10 Feb 1 2006. Added-variable Added variable plots give you a visual sense of whether x2 is a useful addition to the model: E(y|x1) = a + b x1.

Slides:



Advertisements
Similar presentations
Multiple Regression.
Advertisements

Section 10-3 Regression.
Lecture 10 F-tests in MLR (continued) Coefficients of Determination BMTRY 701 Biostatistical Methods II.
4/14/ lecture 81 STATS 330: Lecture 8. 4/14/ lecture 82 Collinearity Aims of today’s lecture: Explain the idea of collinearity and its connection.
Stat 112: Lecture 7 Notes Homework 2: Due next Thursday The Multiple Linear Regression model (Chapter 4.1) Inferences from multiple regression analysis.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Topic 15: General Linear Tests and Extra Sum of Squares.
July 1, 2008Lecture 17 - Regression Testing1 Testing Relationships between Variables Statistics Lecture 17.
Chapter 12 Simple Linear Regression
Session 2. Applied Regression -- Prof. Juran2 Outline for Session 2 More Simple Regression –Bottom Part of the Output Hypothesis Testing –Significance.
Analysis of Economic Data
Lecture 3 Cameron Kaplan
Analysis of Covariance Goals: 1)Reduce error variance. 2)Remove sources of bias from experiment. 3)Obtain adjusted estimates of population means.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Simple Linear Regression Analysis
Week Lecture 3Slide #1 Minimizing e 2 : Deriving OLS Estimators The problem Deriving b 0 Deriving b 1 Interpreting b 0 and b 1.
Analysis of Covariance Goals: 1)Reduce error variance. 2)Remove sources of bias from experiment. 3)Obtain adjusted estimates of population means.
BCOR 1020 Business Statistics Lecture 24 – April 17, 2008.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
SLOPE DUMMY VARIABLES 1 The scatter diagram shows the data for the 74 schools in Shanghai and the cost functions derived from a regression of COST on N.
Multiple Linear Regression Response Variable: Y Explanatory Variables: X 1,...,X k Model (Extension of Simple Regression): E(Y) =  +  1 X 1 +  +  k.
Regression with 2 IVs Generalization of Regression from 1 to 2 Independent Variables.
9/14/ Lecture 61 STATS 330: Lecture 6. 9/14/ Lecture 62 Inference for the Regression model Aim of today’s lecture: To discuss how we assess.
Analysis of Variance: Some Review and Some New Ideas
Lecture 14 Multiple Regression Model
© 2002 Prentice-Hall, Inc.Chap 14-1 Introduction to Multiple Regression Model.
 Combines linear regression and ANOVA  Can be used to compare g treatments, after controlling for quantitative factor believed to be related to response.
The Scientific Method Interpreting Data — Correlation and Regression Analysis.
Lecture 3: Inference in Simple Linear Regression BMTRY 701 Biostatistical Methods II.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Soc 3306a Lecture 9: Multivariate 2 More on Multiple Regression: Building a Model and Interpreting Coefficients.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Warsaw Summer School 2015, OSU Study Abroad Program Regression.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Lecture 14 Summary of previous Lecture Regression through the origin Scale and measurement units.
Lecture 12 Time Series Model Estimation
Environmental Modeling Basic Testing Methods - Statistics III.
Applied Quantitative Analysis and Practices LECTURE#25 By Dr. Osman Sadiq Paracha.
Linear Models Alan Lee Sample presentation for STATS 760.
June 30, 2008Stat Lecture 16 - Regression1 Inference for relationships between variables Statistics Lecture 16.
Chapter 12 Inference for Linear Regression. Reminder of Linear Regression First thing you should do is examine your data… First thing you should do is.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 12 More About Regression 12.1 Inference for.
Essentials of Business Statistics: Communicating with Numbers By Sanjiv Jaggia and Alison Kelly Copyright © 2014 by McGraw-Hill Higher Education. All rights.
Regression Analysis Deterministic model No chance of an error in calculating y for a given x Probabilistic model chance of an error First order linear.
Regression through the origin
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Chapter 8 Linear Regression. Fat Versus Protein: An Example 30 items on the Burger King menu:
Chapter 9 Minitab Recipe Cards. Contingency tests Enter the data from Example 9.1 in C1, C2 and C3.
Faculty of Arts Atkinson College MATH 3330 M W 2003 Welcome Seventh Lecture for MATH 3330 M Professor G.E. Denzel.
© 2001 Prentice-Hall, Inc.Chap 13-1 BA 201 Lecture 19 Measure of Variation in the Simple Linear Regression Model (Data)Data.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 12 More About Regression 12.1 Inference for.
Linear Regression Hypothesis testing and Estimation.
Jacek Wallusch _________________________________ Statistics for International Business Lecture 8: Distributions and Densities.
Stats Methods at IC Lecture 3: Regression.
Quantitative Methods Simple Regression.
POSC 202A: Lecture Lecture: Substantive Significance, Relationship between Variables 1.
Correlation and Regression
CHAPTER 29: Multiple Regression*
Rainfall Example The data set contains cord yield (bushes per acre) and rainfall (inches) in six US corn-producing states (Iowa, Nebraska, Illinois, Indiana,
Analysis of Variance: Some Review and Some New Ideas
Multiple Regression Models
Multiple Linear Regression
Ch 4.1 & 4.2 Two dimensions concept
Introduction to Regression
Adding variables. There is a difference between assessing the statistical significance of a variable acting alone and a variable being added to a model.
Presentation transcript:

Lecture 10 Feb

Added-variable Added variable plots give you a visual sense of whether x2 is a useful addition to the model: E(y|x1) = a + b x1

Steps to making one Regress y on x1 Compute residuals of y on x1: y "-"x1 (remove x1 from y) Regress x2 on x1 Compute residuals of x1 on x2: x2 "-"x1 (remove x1 from x2) Plot y"-"x1 vs. x2 "-"x1

Interpret If there is a "significant" slope, then x2 is useful. Slope of the added variable plot is the same as the coefficient if you fit E(y|x1, x2) = B 0 + B 1 x1 + B 2 x2

Significance Tests Find the slope of an added variable plot, and do a t-test to see if the slope is significant. The value of the t-stat is almost the same as the value of the t-stat for the "full" model. The p-values will differ because the degrees of freedom are different: n-2 for added- variable slope, n-3 for full model

Mussel Beds

Is density related to food levels?

Is density related to human use?

If we know the human use level, do we need to know food leve?

Added-Variable

Summary of added variable plot slope = additional amount of food worth mm of thickness, on avg., controlling for human use t = 4.742, p=1.45e-05

Summary of lm(thickness~food+human.use) E(thickness|food, human) = *food * human.use t_food , p = 1.84e-05

Testing one variable: Plan #1 H0: Beta1 = 0, Beta0, Beta2, Beta3, etc. "arbitrary" Ha: Beta1 <> 0, others arbitrary Fit full model: y = B0+B1*x1+B2*x2+B3*x3 etc. Fit reduced model y = B0 + B2*x2+B3*x3 Compare RSS

partial F-test Compare RSS (full) with RSS (reduced) Note: RSS(reduced) will be ???? than RSS(full) F = ( ( RSS(red) - RSS(full) ) /1 ) /RSS(full)/(n-p) Note: The denominator is just = ????

Formula for partial F (Reduction in RSS)/DF num divided by RSS/DF den DF of a model is n - # of parameters estimated. DF num is DF(full)-DF(reduced) DF den is n - p

In R (long) full <- lm(y ~ x1 + x2 + x3 + x4) red <- lm(y~x2 + x3 + x4) anova(full); anova(red) compute by hand from output

R(short) full <- lm(y ~ x1+x2+x3) anova(full) Read output

Plan #2 t^2 = F full <- lm(y~x1+x2+x3) summary(full) look at t-statistic