Simultaneous inference Estimating (or testing) more than one thing at a time (such as β 0 and β 1 ) and feeling confident about it …

Slides:



Advertisements
Similar presentations
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
Advertisements

Objectives (BPS chapter 24)
Simple Linear Regression
REGRESSION Want to predict one variable (say Y) using the other variable (say X) GOAL: Set up an equation connecting X and Y. Linear regression linear.
Chapter 12 Simple Regression
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Simple Linear Regression Basic Business Statistics 11 th Edition.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Simple Linear Regression Analysis
REGRESSION AND CORRELATION
Notes on Logistic Regression STAT 4330/8330. Introduction Previously, you learned about odds ratios (OR’s). We now transition and begin discussion of.
Simple Linear Regression and Correlation
Fixing problems with the model Transforming the data so that the simple linear regression model is okay for the transformed data.
Business Statistics: Communicating with Numbers
Correlation & Regression
Topic 10: Miscellaneous Topics. Outline Joint estimation of β 0 and β 1 Multiplicity Regression through the origin Measurement error Inverse predictions.
Simple linear regression Linear regression with one predictor variable.
Copyright © Cengage Learning. All rights reserved. 12 Simple Linear Regression and Correlation.
Prediction concerning Y variable. Three different research questions What is the mean response, E(Y h ), for a given level, X h, of the predictor variable?
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.2 Estimating Differences.
Time Series Analysis – Chapter 4 Hypothesis Testing Hypothesis testing is basic to the scientific method and statistical theory gives us a way of conducting.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
CHAPTER 14 MULTIPLE REGRESSION
Introduction to Linear Regression
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Prediction concerning the response Y. Where does this topic fit in? Model formulation Model estimation Model evaluation Model use.
1 Lecture 4 Main Tasks Today 1. Review of Lecture 3 2. Accuracy of the LS estimators 3. Significance Tests of the Parameters 4. Confidence Interval 5.
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.3 Using Multiple Regression to Make Inferences.
CORRELATION: Correlation analysis Correlation analysis is used to measure the strength of association (linear relationship) between two quantitative variables.
Time Series Analysis – Chapter 6 Odds and Ends
Lack of Fit (LOF) Test A formal F test for checking whether a specific type of regression function adequately fits the data.
Multiple regression. Example: Brain and body size predictive of intelligence? Sample of n = 38 college students Response (Y): intelligence based on the.
Regression Analysis Part C Confidence Intervals and Hypothesis Testing
Inference with computer printouts. Coefficie nts Standard Errort StatP-value Lower 95% Upper 95% Intercept
Confidence Intervals vs. Prediction Intervals Confidence Intervals – provide an interval estimate with a 100(1-  ) measure of reliability about the mean.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
Stat 112 Notes 6 Today: –Chapter 4.1 (Introduction to Multiple Regression)
A first order model with one binary and one quantitative predictor variable.
©2011 Brooks/Cole, Cengage Learning Elementary Statistics: Looking at the Big Picture1 Lecture 35: Chapter 13, Section 2 Two Quantitative Variables Interval.
Logistic regression (when you have a binary response variable)
Regression through the origin
Using SPSS Note: The use of another statistical package such as Minitab is similar to using SPSS.
Copyright © Cengage Learning. All rights reserved. 13 Nonlinear and Multiple Regression.
Inference for  0 and 1 Confidence intervals and hypothesis tests.
Variable selection and model building Part I. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Inference with Computer Printouts. Leaning Tower of Pisa Find a 90% confidence interval. Year Lean
Example 9.13 Sample Size Selection for Estimating the Proportion Who Have Tried a New Sandwich Controlling Confidence Interval Length.
Chapter 4 Minitab Recipe Cards. Correlation coefficients Enter the data from Example 4.1 in columns C1 and C2 of the worksheet.
732G21/732G28/732A35 Lecture 3. Properties of the model errors ε 4. ε are assumed to be normally distributed
Regression Analysis Presentation 13. Regression In Chapter 15, we looked at associations between two categorical variables. We will now focus on relationships.
Announcements There’s an in class exam one week from today (4/30). It will not include ANOVA or regression. On Thursday, I will list covered material and.
Multiple Regression.
Lecture #25 Tuesday, November 15, 2016 Textbook: 14.1 and 14.3
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
Notes on Logistic Regression
Multiple Regression.
Chapter 11 Simple Regression
Chapter 13 Simple Linear Regression
Simultaneous Inferences and Other Regression Topics
Multiple Regression.
Regression Models - Introduction
Hypothesis testing and Estimation
Simple Linear Regression
Essentials of Statistics for Business and Economics (8e)
Statistics 350 Lecture 12.
Regression Models - Introduction
Pearson Correlation and R2
Presentation transcript:

Simultaneous inference Estimating (or testing) more than one thing at a time (such as β 0 and β 1 ) and feeling confident about it …

Simultaneous inference we’ll be concerned about … Estimating β 0 and β 1 jointly. Estimating more than one mean response, E(Y), at a time. Predicting more than one new observation at a time.

Why simultaneous inference is important A 95% confidence interval implies a 95% chance that the interval contains β 0. A 95% confidence interval implies a 95% chance that the interval contains β 1. If the intervals are independent, then have only a (0.95×0.95) ×100 = 90.25% chance that both intervals are correct. (Intervals not independent, but point made.)

Terminology Family of estimates (or tests): a set of estimates (or tests) which you want all to be simultaneously correct. Statement confidence level: the confidence level, as you know it, that is, for just one parameter. Family confidence level: the confidence level of the whole family of interval estimates (or tests).

Examples A 95% confidence interval for β 0 – the 95% is a statement confidence level. A 95% confidence interval for β 1 – the 95% is a statement confidence level. Consider family of interval estimates for β 0 and β 1. If a 90.25% chance that both intervals are simultaneously correct, then 90.25% is the family confidence level.

Bonferroni joint confidence intervals for β 0 and β 1 GOAL: To formulate joint confidence intervals for β 0 and β 1 with a specified family confidence level. BASIC IDEA: –Make statement confidence level for β 0 higher –Make statement confidence level for β 1 higher –So that the family confidence level for (β 0, β 1 ) is at least (1-α)×100%.

Recall: Original confidence intervals For β 0 : For β 1 : Goal is to adjust the t-multiples so that family confidence coefficient is 1-α. That is, we need to find the α* to put into the above formulas to achieve the desired family coefficient of 1- α.

A little derivation Let A 1 = the event that first confidence interval does not contain β 0 (i.e., incorrect). So A 1 C = the event that first confidence interval contains β 0 (i.e., correct). P(A 1 ) = α and P(A 1 C ) = 1- α

A little derivation (cont’d) Let A 2 = the event that second confidence interval does not contain β 1 (i.e., incorrect). So A 2 C = the event that second confidence interval contains β 1 (i.e., correct). P(A 2 ) = α and P(A 2 C ) = 1- α

Becoming a not so little derivation… A1A1 A2A2 A 1 or A 2 A 1 C and A 2 C We want P(A 1 C and A 2 C ) to be at least 1-α. P(A 1 C and A 2 C ) = 1 – P(A 1 or A 2 ) = 1 – [P(A 1 )+P(A 2 ) – P(A 1 and A 2 )] = 1 – P(A 1 ) – P(A 2 ) + P(A 1 and A 2 )] ≥ 1 – P(A 1 ) – P(A 2 ) = 1 – α – α = 1 – 2α So, we need α* to be set to α/2.

Bonferroni joint confidence intervals Typically, the t-multiple in this setting is called the Bonferroni multiple and is denoted by the letter B.

Example: 90% family confidence interval The regression equation is punt = leg Predictor Coef SE Coef T P Constant leg n=13 punters t(0.975, 11) = We are 90% confident that β 0 is between and 83.9 and β 1 is between 0.44 and 1.36.

A couple of more points about Bonferroni intervals Bonferroni intervals are most useful when there are only a few interval estimates in the family (o.w., the intervals get too large). Can specify different statement confidence levels to get desired family confidence level. Bonferroni technique easily extends to g interval estimates. Set statement confidence levels at 1-(α/g), so need to look up 1- (α/2g).

Bonferroni intervals for more than one mean response at a time To estimate the mean response E(Y h ) for g different X h values with family confidence coefficient 1-α: where: g is the number of confidence intervals in the family

Example: Mean punting distance for leg strengths of 140, 150, 160 lbs. Predicted Values for New Observations New Fit SE Fit 95.0% CI 95.0% PI (130.55,152.01) (103.23,179.33) (140.13,160.49) (112.41,188.20) (147.72,170.95) (121.03,197.64) n=13 punters t(0.99, 11) = We are 94% confident that the mean responses for leg strengths of 140, 150, 160 pounds are …

Two procedures for predicting g new observations simultaneously Bonferroni procedure Scheffé procedure Use the procedure that gives the narrower prediction limits.

Bonferroni intervals for predicting more than one new obs’n at a time To predict g new observations Y h for g different X h values with family confidence coefficient 1-α: where: g is the number of prediction intervals in the family

Scheffé intervals for predicting more than one new obs’n at a time To predict g new observations Y h for g different X h values with family confidence coefficient 1-α: where: g is the number of prediction intervals in the family

Example: Punting distance for leg strengths of 140 and 150 lbs. n = 13 punters Bonferroni multiple: Suppose we want a 90% family confidence level. Scheffé multiple: Since B is smaller than S, the Bonferroni prediction intervals will be narrower … so use them here instead of the Scheffé intervals.

Example: Punting distance for leg strengths of 140 and 150 lbs. Predicted Values for New Observations New Fit SE Fit 95.0% CI 95.0% PI (130.55,152.01) (103.23,179.33) (140.13,160.49) (112.41,188.20) n=13 punters s(pred(140)) = There is a 90% chance that the punting distances for leg strengths of 140 and 150 pounds will be… s(pred(150)) = 17.21

Simultaneous prediction in Minitab Stat >> Regression >> Regression … Specify predictor and response. Under Options …, In “Prediction intervals for new observations” box, specify a column name containing multiple X values. Specify confidence level. Click on OK. Results appear in session window.