Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections 7.3-7.4): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.

Slides:



Advertisements
Similar presentations
Lecture 17: Tues., March 16 Inference for simple linear regression (Ch ) R2 statistic (Ch ) Association is not causation (Ch ) Next.
Advertisements

Lesson 10: Linear Regression and Correlation
13- 1 Chapter Thirteen McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc., All Rights Reserved.
Stat 112: Lecture 7 Notes Homework 2: Due next Thursday The Multiple Linear Regression model (Chapter 4.1) Inferences from multiple regression analysis.
Inference for Regression
Lecture 6 Outline – Thur. Jan. 29
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Simple Linear Regression. Start by exploring the data Construct a scatterplot  Does a linear relationship between variables exist?  Is the relationship.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Ch11 Curve Fitting Dr. Deshi Ye
Objectives (BPS chapter 24)
Simple Linear Regression
Chapter 10 Simple Regression.
Lecture 23: Tues., Dec. 2 Today: Thursday:
Class 15: Tuesday, Nov. 2 Multiple Regression (Chapter 11, Moore and McCabe).
Chapter 12 Simple Regression
Class 5: Thurs., Sep. 23 Example of using regression to make predictions and understand the likely errors in the predictions: salaries of teachers and.
Lecture 23: Tues., April 6 Interpretation of regression coefficients (handout) Inference for multiple regression.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
Lecture 19: Tues., Nov. 11th R-squared (8.6.1) Review
The Simple Regression Model
Stat 112 – Notes 3 Homework 1 is due at the beginning of class next Thursday.
SIMPLE LINEAR REGRESSION
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Pengujian Parameter Koefisien Korelasi Pertemuan 04 Matakuliah: I0174 – Analisis Regresi Tahun: Ganjil 2007/2008.
Chapter Topics Types of Regression Models
Linear Regression and Correlation Analysis
Class 10: Tuesday, Oct. 12 Hurricane data set, review of confidence intervals and hypothesis tests Confidence intervals for mean response Prediction intervals.
Chapter 11 Multiple Regression.
Simple Linear Regression Analysis
SIMPLE LINEAR REGRESSION
Lecture 17 Interaction Plots Simple Linear Regression (Chapter ) Homework 4 due Friday. JMP instructions for question are actually for.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Lecture 19 Simple linear regression (Review, 18.5, 18.8)
Simple Linear Regression and Correlation
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
SIMPLE LINEAR REGRESSION
Inference for regression - Simple linear regression
STA291 Statistical Methods Lecture 27. Inference for Regression.
Regression Analysis (2)
BPS - 3rd Ed. Chapter 211 Inference for Regression.
1 G Lect 10a G Lecture 10a Revisited Example: Okazaki’s inferences from a survey Inferences on correlation Correlation: Power and effect.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
Stat 112: Notes 2 Today’s class: Section 3.3. –Full description of simple linear regression model. –Checking the assumptions of the simple linear regression.
Copyright ©2011 Brooks/Cole, Cengage Learning Inference about Simple Regression Chapter 14 1.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Lecture 10: Correlation and Regression Model.
Regression Analysis Deterministic model No chance of an error in calculating y for a given x Probabilistic model chance of an error First order linear.
Stat 112 Notes 14 Assessing the assumptions of the multiple regression model and remedies when assumptions are not met (Chapter 6).
Regression Analysis Presentation 13. Regression In Chapter 15, we looked at associations between two categorical variables. We will now focus on relationships.
BPS - 5th Ed. Chapter 231 Inference for Regression.
Regression Analysis AGEC 784.
Inference for Regression (Chapter 14) A.P. Stats Review Topic #3
Virtual COMSATS Inferential Statistics Lecture-26
Stat 112 Notes 4 Today: Review of p-values for one-sided tests
CHAPTER 29: Multiple Regression*
CHAPTER 26: Inference for Regression
Simple Linear Regression
CHAPTER 12 More About Regression
SIMPLE LINEAR REGRESSION
Presentation transcript:

Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence Intervals for mean response –Prediction Intervals Next time: Robustness of least squares inferences, graphical tools for model assessment ( )

Regression Goal of regression: Estimate the mean response Y for subpopulations X=x, Example: Y= neuron activity index, X=years playing stringed instrument Simple linear regression model: Estimate and by least squares – choose to minimize the sum of squared residuals (prediction errors)

Ideal Model Assumptions of ideal simple linear regression model –There is a normally distributed subpopulation of responses for each value of the explanatory variable –The means of the subpopulations fall on a straight-line function of the explanatory variable. –The subpopulation standard deviations are all equal (to ) –The selection of an observation from any of the subpopulations is independent of the selection of any other observation.

The standard deviation is the standard deviation in each subpopulation. measures the accuracy of predictions from the regression. If the simple linear regression model holds, then approximately –68% of the observations will fall within of the regression line –95% of the observations will fall within of the regression line

Estimating Residuals provide basis for an estimate of Degrees of freedom for simple linear regression = n-2 If the simple linear regression models holds, then approximately –68% of the observations will fall within of the least squares line –95% of the observations will fall within of the least squares line

Inference for Simple Linear Regression Inference based on the ideal simple linear regression model holding. Inference based on taking repeated random samples ( ) from the same subpopulations ( ) as in the observed data. Types of inference: –Hypothesis tests for intercept and slope –Confidence intervals for intercept and slope –Confidence interval for mean of Y at X=X 0 –Prediction interval for future Y for which X=X 0

Hypothesis tests for and Hypothesis test of vs. –Based on t-test statistic, –p-value has usual interpretation, probability under the null hypothesis that |t| would be at least as large as its observed value, small p-value is evidence against null hypothesis Hypothesis test for vs. is based on an analogous test statistic. Test statistics and p-values can be found on JMP output under parameter estimates, obtained by using fit line after fit Y by X.

JMP output for example

Confidence Intervals for and Confidence intervals provide a range of plausible values for and 95% Confidence Intervals: Finding CIs in JMP: Can find under parameter estimates after fitting line. Can find in Table A.2. For brain activity study, CIs

Confidence Intervals for Mean of Y at X=X 0 What is a plausible range of values for 95% CI for :, Note about formula –Precision in estimating is not constant for all values of X. Precision decreases as X 0 gets farther away from sample average of X’s JMP implementation: Use Confid Curves fit command under red triangle next to Linear Fit after using Fit Y by X, fit line. Use the crosshair tool to find the exact values of the confidence interval endpoints for a given X 0.

Prediction Intervals What are likely values for a future value Y 0 at some specified value of X (=X 0 )? The best single prediction of a future response at X 0 is the estimated mean response: A prediction interval is an interval of likely values along with a measure of the likelihood that interval will contain response. 95% prediction interval for X 0 : If repeated samples are obtained from the subpopulations and a prediction interval is formed, the prediction interval will contain the value of Y 0 for a future observation from the subpopulation X 0 95% of the time.

Prediction Intervals Cont. Prediction interval must account for two sources of uncertainty: –Uncertainty about the location of the subpopulation mean –Uncertainty about where the future value will be in relation to its mean Prediction Error = Random Sampling Error + Estimation Error

Prediction Interval Formula 95% prediction interval at X 0 Compare to 95% CI for mean at X 0 : –Prediction interval is wider due to random sampling error in future response –As sample size n becomes large, margin of error of CI for mean goes to zero but margin of error of PI doesn’t. JMP implementation: Use Confid Curves Indiv command under red triangle next to Linear Fit after using Fit Y by X, fit line. Use the crosshair tool to find the exact values of the confidence interval endpoints for a given X 0.

Example A building maintenance company is planning to submit a bid on a contract to clean 40 corporate offices scattered throughout an office complex. The costs incurred by the maintenance company are proportional to the number of crews needed for this task. Currently the company has 11 crews. Will 11 crews be enough? Recent data are available for the number of rooms that were cleaned by varying number of crews. The data are in cleaning.jmp. Assuming a simple linear regression model holds, which is more relevant for answering the question of interest – a confidence interval for the mean number of rooms cleaned by 11 crews or a prediction interval for the number of rooms cleaned on a particular day by 11 crews?

Correlation Section Correlation is a measure of the degree of linear association between two variables X and Y. For each unit in population, both X and Y are measured. Population correlation = Correlation is between –1 and 1. Correlation of 0 indicates no linear association. Correlations near +1 indicates strong positive linear association; correlations near –1 indicate strong negative linear association.

Correlation and Regression Features of correlation –Dimension-free. Units of X and Y don’t matter. –Symmetric in X and Y. There is no “response” and “explanatory” variable. Correlation only measures degree of linear association. It is possible for there to be an exact relationship between X and Y and yet sample correlation coefficient is zero. Correlation in JMP: Click multivariate and put variables in Y, columns. Connection to regression –Test of slope vs. is identical to test of vs.. Test of correlation coefficient only makes sense if the pairs (X,Y) are randomly sampled from population.

Correlation in JMP