June 30, 2008Stat 111 - Lecture 16 - Regression1 Inference for relationships between variables Statistics 111 - Lecture 16.

Slides:



Advertisements
Similar presentations
Objectives 10.1 Simple linear regression
Advertisements

Inference for Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Correlation and Regression
CHAPTER 24: Inference for Regression
July 1, 2008Lecture 17 - Regression Testing1 Testing Relationships between Variables Statistics Lecture 17.
Session 2. Applied Regression -- Prof. Juran2 Outline for Session 2 More Simple Regression –Bottom Part of the Output Hypothesis Testing –Significance.
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
Multiple regression analysis
Chapter 10 Simple Regression.
Chapter 12 Simple Regression
BCOR 1020 Business Statistics Lecture 28 – May 1, 2008.
The Simple Regression Model
Stat 112 – Notes 3 Homework 1 is due at the beginning of class next Thursday.
SIMPLE LINEAR REGRESSION
Chapter 11 Multiple Regression.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Business Statistics - QBM117 Interval estimation for the slope and y-intercept Hypothesis tests for regression.
Chapter 12 Section 1 Inference for Linear Regression.
Simple Linear Regression Analysis
Lecture 5 Correlation and Regression
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
June 2, 2008Stat Lecture 18 - Review1 Final review Statistics Lecture 18.
SIMPLE LINEAR REGRESSION
The Chi-Square Distribution 1. The student will be able to  Perform a Goodness of Fit hypothesis test  Perform a Test of Independence hypothesis test.
Introduction to Linear Regression and Correlation Analysis
Inference for regression - Simple linear regression
Chapter 13: Inference in Regression
STA291 Statistical Methods Lecture 27. Inference for Regression.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
June 3, 2008Stat Lecture 5 - Correlation1 Exploring Data Numerical Summaries of Relationships between Statistics Lecture 5.
More About Significance Tests
June 18, 2008Stat Lecture 11 - Confidence Intervals 1 Introduction to Inference Sampling Distributions, Confidence Intervals and Hypothesis Testing.
June 26, 2008Stat Lecture 151 Two-Sample Inference for Proportions Statistics Lecture 15.
June 25, 2008Stat Lecture 14 - Two Means1 Comparing Means from Two Samples Statistics 111 – Lecture 14 One-Sample Inference for Proportions and.
CHAPTER 14 MULTIPLE REGRESSION
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Introduction to Linear Regression
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 4 Section 2 – Slide 1 of 20 Chapter 4 Section 2 Least-Squares Regression.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Basic Concepts of Correlation. Definition A correlation exists between two variables when the values of one are somehow associated with the values of.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
+ Chapter 12: More About Regression Section 12.1 Inference for Linear Regression.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Agresti/Franklin Statistics, 1 of 88 Chapter 11 Analyzing Association Between Quantitative Variables: Regression Analysis Learn…. To use regression analysis.
Applied Quantitative Analysis and Practices LECTURE#25 By Dr. Osman Sadiq Paracha.
Chapter 10 Inference for Regression
Lesson 14 - R Chapter 14 Review. Objectives Summarize the chapter Define the vocabulary used Complete all objectives Successfully answer any of the review.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Regression Analysis Deterministic model No chance of an error in calculating y for a given x Probabilistic model chance of an error First order linear.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Lecture #25 Tuesday, November 15, 2016 Textbook: 14.1 and 14.3
Inference for Regression (Chapter 14) A.P. Stats Review Topic #3
Inference for Regression
CHAPTER 12 More About Regression
Chapter 11 Simple Regression
Correlation and Simple Linear Regression
CHAPTER 26: Inference for Regression
Correlation and Simple Linear Regression
Statistical Inference about Regression
Elementary Statistics: Looking at the Big Picture
CHAPTER 12 More About Regression
CHAPTER 12 More About Regression
Correlation and Simple Linear Regression
Presentation transcript:

June 30, 2008Stat Lecture 16 - Regression1 Inference for relationships between variables Statistics Lecture 16

June 30, 2008Stat Lecture 16 - Regression2 Administrative Notes Homework 5 due on Wednesday, July 1 st

Final is this Thursday –Will start at exactly 10:40 –Will end at exactly 12:10 –Get here a bit early or you will be wasting time on your exam Final will focus on material from after midterm There will be at least two questions that require knowing material from before midterm Calculators are definitely needed! Single 8.5 x 11 cheat sheet (two-sided) allowed June 30, 2008Stat Lecture 16 - Regression3 Final Exam

June 30, 2008Stat Lecture 16 - Regression4 Inference Thus Far Tests and intervals for a single variable Tests and intervals to compare a single variable between two samples For the last couple of classes, we have looked at count data and inference for population proportions Before that, we looked at continuous data and inference for population means Next couple of classes: inference for a relationship between two continuous variables

June 30, 2008Stat Lecture 16 - Regression5 Two Continuous Variables Remember linear relationships between two continuous variables? Scatterplots Correlation Best Fit Lines

June 30, 2008Stat Lecture 16 - Regression6 Scatterplots and Correlation Visually summarize the relationship between two continuous variables with a scatterplot If our X and Y variables show a linear relationship, we can calculate a best fit line between Y and X Education and Mortality: r = Draft Order and Birthdayr = -0.22

June 30, 2008Stat Lecture 16 - Regression7 Linear Regression Best fit line is called Simple Linear Regression Model: Coefficients:  is the intercept and  is the slope Other common notation:  0 for intercept,  1 for slope Our Y variable is a linear function of the X variable but we allow for error (ε i ) in each prediction We approximate the error by using the residual Observed Y i Predicted Y i =  +  X i

June 30, 2008Stat Lecture 16 - Regression8 Residuals and Best Fit Line  0 and  1 that give the best fit line are the values that give smallest sum of squared residuals: Best fit lineis also called the least-squares line e i = residual i

June 30, 2008Stat Lecture 16 - Regression9 Best values for Regression Parameters The best fit line has these values for the regression coefficients: Also can estimate the average squared residual: Best estimate of slope  Best estimate of intercept 

June 30, 2008Stat Lecture 16 - Regression10 Example: Education and Mortality Mortality = · Education Negative association means negative slope b

June 30, 2008Stat Lecture 16 - Regression11 Interpreting the regression line The estimated slope b is the average change you get in the Y variable if you increased the X variable by one Eg. increasing education by one year means an average decrease of ≈ 38 deaths per 100,000 The estimated intercept a is the average value of the Y variable when the X variable is equal to zero Eg. an uneducated city (Education = 0) would have an average mortality of 1353 deaths per 100,000

June 30, 2008Stat Lecture 16 - Regression12 Example: Vietnam Draft Order Draft Order = · Birthday Slightly negative slope means later birthdays have a lower draft order

June 30, 2008Stat Lecture 16 - Regression13 Significance of Regression Line Does the regression line show a significant linear relationship between the two variables? If there is not a linear relationship, then we would expect zero correlation (r = 0) So the estimated slope b should also be close to zero Therefore, our test for a significant relationship will focus on testing whether our true slope  is significantly different from zero: H 0 :  = 0 versus H a :  0 Our test statistic is based on the estimated slope b

June 30, 2008Stat Lecture 16 - Regression14 Test Statistic for Slope Our test statistic for the slope is similar in form to all the test statistics we have seen so far: The standard error of the slope SE(b) has a complicated formula that requires some matrix algebra to calculate We will not be doing this calculation manually because the JMP software does this calculation for us!

June 30, 2008Stat Lecture 16 - Regression15 Example: Education and Mortality

June 30, 2008Stat Lecture 16 - Regression16 p-value for Slope Test Is T = significantly different from zero? To calculate a p-value for our test statistic T, we use the t distribution with n-2 degrees of freedom For testing means, we used a t distribution as well, but we had n-1 degrees of freedom before For testing slopes, we use n-2 degrees of freedom because we are estimating two parameters (intercept and slope) instead of one (a mean) For cities dataset, n = 60, so we have d.f. = 58 Looking at a t-table with 58 df, we discover that the P(T < -4.53) <

June 30, 2008Stat Lecture 16 - Regression17 Conclusion for Cities Example Two-sided alternative: p-value < 2 x = We could get the p-value directly from the JMP output, which is actually more accurate than t-table Since our p-value is far less than the usual  -level of 0.05, we reject our null hypothesis We conclude that there is a statistically significant linear relationship between education and mortality

June 30, 2008Stat Lecture 16 - Regression18 Next Class - Lecture 17 More Inference for Regression Moore, McCabe and Craig: 10.1,