Stat 512 – Lecture 17 Inference for Regression (9.5, 9.6)

Slides:



Advertisements
Similar presentations
Chapter 3 Examining Relationships Lindsey Van Cleave AP Statistics September 24, 2006.
Advertisements

AP Statistics Section 3.2 C Coefficient of Determination
Lecture 17: Tues., March 16 Inference for simple linear regression (Ch ) R2 statistic (Ch ) Association is not causation (Ch ) Next.
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
Correlation and Regression
Chapter 4 The Relation between Two Variables
LECTURE 3 Introduction to Linear Regression and Correlation Analysis
Chapter 10 Simple Regression.
Stat 512 – Lecture 18 Multiple Regression (Ch. 11)
Lecture 19: Tues., Nov. 11th R-squared (8.6.1) Review
Stat 217 – Day 26 Regression, cont.. Last Time – Two quantitative variables Graphical summary  Scatterplot: direction, form (linear?), strength Numerical.
Chapter Topics Types of Regression Models
Stat 217 – Day 25 Regression. Last Time - ANOVA When?  Comparing 2 or means (one categorical and one quantitative variable) Research question  Null.
Stat 512 – Lecture 16 Two Quantitative Variables (Ch. 9)
CHAPTER 3 Describing Relationships
C HAPTER 2 S CATTER PLOTS, C ORRELATION, L INEAR R EGRESSION, I NFERENCES FOR R EGRESSION By: Tasha Carr, Lyndsay Gentile, Darya Rosikhina, Stacey Zarko.
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
Linear Regression/Correlation
Correlation & Regression
Descriptive Methods in Regression and Correlation
Linear Regression.
Regression and Correlation Methods Judy Zhong Ph.D.
MAT 254 – Probability and Statistics Sections 1,2 & Spring.
September In Chapter 14: 14.1 Data 14.2 Scatterplots 14.3 Correlation 14.4 Regression.
Correlation Correlation measures the strength of the LINEAR relationship between 2 quantitative variables. Labeled as r Takes on the values -1 < r < 1.
Ch 3 – Examining Relationships YMS – 3.1
Ms. Khatijahhusna Abd Rani School of Electrical System Engineering Sem II 2014/2015.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
1 Chapter 10 Correlation and Regression 10.2 Correlation 10.3 Regression.
Applied Quantitative Analysis and Practices LECTURE#22 By Dr. Osman Sadiq Paracha.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
Chapter 10 Correlation and Regression
Objective: Understanding and using linear regression Answer the following questions: (c) If one house is larger in size than another, do you think it affects.
Copyright ©2011 Brooks/Cole, Cengage Learning Inference about Simple Regression Chapter 14 1.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 3 Describing Relationships 3.2 Least-Squares.
Examining Bivariate Data Unit 3 – Statistics. Some Vocabulary Response aka Dependent Variable –Measures an outcome of a study Explanatory aka Independent.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Lecture 10: Correlation and Regression Model.
Chapter 4 Summary Scatter diagrams of data pairs (x, y) are useful in helping us determine visually if there is any relation between x and y values and,
 Find the Least Squares Regression Line and interpret its slope, y-intercept, and the coefficients of correlation and determination  Justify the regression.
Chapter 14: Inference for Regression. A brief review of chapter 4... (Regression Analysis: Exploring Association BetweenVariables )  Bi-variate data.
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
LEAST-SQUARES REGRESSION 3.2 Least Squares Regression Line and Residuals.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Simple Linear Regression The Coefficients of Correlation and Determination Two Quantitative Variables x variable – independent variable or explanatory.
Lecture 10 Introduction to Linear Regression and Correlation Analysis.
Chapters 8 Linear Regression. Correlation and Regression Correlation = linear relationship between two variables. Summarize relationship with line. Called.
Warm-Up The least squares slope b1 is an estimate of the true slope of the line that relates global average temperature to CO2. Since b1 = is very.
Topics
AP Statistics Chapter 14 Section 1.
CHAPTER 3 Describing Relationships
CHAPTER 26: Inference for Regression
AP Stats: 3.3 Least-Squares Regression Line
Section 3.3 Linear Regression
AP Statistics, Section 3.3, Part 1
CHAPTER 3 Describing Relationships
Least-Squares Regression
Least-Squares Regression
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
Algebra Review The equation of a straight line y = mx + b
CHAPTER 3 Describing Relationships
Presentation transcript:

Stat 512 – Lecture 17 Inference for Regression (9.5, 9.6)

Last Time – Two Quantitative Variables Question: Is there an association between the two variables? Graphical summary: Scatterplot of response variable vs. explanatory variable (horizontal)  Description: Direction, form, strength Numerical summary: If linear, r, Pearson’s correlation coefficient, -1 < r < 1

Practice Problem r =.504 r =.938

Temperatures vs. time r =.029 Always plot the data!!

Least-Squares Regression Line Model: Least Squares Regression Line  Minimize sum of squared residuals Response-hat = a + b explanatory  a = intercept, predicted value when explanatory=0  b = slope, predicted change in response associated with an increase in explanatory variable by 1 unit  Use regression line for making predictions Warnings: regression line is not resistant  Influential observation = removing the value changes the regression equation  Outliers = extreme residual value

R2R2 If predict everyone to have the same height, lots of “unexplained” variation (SSE = ) If take explanatory variable into account, much less “unexplained” variation (SSE = 235) % change=( ) = 50.6%

R2R2 Of the variability in the heights, 50.6% of that variation is explained by this regression on foot length BAD  50.6% of points lie on the line  50.6% of predictions will be correct

Example 1 (cont.): Airline costs Each flight has a ‘set up’ cost of $151 and each additional mile of travels is associated with an predicted increase in cost of about 7 cents. 19.3% of the variability in airfare is explained by this regression on distance (still lots of unexplained variability) Might investigate further while the cost for ACK was so much higher than expected

Inference in Regression Is the relationship between the two variables statistically significant?  Need to understand how the behavior/variability of regression lines from different random samples

Example 1: House Prices Observational units, variable  Houses, price (quantitative) Positive linear moderately strong association  Larger houses tend to cost more! Predicted price = sq ft, r 2 =42.1% Perhaps houses in Northern CA tend to be a bit more expensive even for the same size, but not a huge difference

Example 1: House Prices H 0 : no association between price and size  H 0 :  = 0 H a :  > 0, there is a positive association between price and size

Example 1: House Prices 1) Curvature? Not really 2) normality? no3) independence? Random sample 4) Equal spread? No

Example 1: House Prices 1) Curvature? No 2) normality? better3) independence? Random sample 4) Equal spread? Yes

Simulating Regression Lines Sampling variabiltiy A slope of.8899 would be quite surprising!.8899 is more than 7 standard errors from 0! p-value <.001, less than.1% of random samples from a population with  =0 would see such an extreme sample slope by chance alone

Example 2: Airfare Costs H 0 :  = 0 (no association between price and distance) H a :  > 0 (cities further away are associated with more expensive flights) p-value =.002/2 =.001 Strong evidence against the null hypothesis, statistically significant evidence of a positive relationship between price and distance BUT residual plots don’t look so great

Example 3: Money-Making Movies

For Tuesday Skim non-starred sections of Ch. 11 Submit PP 15 in Blackboard by 3pm Submit Project Report 3 in class (see syllabus for details)