Simple Linear Regression

Slides:



Advertisements
Similar presentations
Chapter 12 Simple Linear Regression
Advertisements

Regression Analysis Simple Regression. y = mx + b y = a + bx.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Chapter Thirteen McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved. Linear Regression and Correlation.
Chapter 12 Simple Linear Regression
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
ASSESSING THE STRENGTH OF THE REGRESSION MODEL. Assessing the Model’s Strength Although the best straight line through a set of points may have been found.
Probability & Statistics for Engineers & Scientists, by Walpole, Myers, Myers & Ye ~ Chapter 11 Notes Class notes for ISE 201 San Jose State University.
Simple Linear Regression Analysis
Introduction to Probability and Statistics Linear Regression and Correlation.
BCOR 1020 Business Statistics
This Week Continue with linear regression Begin multiple regression –Le 8.2 –C & S 9:A-E Handout: Class examples and assignment 3.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
Least Squares Regression Fitting a Line to Bivariate Data.
Simple Linear Regression and Correlation
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Further Topics in Regression Analysis Objectives: By the end of this section, I will be able to… 1) Explain prediction error, calculate SSE, and.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Class 4 Simple Linear Regression. Regression Analysis Reality is thought to behave in a manner which may be simulated (predicted) to an acceptable degree.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Topic 10 - Linear Regression Least squares principle - pages 301 – – 309 Hypothesis tests/confidence intervals/prediction intervals for regression.
Simple Linear Regression. Deterministic Relationship If the value of y (dependent) is completely determined by the value of x (Independent variable) (Like.
Chapter 5: Regression Analysis Part 1: Simple Linear Regression.
Correlation and Regression Basic Concepts. An Example We can hypothesize that the value of a house increases as its size increases. Said differently,
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Review Lecture 51 Tue, Dec 13, Chapter 1 Sections 1.1 – 1.4. Sections 1.1 – 1.4. Be familiar with the language and principles of hypothesis testing.
ETM U 1 Multiple regression More than one indicator variable may be responsible for the variation we see in the response. Gas mileage is a function.
1 Simple Linear Regression and Correlation Least Squares Method The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES.
Regression Analysis Deterministic model No chance of an error in calculating y for a given x Probabilistic model chance of an error First order linear.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
AP STATISTICS LESSON 3 – 3 (DAY 2) The role of r 2 in regression.
Correlation and Regression Basic Concepts. An Example We can hypothesize that the value of a house increases as its size increases. Said differently,
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Lecture 11: Simple Linear Regression
Simple Linear Regression & Correlation
Regression and Correlation of Data Summary
Correlation & Regression
ENM 310 Design of Experiments and Regression Analysis
Linear Regression Using Excel
Ch12.1 Simple Linear Regression
Simple Linear Regression
Multiple Regression and Model Building
Regression Analysis PhD Course.
Regression Analysis Week 4.
Simple Regression Mary M. Whiteside, PhD.
Linear Regression.
Simple Linear Regression
AP STATISTICS LESSON 3 – 3 (DAY 2)
Least-Squares Regression
Prediction of new observations
Simple Linear Regression
Review of Chapter 2 Some Basic Concepts: Sample center
Least-Squares Regression
Simple Linear Regression
Linear Regression and Correlation
Regression lecture 2 1. Review: deterministic and random components
Simple Linear Regression
Linear Regression and Correlation
Simple Linear Regression
3.2. SIMPLE LINEAR REGRESSION
Regression and Correlation of Data
The Squared Correlation r2 – What Does It Tell Us?
Regression lecture 2 1. Review: deterministic and random components
Simple Linear Regression – Matrix Approach
Decomposition of Sum of Squares
Chapter 14, part C Goodness of Fit..
St. Edward’s University
The Squared Correlation r2 – What Does It Tell Us?
Presentation transcript:

Simple Linear Regression Often we want to understand the relationships among variables, e.g., SAT scores and college GPA car weight and gas mileage amount of a certain pollutant in wastewater and bacteria growth in local streams number of takeoffs and landings and degree of metal fatigue in aircraft structures Simplest relationship  Y = α + βx

Example The owner of a small harness race track in Florida is interested in understanding the relationship between attendance at the track and the total amount bet each night. The data for a two-week period (10 racing nights) is as follows: Attendance, x Amount Bet ($000), Y 117 2.07 128 2.8 122 3.14 119 2.26 131 3.4 135 3.89 125 2.93 120 2.66 130 3.33 127 3.54

Estimating the Regression Coefficients Method of Least Squares Determine a and b (estimates for α and β) so that the sum of the squares of the residuals is minimized.) Steps: Calculate b using and a using

For Our Example b = _______________________________________ Night Attendance, x Amount Bet, Y xiyi xi2 1 117 2.07 242.19 13689 2 128 2.8 358.4 16384 3 122 3.14 383.08 14884 4 119 2.26 268.94 14161 5 131 3.4 445.4 17161 6 135 3.89 525.15 18225 7 125 2.93 366.25 15625 8 120 2.66 319.2 14400 9 130 3.33 432.9 16900 10 127 3.54 449.58 16129 TOTAL 1254 30.02 3791.09 157558 b =((10*3791.09)-(1254*30.02))/((10*157558)-1254^2) = 0.086755875 a = (30.02/10) – 0.0868*(1254/10) = -7.877186684 b = _______________________________________ a = ______________________________

What does this mean? We can draw the regression line that describes the relationship between attendance and amount bet: We can also predict amount bet based on attendance.

How good is our prediction? Estimating the variance: Coefficient of determination, R2 a measure of the “quality of fit,” or the proportion of the variability explained by the fitted model. SSE = sum(residuals2)= 0.639015 s2 = SSE/8 = 0.079876917 SST = Σ(Yi - Y)2 = 2.94516 R2 = 1-(SSE/SST) = 1-(0.639/2.945)=0.783029 (see next page)

Calculations … Night Attendance, x Amount Bet, Y xiyi xi2 yhat residuals2 1 117 2.07 242.19 13689 2.27325 0.0413108 0.86862 2 119 2.26 268.94 14161 2.44676 0.0348802 0.55056 3 120 2.66 319.2 14400 2.53352 0.0159976 0.11696 4 122 3.14 383.08 14884 2.70703 0.187463 0.01904 5 125 2.93 366.25 15625 2.9673 0.0013911 0.00518 6 127 3.54 449.58 16129 3.14081 0.1593531 0.28944 7 128 2.8 358.4 16384 3.22757 0.1828121 0.0408 8 130 3.33 432.9 16900 3.40108 0.0050519 0.10758 9 131 3.4 445.4 17161 3.48783 0.0077146 0.1584 10 135 3.89 525.15 18225 3.83486 0.0030408 0.78854 TOTAL 1254 30.02 3791.09 157558 0.6390153 2.94516 (Y - Y)2

Or … Using Excel Note the confidence interval … we can also draw a confidence interval around our predictions.