Simple Linear Regression

Slides:



Advertisements
Similar presentations
Coefficient of Determination- R²
Advertisements

Topic 12: Multiple Linear Regression
Chapter 12 Simple Linear Regression
Regression Analysis Simple Regression. y = mx + b y = a + bx.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Chapter Thirteen McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved. Linear Regression and Correlation.
Simple Regression Model
Simple Linear Regression. G. Baker, Department of Statistics University of South Carolina; Slide 2 Relationship Between Two Quantitative Variables If.
Definition  Regression Model  Regression Equation Y i =  0 +  1 X i ^ Given a collection of paired data, the regression equation algebraically describes.
Chapter 12 Simple Linear Regression
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
ASSESSING THE STRENGTH OF THE REGRESSION MODEL. Assessing the Model’s Strength Although the best straight line through a set of points may have been found.
Chapter Topics Types of Regression Models
Probability & Statistics for Engineers & Scientists, by Walpole, Myers, Myers & Ye ~ Chapter 11 Notes Class notes for ISE 201 San Jose State University.
Simple Linear Regression Analysis
BCOR 1020 Business Statistics
This Week Continue with linear regression Begin multiple regression –Le 8.2 –C & S 9:A-E Handout: Class examples and assignment 3.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
Simple Linear Regression and Correlation
Multiple Linear Regression Response Variable: Y Explanatory Variables: X 1,...,X k Model (Extension of Simple Regression): E(Y) =  +  1 X 1 +  +  k.
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Chapter 11 Simple Regression
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
Further Topics in Regression Analysis Objectives: By the end of this section, I will be able to… 1) Explain prediction error, calculate SSE, and.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Class 4 Simple Linear Regression. Regression Analysis Reality is thought to behave in a manner which may be simulated (predicted) to an acceptable degree.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Topic 10 - Linear Regression Least squares principle - pages 301 – – 309 Hypothesis tests/confidence intervals/prediction intervals for regression.
Simple Linear Regression. Deterministic Relationship If the value of y (dependent) is completely determined by the value of x (Independent variable) (Like.
Introduction to Regression Analysis. Dependent variable (response variable) Measures an outcome of a study  Income  GRE scores Dependent variable =
Regression Lesson 11. The General Linear Model n Relationship b/n predictor & outcome variables form straight line l Correlation, regression, t-tests,
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
ETM U 1 Multiple regression More than one indicator variable may be responsible for the variation we see in the response. Gas mileage is a function.
Regression Analysis Deterministic model No chance of an error in calculating y for a given x Probabilistic model chance of an error First order linear.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
1 Experimental Statistics - week 11 Chapter 11: Linear Regression and Correlation.
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
The simple linear regression model and parameter estimation
Lecture 11: Simple Linear Regression
Regression and Correlation of Data Summary
EXCEL: Multiple Regression
Chapter 20 Linear and Multiple Regression
Simple Linear Regression
Ch12.1 Simple Linear Regression
Simple Linear Regression
Multiple Regression and Model Building
Relationship with one independent variable
Regression Analysis PhD Course.
Regression Analysis Week 4.
Simple Regression Mary M. Whiteside, PhD.
Simple Linear Regression
Least-Squares Regression
Prediction of new observations
Relationship with one independent variable
Least-Squares Regression
Simple Linear Regression
Linear Regression and Correlation
Example on the Concept of Regression . observation
Simple Linear Regression
Linear Regression and Correlation
3.2. SIMPLE LINEAR REGRESSION
Regression and Correlation of Data
Simple Linear Regression
Nazmus Saquib, PhD Head of Research Sulaiman AlRajhi Colleges
Decomposition of Sum of Squares
Presentation transcript:

Simple Linear Regression Often we want to understand the relationships among variables, e.g., SAT scores and college GPA car weight and gas mileage amount of a certain pollutant in wastewater and bacteria growth in local streams number of takeoffs and landings and degree of metal fatigue in aircraft structures Simplest relationship  Y = α + βx Y = response, dependent variable x = regressor, independent variable, predictor EGR 252 - Ch. 11

Example The owner of a small harness race track in Florida is interested in understanding the relationship between attendance at the track and the total amount bet each night. The data for a two-week period (10 racing nights) is as follows: Attendance, x Amount Bet ($000), Y 117 2.07 128 2.8 122 3.14 119 2.26 131 3.4 135 3.89 125 2.93 120 2.66 130 3.33 127 3.54 EGR 252 - Ch. 11

Estimating the Regression Coefficients Method of Least Squares Determine a and b (estimates for α and β) so that the sum of the squares of the residuals is minimized. Steps: Calculate b using and a using EGR 252 - Ch. 11

For Our Example b = _______________________________________ Night Attendance, x Amount Bet, Y xiyi xi2 1 117 2.07 242.19 13689 2 128 2.8 358.4 16384 3 122 3.14 383.08 14884 4 119 2.26 268.94 14161 5 131 3.4 445.4 17161 6 135 3.89 525.15 18225 7 125 2.93 366.25 15625 8 120 2.66 319.2 14400 9 130 3.33 432.9 16900 10 127 3.54 449.58 16129 TOTAL 1254 30.02 3791.09 157558 b =((10*3791.09)-(1254*30.02))/((10*157558)-1254^2) = 0.086755875 a = (30.02/10) – 0.0868*(1254/10) = -7.877186684 b = _______________________________________ a = ______________________________ EGR 252 - Ch. 11

What does this mean? We can draw the regression line that describes the relationship between attendance and amount bet: We can also predict amount bet based on attendance. EGR 252 - Ch. 11

How good is our prediction? Estimating the variance: Coefficient of determination, R2 a measure of the “quality of fit,” or the proportion of the variability explained by the fitted model. SSE = sum(residuals2)= 0.639015 s2 = SSE/8 = 0.079876917 SST = Σ(Yi - Y)2 = 2.94516 R2 = 1-(SSE/SST) = 1-(0.639/2.945)=0.783029 (see next page) EGR 252 - Ch. 11

Calculations … Night Attendance, x Amount Bet, Y xiyi xi2 yhat residuals2 1 117 2.07 242.19 13689 2.27325 0.0413108 0.86862 2 119 2.26 268.94 14161 2.44676 0.0348802 0.55056 3 120 2.66 319.2 14400 2.53352 0.0159976 0.11696 4 122 3.14 383.08 14884 2.70703 0.187463 0.01904 5 125 2.93 366.25 15625 2.9673 0.0013911 0.00518 6 127 3.54 449.58 16129 3.14081 0.1593531 0.28944 7 128 2.8 358.4 16384 3.22757 0.1828121 0.0408 8 130 3.33 432.9 16900 3.40108 0.0050519 0.10758 9 131 3.4 445.4 17161 3.48783 0.0077146 0.1584 10 135 3.89 525.15 18225 3.83486 0.0030408 0.78854 TOTAL 1254 30.02 3791.09 157558 0.6390153 2.94516 (Y - Y)2 EGR 252 - Ch. 11

Or … Using Excel EGR 252 - Ch. 11 yhat = intercept + “variable name”*x standard error = std dev of the residuals R2 = proportion of the variation in y that is explained by the regression line Note the confidence interval … we can also draw a confidence interval around our predictions. EGR 252 - Ch. 11