Basic Statistics Linear Regression. X Y Simple Linear Regression.

Slides:



Advertisements
Similar presentations
Lesson 10: Linear Regression and Correlation
Advertisements

Section 10-3 Regression.
Statistical Techniques I EXST7005 Simple Linear Regression.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Regression Greg C Elvers.
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
Chapter 10 Regression. Defining Regression Simple linear regression features one independent variable and one dependent variable, as in correlation the.
Scatter Diagrams and Linear Correlation
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
© The McGraw-Hill Companies, Inc., 2000 CorrelationandRegression Further Mathematics - CORE.
Regression Regression: Mathematical method for determining the best equation that reproduces a data set Linear Regression: Regression method applied with.
Chapter 15 (Ch. 13 in 2nd Can.) Association Between Variables Measured at the Interval-Ratio Level: Bivariate Correlation and Regression.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
CORRELATON & REGRESSION
SIMPLE LINEAR REGRESSION
SIMPLE LINEAR REGRESSION
Business Statistics - QBM117 Least squares regression.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
Chapter 2 – Simple Linear Regression - How. Here is a perfect scenario of what we want reality to look like for simple linear regression. Our two variables.
Least Squares Regression Line (LSRL)
Correlation and Linear Regression
Correlation and Linear Regression
McGraw-Hill/Irwin Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 13 Linear Regression and Correlation.
Correlation and Linear Regression Chapter 13 Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Lecture 3-2 Summarizing Relationships among variables ©
Linear Regression.
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
Relationship of two variables
Linear Regression and Correlation
© The McGraw-Hill Companies, Inc., 2000 Business and Finance College Principles of Statistics Lecture 10 aaed EL Rabai week
© 2008 Pearson Addison-Wesley. All rights reserved Chapter 1 Section 13-6 Regression and Correlation.
1 Dr. Jerrell T. Stracener EMIS 7370 STAT 5340 Probability and Statistics for Scientists and Engineers Department of Engineering Management, Information.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Basic Statistics Correlation Var Relationships Associations.
McGraw-Hill/Irwin Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 13 Linear Regression and Correlation.
Linear Regression Handbook Chapter. Experimental Testing Data are collected, in scientific experiments, to test the relationship between various measurable.
Part IV Significantly Different: Using Inferential Statistics
Welcome to MM207 Unit 8 Seminar Dr. R. Correlation, Regression and Excel!
Basic Concepts of Correlation. Definition A correlation exists between two variables when the values of one are somehow associated with the values of.
Regression Regression relationship = trend + scatter
STATISTICS 12.0 Correlation and Linear Regression “Correlation and Linear Regression -”Causal Forecasting Method.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Welcome to MM570 Psychological Statistics Unit 4 Seminar Dr. Srabasti Dutta.
Chapter 6 (cont.) Difference Estimation. Recall the Regression Estimation Procedure 2.
Chapter 9: Correlation and Regression Analysis. Correlation Correlation is a numerical way to measure the strength and direction of a linear association.
Scatter Diagrams scatter plot scatter diagram A scatter plot is a graph that may be used to represent the relationship between two variables. Also referred.
Chapter 10: Determining How Costs Behave 1 Horngren 13e.
Economics 173 Business Statistics Lecture 10 Fall, 2001 Professor J. Petry
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
© 2001 Prentice-Hall, Inc.Chap 13-1 BA 201 Lecture 18 Introduction to Simple Linear Regression (Data)Data.
1 Simple Linear Regression and Correlation Least Squares Method The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES.
STATISTICS 12.0 Correlation and Linear Regression “Correlation and Linear Regression -”Causal Forecasting Method.
Welcome to MM570 Psychological Statistics Unit 4 Seminar Dr. Bob Lockwood.
Chapters 8 Linear Regression. Correlation and Regression Correlation = linear relationship between two variables. Summarize relationship with line. Called.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
PreCalculus 1-7 Linear Models. Our goal is to create a scatter plot to look for a mathematical correlation to this data.
BUSINESS MATHEMATICS & STATISTICS. Module 6 Correlation ( Lecture 28-29) Line Fitting ( Lectures 30-31) Time Series and Exponential Smoothing ( Lectures.
Linear Regression Essentials Line Basics y = mx + b vs. Definitions
Linear Regression.
Ch12.1 Simple Linear Regression
Correlation and Simple Linear Regression
Lecture Notes The Relation between Two Variables Q Q
Equations of Lines and Modeling
Correlation and Simple Linear Regression
Correlation and Regression
11C Line of Best Fit By Eye, 11D Linear Regression
Ch 4.1 & 4.2 Two dimensions concept
Chapter 14 Multiple Regression
Presentation transcript:

Basic Statistics Linear Regression

X Y Simple Linear Regression

Predicting Y from X Recall when we looked at scatter plots in our discussion of correlation, we showed generally the estimate of Y given a value for X, when the correlation was not perfect. We will now look at how to use our knowledge of the correlation to predict a value for Y, when we know a value for X.

Variable X Variable Y The GREEN line shows our prediction or regression line. high low Scatter Plot of Y and X Estimated Y value

Prediction Equation The green line in the previous slide showed us our prediction line. We will use the mathematical formula for a straight line as the method for predicting a value for Y when we know the value for X. The process is called “Linear Regression” because, in this class, we will only deal with relationships that can be fitted by a straight line. The general formula for a straight line is:

The Prediction Equation a y = the intercept or where the prediction line crosses the Y-axis (the value of Y when X = 0) b y = the regression coefficient that indicates the amount of change in Y when the value of X increases one unit.

A Simple Example Suppose that a club charges a flat $25 to use their facilities. They also charge a $10 fee per hour for using the tennis courts. Now, assume that you want to play tennis for 2 hours at this club. How much would you have to pay? Ŷ= $25 + (2) $10 = $25 + $20 = $45 for two hours of tennis

Linking the Simple Example to Regression Ŷ= $25 + (2) $10 = $25 + $20 = $45 for two hours of tennis In our example: –$25 is a y, the intercept. Even if we didn’t play any tennis (X = 0), it would cost $25 to use the club. –$10 is b y, the regression coefficient (it costs $10 for each hour of tennis played) In this case we predicted how much it would cost (Y) when we knew how long we wanted to play tennis.

Formulae for Sums of Squares These were introduced in our discussion of correlation.

Calculating the Regression Coefficient (b) or

Calculating the Intercept (a) You will notice that you must calculate the regression coefficient (b) before you can calculate the intercept (a), since the calculation of a uses b.

An Example From our earlier example, suppose that our college statistics professor is interested in predicting how many errors students might make on the mid-term examination based on how many hours they studied. Specifically, the professor wants to know how many errors a student might make if the student studied for 5 hours.

The Stats Professor’s Data StudentXYX2X2 Y2Y2 XY Total  X = 70  Y = 73  X 2 =546  Y2=695  XY=429

The Resulting Sum of Squares StudentXYX2X2 Y2Y2 XY Total  X = 70  Y = 73  X 2 =546  Y2=695  XY=429 = /10 = = 56 = /10 = = = 429 – (70)(73)/10 = 429 – 511 = -82

Calculating the Regression Coefficient (b) = - 82 / 56 = This can be interpreted as the change in the value of Y (in our case, errors made on the mid-term), for a unit change in X, or for us, each additional hour studied! Thus, study for another hour and make 1.46 fewer mistakes (on average!).

Calculating the Intercept (a) = 7.3 – (-1.46)(7) = = Therefore, our prediction equation is Ŷ = (-1.46) (X)

Using Our Prediction Equation Ŷ = (-1.46) (X) If the professor wanted to predict the number of errors a student might make if the student had studied for 5 hours, then we would substitute 5 for X in the above equation and obtain: Ŷ = (-1.46) (5) = (-7.3) = Thus, the professor would predict errors for a student who had studied for 5 hours.

Measuring Prediction Errors: The Standard Error of the Estimate OR Since we know that the estimate is not exact, as statisticians, we must report how much error we feel is in our estimate. The formula is:

Calculating the Standard Error of the Estimate =  (162.1) / 8 = 2.29 Thus, when we estimated errors, we also would report that the Standard Error of the Estimate is 2.29.

Summarizing Prediction Equations The existence of a relationship between two variables allows us to use that knowledge to make predictions. The prediction based on our equation will result in less error in prediction than using the mean of the dependent variable. Two sums of squares are required to calculate the regression coefficient and the intercept.