Section 3.2: Least Squares Regressions

Slides:



Advertisements
Similar presentations
Least-Squares Regression Section 3.3. Correlation measures the strength and direction of a linear relationship between two variables. How do we summarize.
Advertisements

AP Statistics Section 3.2 C Coefficient of Determination
Least Squares Regression
AP Statistics.  Least Squares regression is a way of finding a line that summarizes the relationship between two variables.
CHAPTER 3 Describing Relationships
Regression, Residuals, and Coefficient of Determination Section 3.2.
C HAPTER 3: E XAMINING R ELATIONSHIPS. S ECTION 3.3: L EAST -S QUARES R EGRESSION Correlation measures the strength and direction of the linear relationship.
Least-Squares Regression Section 3.3. Why Create a Model? There are two reasons to create a mathematical model for a set of bivariate data. To predict.
AP STATISTICS LESSON 3 – 3 LEAST – SQUARES REGRESSION.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 3 Describing Relationships 3.2 Least-Squares.
Examining Bivariate Data Unit 3 – Statistics. Some Vocabulary Response aka Dependent Variable –Measures an outcome of a study Explanatory aka Independent.
Residuals Recall that the vertical distances from the points to the least-squares regression line are as small as possible.  Because those vertical distances.
Response Variable: measures the outcome of a study (aka Dependent Variable) Explanatory Variable: helps explain or influences the change in the response.
Describing Bivariate Relationships. Bivariate Relationships When exploring/describing a bivariate (x,y) relationship: Determine the Explanatory and Response.
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
Unit 4 LSRL.
LSRL.
Chapter 4.2 Notes LSRL.
Least Squares Regression Line.
Sections Review.
Statistics 101 Chapter 3 Section 3.
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Chapter 5 LSRL.
LSRL Least Squares Regression Line
Chapter 3.2 LSRL.
AP Stats: 3.3 Least-Squares Regression Line
Least-Squares Regression
Least Squares Regression Line LSRL Chapter 7-continued
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Least-Squares Regression
^ y = a + bx Stats Chapter 5 - Least Squares Regression
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Least Squares Regression
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 5 LSRL.
Chapter 5 LSRL.
Chapter 5 LSRL.
Least-Squares Regression
Least-Squares Regression
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
3.2 – Least Squares Regression
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
A medical researcher wishes to determine how the dosage (in mg) of a drug affects the heart rate of the patient. Find the correlation coefficient & interpret.
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
Chapter 3: Describing Relationships
9/27/ A Least-Squares Regression.
Chapter 3: Describing Relationships
Correlation/Regression - part 2
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
Presentation transcript:

Section 3.2: Least Squares Regressions Correlation measures the direction and strength of a LINEAR relationship. A line that summarizes the overall pattern of the relationship is called a regression line. Only applicable to bivariate data with explanatory and response variables. The line MODELS the form of the data and can be used to PREDICT response values (y) given an explanatory value (x).

General Equation of any Regression Line y = a + bx y-intercept the value of y when x = 0 …if appropriate for the data slope the approximate amount by which y changes when x increases by one unit

Non exercise activity (NEA) vs. Fat gain This data measure calories burned in everyday activity (non-exercise) compared to the fat gained (in kilograms) Fat gain = a + b(NEA) Predicting values beyond the scale is called extrapolation…these predictions are not always accurate

A good regression line best models the data by MINIMIZING the vertical distance of each point from the line. This line is called the LEAST SQUARES REGRESSION LINE (or LSRL) This is symbolized as where and the line passes through

Using Calculator to find the LSRL: STAT  CALC  #8 LinReg (a +bx) Listx, Listy, Y1 This will store the LSRL for graphing and further analysis

Residual = observed y – predicted y How well does the LSRL model the data? We use RESIDUALS. Residual = observed y – predicted y or Calculator: Must have LSRL stored in Y1 L3 = LRESID Residual for this data value

To investigate all the residuals of the data graphically, we can use a RESIDUAL PLOT. This is a scatterplot of the explanatory variable (L1) vs. residual values (L3). Be sure that Y1 (the LSRL) and any other plot is turned off. You are looking that the points are closely clustered about the x-axis.

To numerically analyze the residuals (and in turn, how well the LSRL models the data) there are two calculations we can use: Standard deviation of the residuals: this will give you a “margin of error” for any predicted value based upon the LSRL. 2) Coefficient of determination (r2): this will give the proportion of data values that are explained by (fit the model of) the LSRL. On Calculator: Do 1 Var Stat on residuals list and find Sx On Calculator: This is listed with the Lin Reg output

Further interpreting the r2 value… If r2 = 1, all data values fall on the LSRL…the LSRL accurately represents the data If r2 = 0, The LSRL does not predict values any better than the mean of the response variable