Practice. Practice Practice Practice Practice r = X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4 (4) 72.

Slides:



Advertisements
Similar presentations
Sleeping and Happiness
Advertisements

Finding Slope – Intercept From an Equation STEPS : 1. Equation must be in y = mx + b form 2. The slope is the coefficient of “x” ( m ) 3. The y - intercept.
Kin 304 Regression Linear Regression Least Sum of Squares
Least Squares Regression
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
AP Statistics.  Least Squares regression is a way of finding a line that summarizes the relationship between two variables.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Chapter 10 Regression. Defining Regression Simple linear regression features one independent variable and one dependent variable, as in correlation the.
Regression Regression: Mathematical method for determining the best equation that reproduces a data set Linear Regression: Regression method applied with.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
LINEAR REGRESSION: What it Is and How it Works Overview What is Bivariate Linear Regression? The Regression Equation How It’s Based on r.
LINEAR REGRESSION: What it Is and How it Works. Overview What is Bivariate Linear Regression? The Regression Equation How It’s Based on r.
Ch 8: Correlation & Regression Wed, Feb 25 th, 2004.
Business 205. Review Chi-Square Preview Regression.
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
REGRESSION Predict future scores on Y based on measured scores on X Predictions are based on a correlation from a sample where both X and Y were measured.
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
1 1 Slide Simple Linear Regression Chapter 14 BA 303 – Spring 2011.
Regression Analysis What is regression ?What is regression ? Best-fit lineBest-fit line Least squareLeast square What is regression ?What is regression.
Section #6 November 13 th 2009 Regression. First, Review Scatter Plots A scatter plot (x, y) x y A scatter plot is a graph of the ordered pairs (x, y)
© The McGraw-Hill Companies, Inc., 2000 Business and Finance College Principles of Statistics Lecture 10 aaed EL Rabai week
Chapter 6 & 7 Linear Regression & Correlation
Finding Slope – Intercept From an Equation STEPS : 1. Equation must be in y = mx + b form 2. The slope is the coefficient of “x” ( m ) 3. The y - intercept.
Linear Regression James H. Steiger. Regression – The General Setup You have a set of data on two variables, X and Y, represented in a scatter plot. You.
Simple Linear Regression One reason for assessing correlation is to identify a variable that could be used to predict another variable If that is your.
Correlation is a statistical technique that describes the degree of relationship between two variables when you have bivariate data. A bivariate distribution.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
10B11PD311 Economics REGRESSION ANALYSIS. 10B11PD311 Economics Regression Techniques and Demand Estimation Some important questions before a firm are.
STATISTICS 12.0 Correlation and Linear Regression “Correlation and Linear Regression -”Causal Forecasting Method.
Practice You collect data from 53 females and find the correlation between candy and depression is Determine if this value is significantly different.
CHAPTER 5 CORRELATION & LINEAR REGRESSION. GOAL : Understand and interpret the terms dependent variable and independent variable. Draw a scatter diagram.
With the growth of internet service providers, a researcher decides to examine whether there is a correlation between cost of internet service per.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
LESSON 6: REGRESSION 2/21/12 EDUC 502: Introduction to Statistics.
Copyright © 2004 by The McGraw-Hill Companies, Inc. All rights reserved.
STATISTICS 12.0 Correlation and Linear Regression “Correlation and Linear Regression -”Causal Forecasting Method.
Method 3: Least squares regression. Another method for finding the equation of a straight line which is fitted to data is known as the method of least-squares.
OLS Regression What is it? Closely allied with correlation – interested in the strength of the linear relationship between two variables One variable is.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Linear Regression.
Regression Analysis AGEC 784.
LEAST – SQUARES REGRESSION
Statistics 101 Chapter 3 Section 3.
Linear Regression Special Topics.
Chapter 13 Linear Regression and Correlation Basic Statistics
Equations of Lines.
Writing an Equation of a Line
Review. Review Statistics Needed Need to find the best place to draw the regression line on a scatter plot Need to quantify the cluster.
Remember No Class on Wednesday No Class on Friday.
S519: Evaluation of Information Systems
Regression and Residual Plots
You can calculate: Central tendency Variability You could graph the data.
Lesson 5.3 How do you write linear equations in point-slope form?
AP Stats: 3.3 Least-Squares Regression Line
Section 4.2 How Can We Define the Relationship between two
LESSON 21: REGRESSION ANALYSIS
LEAST – SQUARES REGRESSION
3.5 Write and Graph Equations of Lines
Introduction to Probability and Statistics Thirteenth Edition
Quick Graphs of Linear Equations
Regression making predictions
Correlation and Regression
4.3 Graphing Equations of Lines From Intercepts
Multivariate Analysis Regression
Slope-intercept Form of Equations of Straight Lines
Sleeping and Happiness
Chapter Thirteen McGraw-Hill/Irwin
Correlation/Regression - part 2
Graphing using Slope-Intercept Form
Presentation transcript:

Practice

Practice

Practice

Practice r = X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4 (4) 72

Practice r = X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4 -92 20 (4) 120 (4) 123 20 19 X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4

Practice r = X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4 -92 20 80 131 X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4

Practice -.90 = X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4 -92 102.37 80 131 X = 20 X2 = 120 Y = 19 Y2 = 123 XY = 72 N = 4

Remember this: Statistics Needed Need to find the best place to draw the regression line on a scatter plot Need to quantify the cluster of scores around this regression line (i.e., the correlation coefficient)

Regression allows us to predict! . . . . .

Straight Line Y = mX + b Where: Y and X are variables representing scores m = slope of the line (constant) b = intercept of the line with the Y axis (constant)

Excel Example

That’s nice but. . . . How do you figure out the best values to use for m and b ? First lets move into the language of regression

Straight Line Y = mX + b Where: Y and X are variables representing scores m = slope of the line (constant) b = intercept of the line with the Y axis (constant)

Regression Equation Y = a + bX Where: Y = value predicted from a particular X value a = point at which the regression line intersects the Y axis b = slope of the regression line X = X value for which you wish to predict a Y value

Practice Y = -7 + 2X What is the slope and the Y-intercept? Determine the value of Y for each X: X = 1, X = 3, X = 5, X = 10

Practice Y = -7 + 2X What is the slope and the Y-intercept? Determine the value of Y for each X: X = 1, X = 3, X = 5, X = 10 Y = -5, Y = -1, Y = 3, Y = 13

Finding a and b Uses the least squares method Minimizes Error Error = Y - Y  (Y - Y)2 is minimized

. . . . .

. . . . . Error = Y - Y  (Y - Y)2 is minimized Error = 1 Error = .5

Finding a and b Ingredients r value between the two variables Sy and Sx Mean of Y and X

b = b r = correlation between X and Y SY = standard deviation of Y SX = standard deviation of X

a a = Y - bX Y = mean of the Y scores b = regression coefficient computed previously X = mean of the X scores

Mean Y = 4.6; SY = 2.41 r = .88 Mean X = 3.0; SX = 1.41

Mean Y = 4.6; SY = 2.41 r = .88 Mean X = 3.0; SX = 1.41

Mean Y = 4.6; SY = 2.41 r = .88 Mean X = 3.0; SX = 1.41 b =

Mean Y = 4.6; SY = 2.41 r = .88 Mean X = 3.0; SX = 1.41 b = .88 1.50 1.41

Mean Y = 4.6; SY = 2.41 r = .88 Mean X = 3.0; SX = 1.41 b = 1.5 a = Y - bX

Mean Y = 4.6; SY = 2.41 r = .88 Mean X = 3.0; SX = 1.41 b = 1.5 0.1 = 4.6 - (1.50)3.0

Regression Equation Y = a + bX Y = 0.1 + (1.5)X

Y = 0.1 + (1.5)X . . . . .

Y = 0.1 + (1.5)X X = 1; Y = 1.6 . . . . . .

Y = 0.1 + (1.5)X X = 5; Y = 7.60 . . . . . . .

Y = 0.1 + (1.5)X . . . . . . .

Practice

Mean Y = 14.50; Sy = 4.43 Mean X = 6.00; Sx= 2.16 r = -.57

Mean Y = 14.50; Sy = 4.43 Mean X = 6.00; Sx= 2.16 r = -.57 b =

Mean Y = 14.50; Sy = 4.43 Mean X = 6.00; Sx= 2.16 r = -.57 b = -.57 -1.17 2.16

Mean Y = 14.50; Sy = 4.43 Mean X = 6.00; Sx= 2.16 b = -1.17 a = Y - bX

Mean Y = 14.50; Sy = 4.43 Mean X = 6.00; Sx= 2.16 b = -1.17 21.52= 14.50 - (-1.17)6.0

Regression Equation Y = a + bX Y = 21.52 + (-1.17)X

Y = 21.52 + (-1.17)X . 22 20 . 18 16 . 14 . 12 10

Y = 21.52 + (-1.17)X . . 22 20 . 18 16 . 14 . 12 10

Y = 21.52 + (-1.17)X . . 22 20 . 18 16 . 14 . . 12 10

Y = 21.52 + (-1.17)X . . 22 20 . 18 16 . 14 . . 12 10