Download presentation
Presentation is loading. Please wait.
Published byJacob Roberts Modified over 6 years ago
1
Analysis of Variance Correlation and Regression Analysis
Chapter 16 & 17 Analysis of Variance Correlation and Regression Analysis Copyright © 2010 Pearson Education, Inc. 16-1
2
Chapter Outline Overview Relationship Among Techniques
One-Way Analysis of Variance Statistics Associated with One-Way Analysis of Variance Conducting One-Way Analysis of Variance Identification of Dependent & Independent Variables Decomposition of the Total Variation Measurement of Effects Significance Testing Interpretation of Results
3
Chapter Outline 1) Overview 2) Product-Moment Correlation 3) Regression Analysis 4) Bivariate Regression 5) Statistics Associated with Bivariate Regression Analysis 6) Conducting Bivariate Regression Analysis i. Scatter Diagram ii. Bivariate Regression Model 7) Multiple Regression
4
Relationship Among Techniques
Analysis of variance (ANOVA) is used as a test of means for two or more populations. The null hypothesis, typically, is that all means are equal. Analysis of variance must have a dependent variable that is metric (measured using an interval or ratio scale). There must also be one or more independent variables that are all categorical (nonmetric). Categorical independent variables are also called factors.
5
One-Way Analysis of Variance
Marketing researchers are often interested in examining the differences in the mean values of the dependent variable for several categories of a single independent variable or factor. For example: Do the various segments differ in terms of their volume of product consumption? Do the brand evaluations of groups exposed to different commercials vary? What is the effect of consumers' familiarity with the store (measured as high, medium, and low) on preference for the store?
6
Statistics Associated with One-Way Analysis of Variance
F statistic. The null hypothesis that the category means are equal in the population is tested by an F statistic based on the ratio of mean square related to X and mean square related to error. Mean square. This is the sum of squares divided by the appropriate degrees of freedom.
7
Conducting One-Way Analysis of Variance Interpret the Results
If the null hypothesis of equal category means is not rejected, then the independent variable does not have a significant effect on the dependent variable. On the other hand, if the null hypothesis is rejected, then the effect of the independent variable is significant. A comparison of the category mean values will indicate the nature of the effect of the independent variable.
8
Illustrative Applications of One-Way Analysis of Variance
We illustrate the concepts discussed in this chapter using the data presented in Table The department store is attempting to determine the effect of in-store promotion (X) on sales (Y). For the purpose of illustrating hand calculations, the data of Table 16.2 are transformed in Table 16.3 to show the store sales (Yij) for each level of promotion. The null hypothesis is that the category means are equal: H0: µ1 = µ2 = µ3
9
Effect of Promotion and Clientele on Sales
Table 16.2
10
One-Way ANOVA: Effect of In-Store Promotion on Store Sales
Table 16.4 Cell means Level of Count Mean Promotion High (1) Medium (2) Low (3) TOTAL Source of Sum of df Mean F ratio F prob. Variation squares square Between groups (Promotion) Within groups (Error) TOTAL
11
Multivariate Analysis of Variance
Multivariate analysis of variance (MANOVA) is similar to analysis of variance (ANOVA), except that instead of one metric dependent variable, we have two or more. In MANOVA, the null hypothesis is that the vectors of means on multiple dependent variables are equal across groups. Multivariate analysis of variance is appropriate when there are two or more dependent variables that are correlated.
12
SPSS Windows: One-Way ANOVA
Select ANALYZE from the SPSS menu bar. Click COMPARE MEANS and then ONE-WAY ANOVA. Move “Sales [sales]” in to the DEPENDENT LIST box. Move “In-Store Promotion[promotion]” to the FACTOR box. Click OPTIONS. Click Descriptive. Click CONTINUE. Click OK.
13
Correlation The product moment correlation, r, summarizes the strength of association between two metric (interval or ratio scaled) variables, say X and Y. It is an index used to determine whether a linear or straight- line relationship exists between X and Y. As it was originally proposed by Karl Pearson, it is also known as the Pearson correlation coefficient. It is also referred to as simple correlation, bivariate correlation, or merely the correlation coefficient.
14
Product Moment Correlation
From a sample of n observations, X and Y, the product moment correlation, r, can be calculated as: r = ( X i - ) Y S 1 n 2 D v s o f t h e u m a d b y g C O V x
15
Product Moment Correlation
r varies between -1.0 and +1.0. The correlation coefficient between two variables will be the same regardless of their underlying units of measurement.
16
Explaining Attitude Toward the City of Residence
Table 17.1
17
Product Moment Correlation
The correlation coefficient may be calculated as follows: = ( )/12 = 9.333 X Y = ( )/12 = 6.583 ( i - ) S = 1 n = ( )(6-6.58) + ( )(9-6.58) + ( )(8-6.58) + (4-9.33)(3-6.58) + ( )( ) + (6-9.33)(4-6.58) + (8-9.33)(5-6.58) + (2-9.33) (2-6.58) + ( )( ) + (9-9.33)(9-6.58) + ( )( ) + (2-9.33)(2-6.58) = =
18
Product Moment Correlation
( X i - ) 2 S = 1 n = ( )2 + ( )2 + ( )2 + (4-9.33)2 + ( )2 + (6-9.33)2 + (8-9.33)2 + (2-9.33)2 + ( )2 + (9-9.33)2 + ( )2 + (2-9.33)2 = = Y = (6-6.58)2 + (9-6.58)2 + (8-6.58)2 + (3-6.58)2 + ( )2+ (4-6.58)2 + (5-6.58)2 + (2-6.58)2 + ( )2 + (9-6.58)2 + ( )2 + (2-6.58)2 = = Thus, r 7 9 . 6 8 3 4 =
19
Correlation Analysis Pearson Correlation Coefficient–statistical measure of the strength of a linear relationship between two metric variables Varies between – 1.00 and +1.00 The higher the correlation coefficient–the stronger the level of association Correlation coefficient can be either positive or negative
20
Strength of Correlation Coefficients
21
SPSS Pearson Correlation Example
22
Regression Analysis Regression analysis examines associative relationships between a metric dependent variable and one or more independent variables in the following ways: Determine whether the independent variables explain a significant variation in the dependent variable: whether a relationship exists. Determine how much of the variation in the dependent variable can be explained by the independent variables: strength of the relationship. Determine the structure or form of the relationship: the mathematical equation relating the independent and dependent variables. Predict the values of the dependent variable. Control for other independent variables when evaluating the contributions of a specific variable or set of variables. Regression analysis is concerned with the nature and degree of association between variables and does not imply or assume any causality.
23
Relationships between Variables
Is there a relationship between the two variables we are interested in? How strong is the relationship? How can that relationship be best described?
24
Conducting Bivariate Regression Analysis Formulate the Bivariate Regression Model
In the bivariate regression model, the general form of a straight line is: Y = X b + 1 where Y = dependent or criterion variable X = independent or predictor variable = intercept of the line = slope of the line The regression procedure adds an error term to account for the probabilistic or stochastic nature of the relationship: Yi = Xi + ei where ei is the error term associated with the i th observation.
25
Covariation and Variable Relationships
First we should understand the covariation between variables Covariation is amount of change in one variable that is consistently related to the change in another variable A scatter diagram graphically plots the relative position of two variables using a horizontal and a vertical axis to represent the variable values
26
Plot of Attitude with Duration
Fig. 17.3 4.5 2.25 6.75 11.25 9 13.5 3 6 15.75 18 Duration of Residence Attitude
27
Scatter Diagram Illustrates No Relationship
28
Positive Relationship between X and Y
29
Negative Relationship between X and Y
30
Curvilinear Relationship between X and Y
31
Straight Line Relationship in Regression
32
Formula for a Straight Line
y = a bX + ei y = the dependent variable a = the intercept b = the slope X = the independent variable used to predict y ei = the error for the prediction
33
Ordinary Least Squares (OLS)
OLS is a statistical procedure that estimates regression equation coefficients which produce the lowest sum of squared differences between the actual and predicted values of the dependent variable
34
SPSS Results for Bivariate Regression
35
SPSS Results say... Percieved reasonableness of prices is positively related to overall customer satisfaction Th relationship is positive But weak! Prices and satisfaction is associated, but there are other factors as well!!
36
Multiple Regression Analysis
Multiple regression analysis is a statistical technique which analyzes the linear relationship between a dependent variable and multiple independent variables by estimating coefficients for the equation for a straight line
37
Y = b + X . e Multiple Regression
The general form of the multiple regression model is as follows: which is estimated by the following equation: = a + b1X1 + b2X2 + b3X bkXk As before, the coefficient a represents the intercept, but the b's are now the partial regression coefficients. Y = b + 1 X 2 3 . k e
38
Statistics Associated with Multiple Regression
Adjusted R2. R2, coefficient of multiple determination, is adjusted for the number of independent variables and the sample size to account for the diminishing returns. After the first few variables, the additional independent variables do not make much contribution. Coefficient of multiple determination. The strength of association in multiple regression is measured by the square of the multiple correlation coefficient, R2, which is also called the coefficient of multiple determination. F test. The F test is used to test the null hypothesis that the coefficient of multiple determination in the population, R2pop, is zero. This is equivalent to testing the null hypothesis. The test statistic has an F distribution with k and (n - k - 1) degrees of freedom.
39
VARIABLES IN THE EQUATION
Multiple Regression Table 17.3 Multiple R R Adjusted R Standard Error ANALYSIS OF VARIANCE df Sum of Squares Mean Square Regression Residual F = Significance of F = VARIABLES IN THE EQUATION Variable b SEb Beta (ß) T Significance of T IMPORTANCE DURATION (Constant)
40
SPSS Windows The CORRELATE program computes Pearson product moment correlations and partial correlations with significance levels. Univariate statistics, covariance, and cross-product deviations may also be requested. Significance levels are included in the output. To select these procedures using SPSS for Windows, click: Analyze>Correlate>Bivariate … Analyze>Correlate>Partial … Scatterplots can be obtained by clicking: Graphs>Scatter >Simple>Define … REGRESSION calculates bivariate and multiple regression equations, associated statistics, and plots. It allows for an easy examination of residuals. This procedure can be run by clicking: Analyze>Regression Linear …
41
SPSS Windows: Correlations
Select ANALYZE from the SPSS menu bar. Click CORRELATE and then BIVARIATE. Move “Attitude[attitude]” into the VARIABLES box. Then move “Duration[duration]” into the VARIABLES box. Check PEARSON under CORRELATION COEFFICIENTS. Check ONE-TAILED under TEST OF SIGNIFICANCE. Check FLAG SIGNIFICANT CORRELATIONS. Click OK.
42
SPSS Windows: Bivariate Regression
Select ANALYZE from the SPSS menu bar. Click REGRESSION and then LINEAR. Move “Attitude[attitude]” into the DEPENDENT box. Move “Duration[duration]” into the INDEPENDENT(S) box. Select ENTER in the METHOD box. Click on STATISTICS and check ESTIMATES under REGRESSION COEFFICIENTS. Check MODEL FIT. Click CONTINUE. Click OK.
43
Copyright © 2010 Pearson Education, Inc.
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior written permission of the publisher. Printed in the United States of America. Copyright © 2010 Pearson Education, Inc.
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.