Lecture 8 Analysis of Variance and Covariance. 16-2 Effect of Coupons, In-Store Promotion and Affluence of the Clientele on Sales.

Slides:



Advertisements
Similar presentations
Analysis of Variance and Covariance Chapter Outline 1)Overview 2)Relationship Among Techniques 3) One-Way Analysis of Variance 4)Statistics Associated.
Advertisements

Analysis of Variance and Covariance. Chapter Outline 1) Overview 2) Relationship Among Techniques 3) One-Way Analysis of Variance 4) Statistics Associated.
ANOVA and Linear Models. Data Data is from the University of York project on variation in British liquids. Data is from the University of York project.
Analysis of variance (ANOVA)-the General Linear Model (GLM)
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 12: Analysis of Variance: Differences among Means of Three or More Groups.
Comparing k Populations Means – One way Analysis of Variance (ANOVA)
Chapter 17 Overview of Multivariate Analysis Methods
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
14-1 Statistically Adjusting the Data Variable Respecification Variable respecification involves the transformation of data to create new variables or.
Part I – MULTIVARIATE ANALYSIS
Comparing Means.
Correlation Patterns. Correlation Coefficient A statistical measure of the covariation or association between two variables. Are dollar sales.
Lecture 9: One Way ANOVA Between Subjects
Two Groups Too Many? Try Analysis of Variance (ANOVA)
8. ANALYSIS OF VARIANCE 8.1 Elements of a Designed Experiment
Topic 3: Regression.
Analysis of Variance & Multivariate Analysis of Variance
Comparing Means.
Today Concepts underlying inferential statistics
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Chapter 14 Inferential Data Analysis
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Leedy and Ormrod Ch. 11 Gray Ch. 14
Example of Simple and Multiple Regression
QNT 531 Advanced Problems in Statistics and Research Methods
Comparing Means. Anova F-test can be used to determine whether the expected responses at the t levels of an experimental factor differ from each other.
Chapter Sixteen Analysis of Variance and Covariance 16-1 © 2007 Prentice Hall.
Chapter Sixteen Analysis of Variance and Covariance.
Chapter Sixteen Analysis of Variance and Covariance.
Chapter Sixteen Analysis of Variance and Covariance 16-1 Copyright © 2010 Pearson Education, Inc.
بسم الله الرحمن الرحیم.. Multivariate Analysis of Variance.
1 Multivariate Analysis (Source: W.G Zikmund, B.J Babin, J.C Carr and M. Griffin, Business Research Methods, 8th Edition, U.S, South-Western Cengage Learning,
t(ea) for Two: Test between the Means of Different Groups When you want to know if there is a ‘difference’ between the two groups in the mean Use “t-test”.
© Copyright McGraw-Hill CHAPTER 12 Analysis of Variance (ANOVA)
Chapter 10 Analysis of Variance.
ANOVA (Analysis of Variance) by Aziza Munir
Psychology 301 Chapters & Differences Between Two Means Introduction to Analysis of Variance Multiple Comparisons.
Examining Relationships in Quantitative Research
Inferential Statistics
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
CHAPTER 11 SECTION 2 Inference for Relationships.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Chapter 16 Data Analysis: Testing for Associations.
ANOVA: Analysis of Variance.
Analysis of Variance and Covariance Effect of Coupons, In-Store Promotion and Affluence of the Clientele on Sales.
Chapter Seventeen. Figure 17.1 Relationship of Hypothesis Testing Related to Differences to the Previous Chapter and the Marketing Research Process Focus.
Chapter Seventeen Analysis of Variance and Covariance.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: One-way ANOVA Marshall University Genomics Core.
MARKETING RESEARCH CHAPTER 17: Hypothesis Testing Related to Differences.
ANOVA P OST ANOVA TEST 541 PHL By… Asma Al-Oneazi Supervised by… Dr. Amal Fatani King Saud University Pharmacy College Pharmacology Department.
Correlation & Regression Analysis
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Lecture Slides Elementary Statistics Eleventh Edition and the Triola.
Kin 304 Inferential Statistics Probability Level for Acceptance Type I and II Errors One and Two-Tailed tests Critical value of the test statistic “Statistics.
Analysis of Variance and Covariance
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 11 Testing for Differences Differences betweens groups or categories of the independent.
Analysis of Variance and Covariance Chapter Outline 1)Overview 2)Relationship Among Techniques 3) One-Way Analysis of Variance 4)Statistics Associated.
SUMMARY EQT 271 MADAM SITI AISYAH ZAKARIA SEMESTER /2015.
Analysis of Variance and Covariance Chapter Outline 1)Overview 2)Relationship Among Techniques 3) One-Way Analysis of Variance 4)Statistics Associated.
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Lecture Slides Elementary Statistics Tenth Edition and the.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Chapter Sixteen Analysis of Variance and Covariance 16-1 Copyright © 2010 Pearson Education, Inc.
Posthoc Comparisons finding the differences. Statistical Significance What does a statistically significant F statistic, in a Oneway ANOVA, tell us? What.
UNIT 4-B: DATA ANALYSIS and REPORTING
Analysis of Variance and Covariance
Analysis of Variance and Covariance
Analysis of Variance Correlation and Regression Analysis
MOHAMMAD NAZMUL HUQ, Assistant Professor, Department of Business Administration. Chapter-16: Analysis of Variance and Covariance Relationship among techniques.
Comparing Means.
Chapter 9: Differences among Groups
Product moment correlation
Presentation transcript:

Lecture 8 Analysis of Variance and Covariance

16-2 Effect of Coupons, In-Store Promotion and Affluence of the Clientele on Sales

16-3 Relationship Among Techniques Analysis of variance (ANOVA) is used as a test of means for two or more populations. The null hypothesis, typically, is that all means are equal. Analysis of variance must have a dependent variable that is metric (measured using an interval or ratio scale). There must also be one or more independent variables that are all categorical (nonmetric). Categorical independent variables are also called factors.

16-4 Relationship Among Techniques A particular combination of factor levels, or categories, is called a treatment. One-way analysis of variance involves only one categorical variable, or a single factor. In one-way analysis of variance, a treatment is the same as a factor level. If two or more factors are involved, the analysis is termed n-way analysis of variance. If the set of independent variables consists of both categorical and metric variables, the technique is called analysis of covariance (ANCOVA). In this case, the categorical independent variables are still referred to as factors, whereas the metric- independent variables are referred to as covariates.

16-5 Relationship Amongst t-Test, Analysis of Variance, Analysis of Covariance, & Regression One IndependentOne or More Metric Dependent Variable t Test Binary Variable One-Way Analysis of Variance One Factor N-Way Analysis of Variance More than One Factor Analysis of Variance Categorical (factors) Analysis of Covariance Categorical (f-rs) Interval covariates Regression Metric Independent Variables

16-6 One-way Analysis of Variance Business researchers are often interested in examining the differences in the mean values of the dependent variable for several categories of a single independent variable or factor. For example: Do the various segments differ in terms of their volume of product consumption? Do the brand evaluations of groups exposed to different commercials vary? What is the effect of consumers' familiarity with the store (measured as high, medium, and low) on preference for the store?

16-7 Statistics Associated with One-way Analysis of Variance eta 2 ( 2 ). The strength of the effects of X (independent variable or factor) on Y (dependent variable) is measured by eta 2 ( 2 ). The value of 2 varies between 0 and 1. F statistic. The null hypothesis that the category means are equal in the population is tested by an F statistic based on the ratio of mean square related to X and mean square related to error. Mean square. This is the sum of squares divided by the appropriate degrees of freedom.

16-8 Conducting One-way Analysis of Variance Interpret the Results If the null hypothesis of equal category means is not rejected, then the independent variable does not have a significant effect on the dependent variable. On the other hand, if the null hypothesis is rejected, then the effect of the independent variable is significant. A comparison of the category mean values will indicate the nature of the effect of the independent variable.

16-9 Illustrative Applications of One-way Analysis of Variance We illustrate the concepts discussed in this chapter using the data presented in Table The department store is attempting to determine the effect of in-store promotion (X) on sales (Y). For the purpose of illustrating hand calculations, the data of Table 16.2 are transformed in Table 16.3 to show the store sales (Y ij ) for each level of promotion. The null hypothesis is that the category means are equal: H 0 : µ 1 = µ 2 = µ 3.

16-10 Effect of Promotion and Clientele on Sales Table 16.2

16-11 One-Way ANOVA: Effect of In-store Promotion on Store Sales Table 16.3 Cell means Level of CountMean Promotion High (1) Medium (2) Low (3) TOTAL Source of Sum ofdfMean F ratio F prob. Variationsquaressquare Between groups (Promotion) Within groups (Error) TOTAL

16-12 N-way Analysis of Variance In business research, one is often concerned with the effect of more than one factor simultaneously. For example: How do advertising levels (high, medium, and low) interact with price levels (high, medium, and low) to influence a brand's sale? Do educational levels (less than high school, high school graduate, some college, and college graduate) and age (less than 35, 35-55, more than 55) affect consumption of a brand? What is the effect of consumers' familiarity with a department store (high, medium, and low) and store image (positive, neutral, and negative) on preference for the store?

16-13 Two-way Analysis of Variance Source ofSum ofMean Sig. of Variationsquares dfsquare F F  Main Effects Promotion Coupon Combined Two-way interaction Model Residual (error) TOTAL Table 16.4

16-14 Two-way Analysis of Variance Table 16.4 cont. Cell Means PromotionCoupon Count Mean High Yes High No Medium Yes Medium No Low Yes Low No TOTAL 30 Factor Level Means PromotionCoupon Count Mean High Medium Low Yes No Grand Mean

16-15 Analysis of Covariance When examining the differences in the mean values of the dependent variable related to the effect of the controlled independent variables, it is often necessary to take into account the influence of uncontrolled (usually metric) independent variables. For example: In determining how different groups exposed to different commercials evaluate a brand, it may be necessary to control for prior knowledge. In determining how different price levels will affect a household's cereal consumption, it may be essential to take household size into account. We again use the data of Table 16.2 to illustrate analysis of covariance. Suppose that we wanted to determine the effect of in-store promotion and couponing on sales while controlling for the effect of clientele. The results are shown in Table 16.6.

16-16 Analysis of Covariance Sum ofMeanSig. Source of Variation SquaresdfSquare Fof F Covariance Clientele Main effects Promotion Coupon Combined Way Interaction Promotion* Coupon Model Residual (Error) TOTAL CovariateRaw Coefficient Clientele Table 16.5

16-17 Issues in Interpretation Multiple Comparisons If the null hypothesis of equal means is rejected, we can only conclude that not all of the group means are equal. We may wish to examine differences among specific means. This can be done by specifying appropriate contrasts, or comparisons used to determine which of the means are statistically different. A priori contrasts are determined before conducting the analysis, based on the researcher's theoretical framework. Generally, a priori contrasts are used in lieu of the ANOVA F test. The contrasts selected are orthogonal (they are independent in a statistical sense).

16-18 Issues in Interpretation Multiple Comparisons A posteriori contrasts are made after the analysis. These are generally multiple comparison tests. They enable the researcher to construct generalized confidence intervals that can be used to make pairwise comparisons of all treatment means. These tests, listed in order of decreasing power, include least significant difference, Duncan's multiple range test, Student-Newman-Keuls, Tukey's alternate procedure, honestly significant difference, modified least significant difference, and Scheffe's test. Of these tests, least significant difference is the most powerful, Scheffe's the most conservative.

16-19 Multivariate Analysis of Variance Multivariate analysis of variance (MANOVA) is similar to analysis of variance (ANOVA), except that instead of one metric dependent variable, we have two or more. In MANOVA, the null hypothesis is that the vectors of means on multiple dependent variables are equal across groups. Multivariate analysis of variance is appropriate when there are two or more dependent variables that are correlated. If they are uncorrelated, use ANOVA on each of the dependent variables separately rather than MANOVA.

16-20 SPSS Windows One-way ANOVA can be efficiently performed using the program COMPARE MEANS and then One-way ANOVA. To select this procedure using SPSS for Windows click: Analyze>Compare Means>One-Way ANOVA … N-way analysis of variance and analysis of covariance can be performed using GENERAL LINEAR MODEL. To select this procedure using SPSS for Windows click: Analyze>General Linear Model>Univariate …