Multivariate Analysis. One-way ANOVA Tests the difference in the means of 2 or more nominal groups Tests the difference in the means of 2 or more nominal.

Slides:



Advertisements
Similar presentations
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Advertisements

General Linear Model Introduction to ANOVA.
Forecasting Using the Simple Linear Regression Model and Correlation
EPI 809/Spring Probability Distribution of Random Error.
Simple Linear Regression and Correlation
Definition  Regression Model  Regression Equation Y i =  0 +  1 X i ^ Given a collection of paired data, the regression equation algebraically describes.
Chapter 12 Simple Linear Regression
Chapter 13 Multiple Regression
ONE-WAY BETWEEN SUBJECTS ANOVA Also called One-Way Randomized ANOVA Purpose: Determine whether there is a difference among two or more groups – used mostly.
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Chapter 12 Simple Regression
Chapter 12 Multiple Regression
Statistics for Business and Economics
Chapter Topics Types of Regression Models
Ch. 14: The Multiple Regression Model building
Lecture 17 Interaction Plots Simple Linear Regression (Chapter ) Homework 4 due Friday. JMP instructions for question are actually for.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
Chapter 7 Forecasting with Simple Regression
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression. Introduction In Chapters 17 to 19, we examine the relationship between interval variables via a mathematical equation. The motivation.
Chapter 6 (cont.) Regression Estimation. Simple Linear Regression: review of least squares procedure 2.
Linear Regression/Correlation
Basic Analysis of Variance and the General Linear Model Psy 420 Andrew Ainsworth.
1 1 Slide Simple Linear Regression Chapter 14 BA 303 – Spring 2011.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
Regression and Correlation Methods Judy Zhong Ph.D.
Introduction to Linear Regression and Correlation Analysis
Section #6 November 13 th 2009 Regression. First, Review Scatter Plots A scatter plot (x, y) x y A scatter plot is a graph of the ordered pairs (x, y)
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Statistics for Business and Economics Chapter 10 Simple Linear Regression.
EQT 272 PROBABILITY AND STATISTICS
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
INTRODUCTORY LINEAR REGRESSION SIMPLE LINEAR REGRESSION - Curve fitting - Inferences about estimated parameter - Adequacy of the models - Linear.
© 2001 Prentice-Hall, Inc. Statistics for Business and Economics Simple Linear Regression Chapter 10.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
Statistical Methods Statistical Methods Descriptive Inferential
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Chapter 10: Analyzing Experimental Data Inferential statistics are used to determine whether the independent variable had an effect on the dependent variance.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1 Multiple Regression A single numerical response variable, Y. Multiple numerical explanatory variables, X 1, X 2,…, X k.
Regression Lesson 11. The General Linear Model n Relationship b/n predictor & outcome variables form straight line l Correlation, regression, t-tests,
Simple Linear Regression ANOVA for regression (10.2)
1 Regression & Correlation (1) 1.A relationship between 2 variables X and Y 2.The relationship seen as a straight line 3.Two problems 4.How can we tell.
Applied Quantitative Analysis and Practices LECTURE#25 By Dr. Osman Sadiq Paracha.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
© 2001 Prentice-Hall, Inc.Chap 13-1 BA 201 Lecture 18 Introduction to Simple Linear Regression (Data)Data.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Handout Eight: Two-Way Between- Subjects Design with Interaction- Assumptions, & Analyses EPSE 592 Experimental Designs and Analysis in Educational Research.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
Linear Regression Linear Regression. Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Purpose Understand Linear Regression. Use R functions.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 17 Simple Linear Regression and Correlation.
Formula for Linear Regression y = bx + a Y variable plotted on vertical axis. X variable plotted on horizontal axis. Slope or the change in y for every.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
BUSINESS MATHEMATICS & STATISTICS. Module 6 Correlation ( Lecture 28-29) Line Fitting ( Lectures 30-31) Time Series and Exponential Smoothing ( Lectures.
REGRESSION AND CORRELATION SIMPLE LINEAR REGRESSION 10.2 SCATTER DIAGRAM 10.3 GRAPHICAL METHOD FOR DETERMINING REGRESSION 10.4 LEAST SQUARE METHOD.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Chapter 14 Introduction to Multiple Regression
REGRESSION G&W p
Statistics for Managers using Microsoft Excel 3rd Edition
Chapter 11: Simple Linear Regression
Chapter 11: Simple Linear Regression
Chapter 11 Simple Regression
CHAPTER 29: Multiple Regression*
PENGOLAHAN DAN PENYAJIAN
Correlation and Regression
Introduction to Regression
Chapter 14 Multiple Regression
Presentation transcript:

Multivariate Analysis

One-way ANOVA Tests the difference in the means of 2 or more nominal groups Tests the difference in the means of 2 or more nominal groups E.g., High vs. Medium vs. Low exposure E.g., High vs. Medium vs. Low exposure Can be used with more than one IV Can be used with more than one IV Two-way ANOVA, Three-way ANOVA etc. Two-way ANOVA, Three-way ANOVA etc.

ANOVA _______-way ANOVA _______-way ANOVA Number refers to the number of IVs Number refers to the number of IVs Tests whether there are differences in the means of IV groups Tests whether there are differences in the means of IV groups E.g.: E.g.: Experimental vs. control group Experimental vs. control group Women vs. Men Women vs. Men High vs. Medium vs. Low exposure High vs. Medium vs. Low exposure

Logic of ANOVA Variance partitioned into: Variance partitioned into: 1. Systematic variance: 1. Systematic variance: the result of the influence of the Ivs the result of the influence of the Ivs 2. Error variance: 2. Error variance: the result of unknown factors the result of unknown factors Variation in scores partitions the variance into two parts by calculating the “sum of squares”: Variation in scores partitions the variance into two parts by calculating the “sum of squares”: 1. Between groups variation (systematic) 1. Between groups variation (systematic) 2. Within groups variation (error) 2. Within groups variation (error) SS total = SS between + SS within SS total = SS between + SS within

Significant and Non-significant Differences Significant: Between > Within Non-significant: Within > Between

Partitioning the Variance Comparisons Total variation = score – grand mean Total variation = score – grand mean Between variation = group mean – grand mean Between variation = group mean – grand mean Within variation = score – group mean Within variation = score – group mean Deviation is taken, then squared, then summed across cases Deviation is taken, then squared, then summed across cases Hence the term “Sum of squares” (SS) Hence the term “Sum of squares” (SS)

One-way ANOVA example Total SS (deviation from grand mean) Group AGroup BGroup C Mean = Grand mean = 54

One-way ANOVA example Total SS (deviation from grand mean) Group A Group B Group C Sum of squares = = 178

One-way ANOVA example Between SS (group mean – grand mean) A B C Group means Group deviation from grand mean Squared deviation n(squared deviation) Between SS = = 130 Grand mean = 54

One-way ANOVA example Within SS (score - group mean) ABC Deviation from group means Squared deviations Within SS = = 48

The F equation for ANOVA F = Between groups sum of squares/(k-1) Within groups sum of squares/(N-k) Within groups sum of squares/(N-k) N = total number of subjects k = number of groups Numerator = Mean square between groups Denominator = Mean square within groups

Significance of F F-critical is 3.89 (2,12 df) F observed > F critical 3.89 Groups are significantly different -T-tests could then be run to determine which groups are significantly different from which other groups

Computer Printout Example

Two-way ANOVA ANOVA compares: ANOVA compares: Between and within groups variance Between and within groups variance Adds a second IV to one-way ANOVA Adds a second IV to one-way ANOVA 2 IV and 1 DV 2 IV and 1 DV Analyzes significance of: Analyzes significance of: Main effects of each IV Main effects of each IV Interaction effect of the IVs Interaction effect of the IVs

Graphs of potential outcomes No main effects or interactions No main effects or interactions Main effects of color only Main effects of color only Main effects for motion only Main effects for motion only Main effects for color and motion Main effects for color and motion Interactions Interactions

Graphs Color B&W x Motion * Still AROUSALAROUSAL

No main effects for interactions Color B&W x Motion * Still AROUSALAROUSAL

No main effects for interactions Color B&W x Motion * Still x x * * AROUSALAROUSAL

Main effects for color only Color B&W x Motion * Still AROUSALAROUSAL

Main effects for color only Color B&W x Motion * Still x x * * AROUSALAROUSAL

Main effects for motion only Color B&W x Motion * Still AROUSALAROUSAL

Main effects for motion only Color B&W x Motion * Still x x ** AROUSALAROUSAL

Main effects for color and motion Color B&W x Motion * Still AROUSALAROUSAL

Main effects for color and motion Color B&W x Motion * Still x x * * AROUSALAROUSAL

Transverse interaction Color B&W x Motion * Still AROUSALAROUSAL

Transverse interaction Color B&W x Motion * Still x x * * AROUSALAROUSAL

Interaction—color only makes a difference for motion Color B&W x Motion * Still AROUSALAROUSAL

Interaction—color only makes a difference for motion Color B&W x Motion * Still x x * * AROUSALAROUSAL

Partitioning the variance for Two- way ANOVA Total variation = Main effect variable 1 + Main effect variable 2 + Interaction + Residual (within)

Summary Table for Two-way ANOVA SourceSSdfMSF Main effect 1 Main effect 2 InteractionWithinTotal

Printout Example

Printout plot

Scatter Plot of Price and Attendance Price is the average seat price for a single regular season game in today’s dollars Price is the average seat price for a single regular season game in today’s dollars Attendance is total annual attendance and is in millions of people per annum. Attendance is total annual attendance and is in millions of people per annum.

Is there a relation there? Lets use linear regression to find out, that is Lets use linear regression to find out, that is Let’s fit a straight line to the data. Let’s fit a straight line to the data. But aren’t there lots of straight lines that could fit? But aren’t there lots of straight lines that could fit? Yes! Yes!

Desirable Properties We would like the “closest” line, that is the one that minimizes the error We would like the “closest” line, that is the one that minimizes the error The idea here is that there is actually a relation, but there is also noise. We would like to make sure the noise (i.e., the deviation from the postulated straight line) to be as small as possible. The idea here is that there is actually a relation, but there is also noise. We would like to make sure the noise (i.e., the deviation from the postulated straight line) to be as small as possible. We would like the error (or noise) to be unrelated to the independent variable (in this case price). We would like the error (or noise) to be unrelated to the independent variable (in this case price). If it were, it would not be noise --- right! If it were, it would not be noise --- right!

Scatter Plot of Price and Attendance Price is the average seat price for a single regular season game in today’s dollars Price is the average seat price for a single regular season game in today’s dollars Attendance is total annual attendance and is in millions of people per annum. Attendance is total annual attendance and is in millions of people per annum.

Simple Regression The simple linear regression MODEL is: The simple linear regression MODEL is: y =  0 +  1 x +  describes how y is related to x describes how y is related to x  0 and  1 are called parameters of the model.  0 and  1 are called parameters of the model.  is a random variable called the error term.  is a random variable called the error term. xy e

Simple Regression Graph of the regression equation is a straight line. Graph of the regression equation is a straight line. β is the population y-intercept of the regression line. β 0 is the population y-intercept of the regression line. β 1 is the population slope of the regression line. β 1 is the population slope of the regression line. E(y) is the expected value of y for a given x value E(y) is the expected value of y for a given x value

Simple Regression E(y)E(y)E(y)E(y) x Slope  1 is positive Regression line Intercept  0

Simple Regression E(y)E(y)E(y)E(y) x Slope  1 is 0 Regression line Intercept  0

Types of Regression Models

Regression Modeling Steps 1.Hypothesize Deterministic Components 1.Hypothesize Deterministic Components 2.Estimate Unknown Model Parameters 2.Estimate Unknown Model Parameters 3.Specify Probability Distribution of Random Error Term 3.Specify Probability Distribution of Random Error Term Estimate Standard Deviation of Error Estimate Standard Deviation of Error 4.Evaluate Model 4.Evaluate Model 5.Use Model for Prediction & Estimation 5.Use Model for Prediction & Estimation

Linear Multiple Regression Model 1.Relationship between 1 dependent & 2 or more independent variables is a linear function 1.Relationship between 1 dependent & 2 or more independent variables is a linear function Dependent (response) variable Independent (explanatory) variables Population slopes Population Y-intercept Random error

Multiple Regression Model Multivariate model