Canonical Correlation simple correlation -- y 1 = x 1 multiple correlation -- y 1 = x 1 x 2 x 3 canonical correlation -- y 1 y 2 y 3 = x 1 x 2 x 3 The.

Slides:



Advertisements
Similar presentations
Canonical Correlation
Advertisements

Automated Regression Modeling Descriptive vs. Predictive Regression Models Four common automated modeling procedures Forward Modeling Backward Modeling.
Research Hypotheses and Multiple Regression: 2 Comparing model performance across populations Comparing model performance across criteria.
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
Chapter 17 Making Sense of Advanced Statistical Procedures in Research Articles.
Metrics, Algorithms & Follow-ups Profile Similarity Measures Cluster combination procedures Hierarchical vs. Non-hierarchical Clustering Statistical follow-up.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 14 Using Multivariate Design and Analysis.
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Principal Components An Introduction Exploratory factoring Meaning & application of “principal components” Basic steps in a PC analysis PC extraction process.
Linear Discriminant Function LDF & MANOVA LDF & Multiple Regression Geometric example of LDF & multivariate power Evaluating & reporting LDF results 3.
Discrim Continued Psy 524 Andrew Ainsworth. Types of Discriminant Function Analysis They are the same as the types of multiple regression Direct Discrim.
Multiple Regression Models Advantages of multiple regression Important preliminary analyses Parts of a multiple regression model & interpretation Differences.
Canonical Correlation: Equations Psy 524 Andrew Ainsworth.
Multiple Regression Models: Some Details & Surprises Review of raw & standardized models Differences between r, b & β Bivariate & Multivariate patterns.
The Simple Regression Model
Multiple Regression Models
Regression and Correlation
MANOVA LDF & MANOVA Geometric example of MANVOA & multivariate power MANOVA dimensionality Follow-up analyses if k > 2 Factorial MANOVA.
Multiple-group linear discriminant function maximum & contributing ldf dimensions concentrated & diffuse ldf structures follow-up analyses evaluating &
2.3. Measures of Dispersion (Variation):
Bivariate & Multivariate Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas process of.
Today Concepts underlying inferential statistics
An Introduction to Classification Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification.
Norms & Norming Raw score: straightforward, unmodified accounting of performance Norms: test performance data of a particular group of test takers that.
Discriminant Analysis Testing latent variables as predictors of groups.
Factor Analysis Psy 524 Ainsworth.
ANCOVA Lecture 9 Andrew Ainsworth. What is ANCOVA?
Regression with 2 IVs Generalization of Regression from 1 to 2 Independent Variables.
Some matrix stuff.
Correlation and Regression PS397 Testing and Measurement January 16, 2007 Thanh-Thanh Tieu.
Chapter 18 Some Other (Important) Statistical Procedures You Should Know About Part IV Significantly Different: Using Inferential Statistics.
BIOL 582 Lecture Set 11 Bivariate Data Correlation Regression.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
CANONICAL CORRELATION. Return to MR  Previously, we’ve dealt with multiple regression, a case where we used multiple independent variables to predict.
Regression Models w/ 2 Quant Variables Sources of data for this model Variations of this model Main effects version of the model –Interpreting the regression.
Aron, Aron, & Coups, Statistics for the Behavioral and Social Sciences: A Brief Course (3e), © 2005 Prentice Hall Chapter 12 Making Sense of Advanced Statistical.
Adjusted from slides attributed to Andrew Ainsworth
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 12 Testing for Relationships Tests of linear relationships –Correlation 2 continuous.
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
UNDERSTANDING DESCRIPTION AND CORRELATION. CORRELATION COEFFICIENTS: DESCRIBING THE STRENGTH OF RELATIONSHIPS Pearson r Correlation Coefficient Strength.
Chapter 9: Correlation and Regression Analysis. Correlation Correlation is a numerical way to measure the strength and direction of a linear association.
Correlation & Regression Analysis
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
 Seeks to determine group membership from predictor variables ◦ Given group membership, how many people can we correctly classify?
SOCW 671 #11 Correlation and Regression. Uses of Correlation To study the strength of a relationship To study the direction of a relationship Scattergrams.
Applied Quantitative Analysis and Practices LECTURE#28 By Dr. Osman Sadiq Paracha.
Unit 7 Statistics: Multivariate Analysis of Variance (MANOVA) & Discriminant Functional Analysis (DFA) Chat until class starts.
D/RS 1013 Discriminant Analysis. Discriminant Analysis Overview n multivariate extension of the one-way ANOVA n looks at differences between 2 or more.
Université d’Ottawa / University of Ottawa 2001 Bio 8100s Applied Multivariate Biostatistics L11.1 Lecture 11: Canonical correlation analysis (CANCOR)
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 18 Multivariate Statistics.
Topics, Summer 2008 Day 1. Introduction Day 2. Samples and populations Day 3. Evaluating relationships Scatterplots and correlation Day 4. Regression and.
Linear Discriminant Function Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification & LDF.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Methods of Presenting and Interpreting Information Class 9.
Multiple Regression.
Analysis and Interpretation: Multiple Variables Simultaneously
Simple Linear Regression
Correlation and Regression
Quantitative Methods Simple Regression.
Making Sense of Advanced Statistical Procedures in Research Articles
Multiple Regression.
CORRELATION ANALYSIS.
Multivariate Statistics
Chapter 9: Differences among Groups
Product moment correlation
Prediction/Regression
2.3. Measures of Dispersion (Variation):
Correlation and Simple Linear Regression
Presentation transcript:

Canonical Correlation simple correlation -- y 1 = x 1 multiple correlation -- y 1 = x 1 x 2 x 3 canonical correlation -- y 1 y 2 y 3 = x 1 x 2 x 3 The “most multivariate” of the correlation models All other models (including ANOVAs, ANCOVAs and MANOVA and LDF) are part of this model

Uses and Alternatives for Canonical Correlation Just because there are multiple criterion variables and multiple predictors doesn’t mean that a canonical correlation is the best analysis for the data! If your research hypotheses/questions are about correlations between specific criterion variables and specific predictor variables... use the specified set of Pearson’s correlations use Hotelling’s t or Rosenthal’s Z to test hypotheses about which correlations are larger than others (which predictors are more correlated with which criterion variables) If your research hypotheses/questions are about differences between the best models to understand/predict specific criterion variables build the specified set of multiple regression models use “cross-validation” procedures and Hotelling’s/Rosenthal’s tests to compare multiple regression models across the criterion variables

Let’s take a look at how canonical correlation “works”, to help understand when to use it (instead of simple or multiple reg)… Start with multiple y and x variables y 1 y 2 y 3 = x 1 x 2 x 3 construct a “canonical variate” as the combination of y variables CV y1 = b 1 y 1 + b 2 y 2 + b 3 y 3 construct a “canonical variate” as the combination of x variables CV x1 = b 1 x 1 + b 2 x 2 + b 3 x 3 The canonical correlation is the correlation of the canonical variates R c = r cvy1, cvx1

1st thing -- The weights used to construct CV y1 and CV x1 are chosen to produce the largest possible R c This is different from simple correlation weighting for X 1 is chosen so y’ mimics y This is different from multiple correlation weighting for x 1, x 2, etc chosen to maximize the correlation of y’ with y In canonical correlation the weightings for the Y variables and for the X variables are chosen simultaneously to maximize the correlation between the constructed X variate and the constructed Y variate 2nd thing -- There can be multiple canonicals max # is the smaller of the number of criterion & number of predictors like ldf, have to decide if we have a concentrated or diffuse structure

So, canonical correlation is used (instead of comparisons among selected simple and/or multiple correlations) when … research hypotheses are about which combination(s) of criterion variables are related to which combination(s) or predictors research hypotheses are about whether the set of variables have a concentrated or a diffuse correlational structure Examples: have a set of interrelated criterion variables (e.g., five measures of depression) and want to know if they all share the same relationship with the predictors (a concentrated structure) have two sets of predictor variables and want to show that one set is most useful for one understanding the one criterion variable and the other set is most useful for understanding a different set of criterion variables

Steps of inspecting and describing a canonical correlation… Determine the number of canonical variates and correlations maximum number is smaller of # predictors or # criterion each R c is tested for H0: R c = 0 (same test as for MR) Evaluate the strength of each canonical relationship R c ² is “shared variance” between that x and y variate Interpret each canonical variate Structure weights tell which variables related to which variates Standardized regression weights tell about the unique contribution of variables to variates

Depicting and Describing “Shared Variance” in Canonical Correlation There are 3 kinds of “shared variances” in cancorr: R C ² squared canonical correlation variance shared between the corresponding Y and X canonical variates at PC variance shared between a canonical variate and the set of variables from which it is constructed an index of how well that canonical variate “represents” the set of variables PC is computed as the average structure weight for that canonical variate RC redundancy coefficient variance shared between a canonical variate from one set of variables and the other set of variables an index of how well a canonical variate “represents” the other set of variables

X1 x2 x3 x4 y1 y2 y3 We start with dependent (y) variables and covariate/predictor (x) variables a canonical variate is constructed for each set so that the variates are as highly correlated as possible the R C ² tells the variance shared between the canonical variates the PC tells the variance shared between each canonical variate and the variables from which it was constructed CV y1 = b 1 y1 + b 2 y2 + b 3 y3 + aCV x1 = b 1 x1 + b 2 x2 + b 3 x3 + b 4 x4 + a PC y1 PC x1 RC²RC²

X1 x2 x3 x4 y1 y2 y3 We can also ask how much variance is shared between the CV of one set of variables (the covariate/predictor variables) and the other set of variables (the dependent variables)-- a redundancy coefficient shown is “how much variance among the dependent variables (Ys) is accounted for by the covariate/predictor canonical variate (CV x1) a redundancy coefficient is calculated as R C ² * PC of the variables involved (R C ² * PC y1, here) CV y1 CV x1 PC y1 PC x1 RC²RC² redundancy coef.

About the size of the redundancy coefficient the “old” story was that a small redundancy coefficient meant the canonical correlation analysis was uninteresting or misleading the “new” story a redundancy coefficient can be “small” for two different reasons that have very different interpretations “good” -- the CV is “strongly” composed of a minority of the variables in that set (few vars, but with large weights) often denotes specificity among the variables -- a prerequisite for a diffuse structure “yucky” -- the CV is “weakly” composed of a majority of the variables in that set (many variables, but small weights) often denotes that the construct related to the “other vars” is not well-represented by these vars, and so must be “put together” using small portions of ` many vars