Canonical Correlation Analysis: An overview with application to learning methods By David R. Hardoon, Sandor Szedmak, John Shawe-Taylor School of Electronics.

Slides:



Advertisements
Similar presentations
Ordinary Least-Squares
Advertisements

A. The Basic Principle We consider the multivariate extension of multiple linear regression – modeling the relationship between m responses Y 1,…,Y m and.
Canonical Correlation
Canonical Correlation simple correlation -- y 1 = x 1 multiple correlation -- y 1 = x 1 x 2 x 3 canonical correlation -- y 1 y 2 y 3 = x 1 x 2 x 3 The.
Mutidimensional Data Analysis Growth of big databases requires important data processing.  Need for having methods allowing to extract this information.
1 Simple Linear Regression and Correlation The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES Assessing the model –T-tests –R-square.
Simple Linear Regression and Correlation
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Subspace and Kernel Methods April 2004 Seong-Wook Joo.
Principal Component Analysis
Simple Linear Regression
Principal component analysis (PCA)
Expectation Maximization Method Effective Image Retrieval Based on Hidden Concept Discovery in Image Database By Sanket Korgaonkar Masters Computer Science.
Statistics for Business and Economics
Duality Dual problem Duality Theorem Complementary Slackness
Canonical Correlation: Equations Psy 524 Andrew Ainsworth.
The Simple Regression Model
Canonical correlations
SIMPLE LINEAR REGRESSION
Independent Component Analysis (ICA) and Factor Analysis (FA)
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
SIMPLE LINEAR REGRESSION
Tables, Figures, and Equations
Simple Linear Regression. Introduction In Chapters 17 to 19, we examine the relationship between interval variables via a mathematical equation. The motivation.
1 Simple Linear Regression 1. review of least squares procedure 2. inference for least squares lines.
Copyright ©2011 Pearson Education 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft Excel 6 th Global Edition.
Separate multivariate observations
SIMPLE LINEAR REGRESSION
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Empirical Modeling Dongsup Kim Department of Biosystems, KAIST Fall, 2004.
Summarized by Soo-Jin Kim
Chapter 2 Dimensionality Reduction. Linear Methods
Presented By Wanchen Lu 2/25/2013
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft.
The use of machine translation tools for cross-lingual text-mining Blaz Fortuna Jozef Stefan Institute, Ljubljana John Shawe-Taylor Southampton University.
Some matrix stuff.
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
SINGULAR VALUE DECOMPOSITION (SVD)
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Reduces time complexity: Less computation Reduces space complexity: Less parameters Simpler models are more robust on small datasets More interpretable;
Principle Component Analysis and its use in MA clustering Lecture 12.
Principal Component Analysis (PCA)
Regression Analysis Deterministic model No chance of an error in calculating y for a given x Probabilistic model chance of an error First order linear.
Advanced Statistics Factor Analysis, I. Introduction Factor analysis is a statistical technique about the relation between: (a)observed variables (X i.
Multiple Regression Learning Objectives n Explain the Linear Multiple Regression Model n Interpret Linear Multiple Regression Computer Output n Test.
Multivariate Transformation. Multivariate Transformations  Started in statistics of psychology and sociology.  Also called multivariate analyses and.
Université d’Ottawa / University of Ottawa 2001 Bio 8100s Applied Multivariate Biostatistics L11.1 Lecture 11: Canonical correlation analysis (CANCOR)
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Principal Components Analysis ( PCA)
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
1 Modification of Correlation Kernels in SVM, KPCA and KCCA in Texture Classification Yo Horikawa Kagawa University, Japan.
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Canonical Correlation Analysis (CCA). CCA This is it! The mother of all linear statistical analysis When ? We want to find a structural relation between.
Statistics made easy Basic principles. Statistics made easy Assume we are interested in getting some information about the average height of people in.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Inference for Least Squares Lines
Decomposition of Sum of Squares
LECTURE 11: Advanced Discriminant Analysis
Multiple Regression Analysis and Model Building
Regression.
Principal Component Analysis (PCA)
Efficient Image Classification on Vertically Decomposed Data
Image Segmentation Techniques
Contact: Machine Learning – (Linear) Regression Wilson Mckerrow (Fenyo lab postdoc) Contact:
Large scale multilingual and multimodal integration
Canonical Correlation Analysis
Presentation transcript:

Canonical Correlation Analysis: An overview with application to learning methods By David R. Hardoon, Sandor Szedmak, John Shawe-Taylor School of Electronics and Computer Science, University of Southampton Published in Neural Computaion, 2004 Presented by: Shankar Bhargav

Measuring the linear relationship between two multi dimensional variables Finding two sets of basis vectors such that the correlation between the projections of the variables onto these basis vectors is maximized Determine Correlation Coefficients Canonical Correlation Analysis

More than one canonical correlations will be found each corresponding to a different set of basis vectors/Canonical variates Correlations between successively extracted canonical variates are smaller and smaller Correlation coefficients : Proportion of correlation between the canonical variates accounted for by the particular variable. Canonical Correlation Analysis

Differences with Correlation Not dependent on the coordinate system of variables Finds direction that yield maximum correlations

Find basis vectors for two sets of variables x, y such that the correlations between the projections of the variables onto these basis vector S x = (x.w x ) and S y = (y.w y ) ρ = E[S x S y ] √ E[S x 2 ] E[S y 2 ] ρ = E[(x T w x y T w y )] √E[(x T w x x T w x ) ] E[(y T w y y T w y )]

ρ = max w x w y E[w x T x y T w y ] √E[w x T x x T w x ] E[w y T y y T w y ] ρ = max w x w y w x T C xy w y √ w x T C xx w x w y T C yy w y Solving this with constraint with constraint w x T C xx w x =1 w y T C yy w y =1

C xx -1 C xy C yy -1 C yx w x = ρ 2 w x C yy -1 C yx C xx -1 C xy w y = ρ 2 w y C xy w y = ρλ x C xx w x C yx w x = ρλ y C yy w y λ x =λ y -1 = w y T C yy w y √ √ w x T C xx w x

CCA in Matlab [ A, B, r, U, V ] = canoncorr(x, y) [ A, B, r, U, V ] = canoncorr(x, y) x, y : set of variables in the form of matrices  Each row is an observation  Each column is an attribute/feature A, B: Matrices containing the correlation coefficient r : Column matrix containing the canonical correlations (Successively decreasing) U, V: Canonical variates/basis vectors for A,B respectively

Interpretation of CCA Correlation coefficient represents unique contribution of each variable to relation Multicollinearity may obscure relationships Factor Loading : Correlations between the canonical variates (basis vector) and the variables in each set Proportion of variance explained by the canonical variates can be inferred by factor loading

Redundancy Calculation Redundancy left =[ ∑ (loadings left 2 )/p]*R c 2 Redundancy right =[ ∑ (loadings right 2 )/q]*R c 2 p – Number of variable in the first (left) set of variables p – Number of variable in the first (left) set of variables q – Number of variable in the second (right) set of variables q – Number of variable in the second (right) set of variables Rc2 – Respective squared canonical correlation Since successively extracted roots are uncorrelated we can sum the redundancies across all correlations to get a single index of redundancy. Since successively extracted roots are uncorrelated we can sum the redundancies across all correlations to get a single index of redundancy.

Application Kernel CCA can be used to find non linear relationships between multi variates Two views of the same semantic object to extract the representation of the semantics Speaker Recognition – Audio and Lip movement Speaker Recognition – Audio and Lip movement Image retrieval – Image features (HSV, Texture) and Associated text Image retrieval – Image features (HSV, Texture) and Associated text

Use of KCCA in cross-modal retrieval 400 records of JPEG images for each class with associated text and a total of 3 classes 400 records of JPEG images for each class with associated text and a total of 3 classes Data was split randomly into 2 parts for training and test Data was split randomly into 2 parts for training and test Features Features Image – HSV Color, Gabor texture Text – Term frequencies Results were taken for an average of 10 runs Results were taken for an average of 10 runs

Cross-modal retrieval Content based retrieval: Retrieve images in the same class Tested with 10 and 30 images sets where count j k = 1 if the image k in the set is of the same label as the text query present in the set, else count j k = 0. where count j k = 1 if the image k in the set is of the same label as the text query present in the set, else count j k = 0.

Comparison of KCCA (with 5 and 30 Eigen vectors) with GVSM Content based retrieval

`

Mate based retrieval Match the exact image among the selected retrieved images Tested with 10 and 30 images sets where count j = 1 if the exact matching image was present in the set else it is 0 where count j = 1 if the exact matching image was present in the set else it is 0

Comparison of KCCA (with 30 and 150 Eigen vectors) with GVSM Mate based retrieval

Comments The good Good explanation of CCA and KCCA Good explanation of CCA and KCCA Innovative use of KCCA in image retrieval application Innovative use of KCCA in image retrieval application The bad The data set and the number of classes used were small The data set and the number of classes used were small The image set size is not taken into account while calculating accuracy in Mate based retrieval The image set size is not taken into account while calculating accuracy in Mate based retrieval Could have done cross-validation tests Could have done cross-validation tests

Limitations and Assumptions of CCA At least 40 to 60 times as many cases as variables is recommended to get relliable estimates for two roots– BarciKowski & Stevens(1986) Outliers can greatly affect the canonical correlation Variables in two sets should not be completely redundant

Thank you