Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 IBM Research, Tokyo Research Lab. Tsuyoshi Idé Pairwise Symmetry Decomposition.

Slides:



Advertisements
Similar presentations
Spatial point patterns and Geostatistics an introduction
Advertisements

Group Theory II.
5.4 Basis And Dimension.
5.1 Real Vector Spaces.
3D Geometry for Computer Graphics
Mathematics of Cryptography Part II: Algebraic Structures
Cryptography and Network Security
Tokyo Research Laboratory © Copyright IBM Corporation 2006 | 2006/09/19 | PKDD 2006 Why does subsequence time-series clustering produce sine waves? IBM.
CLUSTERING PROXIMITY MEASURES
On the alternative approaches to ITRF formulation. A theoretical comparison. Department of Geodesy and Surveying Aristotle University of Thessaloniki Athanasios.
Part 2.4: Rules that Govern Symmetry 1. Define Group Theory The Rules for Groups Combination Tables Additional Rules/Definitions – Subgroups – Representations.
Symmetries in Nuclei, Tokyo, 2008 Scale invariance Object of which a detail when enlarged becomes (approximately) identical to the object itself. Condition.
What role should probabilistic sensitivity analysis play in SMC decision making? Andrew Briggs, DPhil University of Oxford.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
International Workshop on Computer Vision - Institute for Studies in Theoretical Physics and Mathematics, April , Tehran 1 IV COMPUTING SIZE.
2.III. Basis and Dimension 1.Basis 2.Dimension 3.Vector Spaces and Linear Systems 4.Combining Subspaces.
QA-2 FRM-GARP Sep-2001 Zvi Wiener Quantitative Analysis 2.
FRM Zvi Wiener Following P. Jorion, Financial Risk Manager Handbook Financial Risk Management.
Probability theory 2011 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different definitions.
Basic Concepts and Definitions Vector and Function Space. A finite or an infinite dimensional linear vector/function space described with set of non-unique.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Probability theory 2008 Outline of lecture 5 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different.
Correlation. The sample covariance matrix: where.
Review of Probability.
Nidhi Joshi Centre for Theoretical Physics Jamia Millia Islamia Collaborators: Aditya Rotti, Tarun Souradeep Confronting particle-cosmology with Planck.
Tokyo Research Laboratory © Copyright IBM Corporation 2009 | 2009/04/03 | SDM 09 / Travel-Time Prediction Travel-Time Prediction using Gaussian Process.
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
Statistical Methods For Engineers ChE 477 (UO Lab) Larry Baxter & Stan Harding Brigham Young University.
2009 Mathematics Standards of Learning Training Institutes Algebra II Virginia Department of Education.
Copyright © Cengage Learning. All rights reserved. CHAPTER 11 ANALYSIS OF ALGORITHM EFFICIENCY ANALYSIS OF ALGORITHM EFFICIENCY.
Correlations and Copulas 1. Measures of Dependence 2 The risk can be split into two parts: the individual risks and the dependence structure between them.
GROUPS & THEIR REPRESENTATIONS: a card shuffling approach Wayne Lawton Department of Mathematics National University of Singapore S ,
Correlation.
1 Physical Chemistry III (728342) Chapter 5: Molecular Symmetry Piti Treesukol Kasetsart University Kamphaeng Saen Campus.
Some matrix stuff.
1 Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Estimating a Population Mean: σ Known 7-3, pg 355.
Week 15 - Wednesday.  What did we talk about last time?  Review first third of course.
2 2  Background  Vision in Human Brain  Efficient Coding Theory  Motivation  Natural Pictures  Methodology  Statistical Characteristics  Models.
Invariants to affine transform What is affine transform ?
General Introduction to Symmetry in Crystallography A. Daoud-Aladine (ISIS)
17. Group Theory 1.Introduction to Group Theory 2.Representation of Groups 3.Symmetry & Physics 4.Discrete Groups 5.Direct Products 6.Symmetric Groups.
Statistical Description of Charged Particle Beams and Emittance Measurements Jürgen Struckmeier HICforFAIR Workshop.
Elementary Linear Algebra Anton & Rorres, 9th Edition
Computer Graphics and Image Processing (CIS-601).
Sparse Kernel Methods 1 Sparse Kernel Methods for Classification and Regression October 17, 2007 Kyungchul Park SKKU.
AGC DSP AGC DSP Professor A G Constantinides©1 Signal Spaces The purpose of this part of the course is to introduce the basic concepts behind generalised.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
J. Flusser, T. Suk, and B. Zitová Moments and Moment Invariants in Pattern Recognition The slides accompanying.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Quantum Two 1. 2 Angular Momentum and Rotations 3.
Tokyo Research Laboratory © Copyright IBM Corporation 2003KDD 04 | 2004/08/24 | Eigenspace-based anomaly detection in computer systems IBM Research, Tokyo.
Joint Moments and Joint Characteristic Functions.
Ch 6 Vector Spaces. Vector Space Axioms X,Y,Z elements of  and α, β elements of  Def of vector addition Def of multiplication of scalar and vector These.
STA347 - week 91 Random Vectors and Matrices A random vector is a vector whose elements are random variables. The collective behavior of a p x 1 random.
Pattern Recognition Mathematic Review Hamid R. Rabiee Jafar Muhammadi Ali Jalali.
Lecture 7: Bivariate Statistics. 2 Properties of Standard Deviation Variance is just the square of the S.D. If a constant is added to all scores, it has.
Introduction to Symmetry Analysis Brian Cantwell Department of Aeronautics and Astronautics Stanford University Chapter 1 - Introduction to Symmetry.
Unsupervised Learning II Feature Extraction
Statistical Properties of Digital Piecewise Linear Chaotic Maps and Their Roles in Cryptography & Pseudo-Random Coding Li ShujunLi Shujun 1, Li Qi 2, Li.
Slide 1 Copyright © 2004 Pearson Education, Inc.  Descriptive Statistics summarize or describe the important characteristics of a known set of population.
SIGNALS PROCESSING AND ANALYSIS
GROUPS & THEIR REPRESENTATIONS: a card shuffling approach
Basis and Dimension Basis Dimension Vector Spaces and Linear Systems
Tutorial 9: Further Topics on Random Variables 2
Singular Value Decomposition SVD
2.III. Basis and Dimension
Chapter 4 Mathematical Expectation.
Marios Mattheakis and Pavlos Protopapas
Presentation transcript:

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 IBM Research, Tokyo Research Lab. Tsuyoshi Idé Pairwise Symmetry Decomposition Method for Generalized Covariance Analysis

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 2 Summary: We generalize the notion of covariance using group theory. covariance theoretical properties 2-variate cross cumulant An irreducible representation of a group generalization New approach to pattern recognition 2-variate CC as irreducible representations

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 3 Motivation. The traditional covariance cannot capture nonlinearities. x y data point (x i, y i )  The traditional covariance cannot capture nonlinearities. # of data points  C xy would be useless in this case.  We wish to explicitly define useful metrics for nonlinear correlations.  cf. kernel methods are black-boxes

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 4 The lowest order cross-cumulant (CC) is identical to the covariance. A generalized covariance could be higher-order CC. 2 nd order CC is identical to the covariance Use higher-order CC for generalizing the covariance Multivariate systems can be described with PDF. Notation of “cumulant average” The cumulants of p completely characterize p. * We assume zero-mean data hereafter. principle #1

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 5 Relation between the covariance and symmetries. The axioms of group can be used to characterize the symmetries. symmetric w.r.t. x and y, etc. A collection of such symmetry operations may be used for characterizing the symmetries. What is the guiding principle to define it? The axioms of group can be the guiding principle. Closure, Associativity, Identity, Inverse. PDF

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 6 The set of OPs is almost unique --- C 4v is the most appropriate group for characterizing the pairwise correlations.  Requirements on the group G  G should include OPs describing the symmetries within the xy plane.  G should include an OP to exchange x with y. Most general one is a group named C 4v Point group ? only rotations and mirror reflections Point group is a natural choice x y only xy

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 7 What is the C 4v group ? It contains 8 symmetry operations within the xy space.

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 8 Only a single IRR has been used for recognizing correlation patterns. We unveil the other representations ! One can easily prove: x y any correlation pattern A1A1 A2A2 B1B1 B2B2 E Further, a “symmetry decomposition theorem” holds: linear combination of the five IRRs Only the B 2 component has been used so far. Find the other IRRs, which haven’t been used so far. principle #2

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 9 The two consequences lead to the definition of the generalized covariances, which are symmetrized cross cumulants in the C 4v sense. Note There is arbitrariness in prefactors x and y should be standardized (unit variance) to be scale invariant 2-variate cross cumulantsirreducible representations of C 4v A 1, A 2, B 1, (B 2 ), E Construct IRRs as linear combinations of CCs result: principle #2 principle #1

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 10 Experiment with Lissajous’ trajectories. The generalized covariances detect the nonlinearities while the standard covariance C(B 2 ) fails. C(B 2 ) should be minus 1 due to the perfect Inverse linear correlation C(B 2 ) fails to capture the correlation C(A 1 ) succeeds to detect the nonlinear correlation C(B 2 ) fails to capture the correlation C(E 2 ) and C(E 2 ) succeed to detect the nonlinear correlation C(B 2 ) fails to capture the correlation C(A 2 ) succeeds to detect the nonlinear correlation

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 11 Summary: We have generalized the notion of covariance using group theory. covariance theoretical properties 2-variate cross cumulant B 2 irreducible representation of C 4v generalization Symmetry decomposition: new view to pattern recognition 2-variate CC as A 1, A 2, B 2, and E representations

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 12 Thank you !!

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 13 Background. How can you tell the difference of the two states quantitatively? The traditional covariance is not helpful. x y state B x y state A data point (x i, y i )  Traditional covariance # of data points  C xy would be useless in this case.  The traditional covariance cannot capture nonlinearities.  We wish to explicitly define useful metrics for nonlinear correlations.  cf. kernel methods are black-boxes

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 14 Take only 2-variate CC like Take only lower order ones up to Covariance as the lowest order CC: Summary of this section. We focus on cross cumulants (CC) as a theoretical basis. Describe n-variate systems using its PDF p employ cumulants as features of p assumption characterization approximation cumulants are expansion coeff. w.r.t. s Cumulant generating function Definition of cumulants Notation of “cumulant average”

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 15 Sparse Correlation Approximation of the cumulant generating function. What kind of terms are omitted? cumulant generating function 2-body cluster SCA terms taken * A pair of variable will be represented as x and y hereafter.

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 16 Mathematical preliminaries. Position operator, Hilbert space, Dirac’s bra-ket notation, and moments in the bra-ket notation x y D: The domain of p(r) Def. of the position eigenstate Hilbert space spanned by the position eigenstate thenetc. position operator where is the marginal DF wrt (x,y)

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 17 Verifying the definition of the generalized covariances: a few examples. representation matrices are all 1  A1 representation You can use the method of projection operators if want to construct IRRs systematically (See a textbook of group theory)

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 18 Experiment : Calculating the generalized covariances for Lissajous’ trajectories analytically.  Model of correlated variables →  assume the uniformly distribution over t  mean is zero for both x and y  variance is 1 for both x and y  Generalized covariances can be explicitly calculated  x y

Tokyo Research Laboratory © Copyright IBM Corporation 2005 | 2005/11/30 | ICDM 05 Page 19 Detailed summary.  We generalized the traditional notion of covariance based on the two theoretical properties  1 . Standard covariance is the lowest order 2- variate cross cumulant  2 . Standard covariance is the B 2 irreducible representation of the C 4v group.  Our result suggests a new approach to pattern recognition where patterns are characterized by the irreducible representation of a finite group  Practically, we found that  C(B 2 ) would be greatly enhanced for linear correlations.  C(E 1 ) and C(E 2 ) reflect some asymmetries in the distribution.  C(A 1 ) clearly takes a large value when the distribution has a donut-like shape.  Finally, C(A 2 ) would be enhanced by distributions with some Hakenkreuz-like correlations.  These features can be used in anomaly detection tasks where nonlinear correlations plays some important role