Principal Nested Spheres Analysis

Slides:



Advertisements
Similar presentations
Self-Organizing Maps Projection of p dimensional observations to a two (or one) dimensional grid space Constraint version of K-means clustering –Prototypes.
Advertisements

Dimensionality Reduction PCA -- SVD
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
Clustering and Dimensionality Reduction Brendan and Yifang April
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Started detailed look at PCA Reviewed linear algebra Today: More linear algebra Multivariate.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
“Random Projections on Smooth Manifolds” -A short summary
Principal Component Analysis
3D Geometry for Computer Graphics. 2 The plan today Least squares approach  General / Polynomial fitting  Linear systems of equations  Local polynomial.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
3D Geometry for Computer Graphics
PATTERN RECOGNITION : PRINCIPAL COMPONENTS ANALYSIS Prof.Dr.Cevdet Demir
Dimensionality Reduction
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
1cs542g-term Notes  Extra class next week (Oct 12, not this Friday)  To submit your assignment: me the URL of a page containing (links to)
SVD(Singular Value Decomposition) and Its Applications
Statistics – O. R. 892 Object Oriented Data Analysis J. S. Marron Dept. of Statistics and Operations Research University of North Carolina.
Object Orie’d Data Analysis, Last Time Finished Algebra Review Multivariate Probability Review PCA as an Optimization Problem (Eigen-decomp. gives rotation,
Summarized by Soo-Jin Kim
Chapter 2 Dimensionality Reduction. Linear Methods
Presented By Wanchen Lu 2/25/2013
Non Negative Matrix Factorization
1 UNC, Stat & OR Nonnegative Matrix Factorization.
A Challenging Example Male Pelvis –Bladder – Prostate – Rectum.
Statistics – O. R. 891 Object Oriented Data Analysis J. S. Marron Dept. of Statistics and Operations Research University of North Carolina.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Object Orie’d Data Analysis, Last Time Discrimination for manifold data (Sen) –Simple Tangent plane SVM –Iterated TANgent plane SVM –Manifold SVM Interesting.
Data Mining Course 0 Manifold learning Xin Yang. Data Mining Course 1 Outline Manifold and Manifold Learning Classical Dimensionality Reduction Semi-Supervised.
Maximal Data Piling Visual similarity of & ? Can show (Ahn & Marron 2009), for d < n: I.e. directions are the same! How can this be? Note lengths are different.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Common Property of Shape Data Objects: Natural Feature Space is Curved I.e. a Manifold (from Differential Geometry) Shapes As Data Objects.
1 UNC, Stat & OR PCA Extensions for Data on Manifolds Fletcher (Principal Geodesic Anal.) Best fit of geodesic to data Constrained to go through geodesic.
PATTERN RECOGNITION : PRINCIPAL COMPONENTS ANALYSIS Richard Brereton
Participant Presentations Please Sign Up: Name (Onyen is fine, or …) Are You ENRolled? Tentative Title (???? Is OK) When: Thurs., Early, Oct., Nov.,
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
Object Orie’d Data Analysis, Last Time PCA Redistribution of Energy - ANOVA PCA Data Representation PCA Simulation Alternate PCA Computation Primal – Dual.
Object Orie’d Data Analysis, Last Time PCA Redistribution of Energy - ANOVA PCA Data Representation PCA Simulation Alternate PCA Computation Primal – Dual.
GWAS Data Analysis. L1 PCA Challenge: L1 Projections Hard to Interpret (i.e. Little Data Insight) Solution: 1)Compute PC Directions Using L1 2)Compute.
Unsupervised Learning II Feature Extraction
Dimension reduction (2) EDR space Sliced inverse regression Multi-dimensional LDA Partial Least Squares Network Component analysis.
SigClust Statistical Significance of Clusters in HDLSS Data When is a cluster “really there”? Liu et al (2007), Huang et al (2014)
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Linear Algebra Review Multivariate Probability Review PCA as an Optimization Problem (Eigen-decomp.
Landmark Based Shapes As Data Objects
Efficient non-linear analysis of large data sets
Unsupervised Learning
PREDICT 422: Practical Machine Learning
Motion Segmentation with Missing Data using PowerFactorization & GPCA
Unsupervised Riemannian Clustering of Probability Density Functions
René Vidal and Xiaodong Fan Center for Imaging Science
René Vidal Time/Place: T-Th 4.30pm-6pm, Hodson 301
A Unified Algebraic Approach to 2D and 3D Motion Segmentation
Functional Data Analysis
Statistics – O. R. 881 Object Oriented Data Analysis
Machine Learning Dimensionality Reduction
Statistics – O. R. 881 Object Oriented Data Analysis
Maximal Data Piling MDP in Increasing Dimensions:
An Interesting Question
Landmark Based Shape Analysis
Today is Last Class Meeting
EE 290A Generalized Principal Component Analysis
Probabilistic Models with Latent Variables
Presented by Kojo Essuman Ackah Spring 2018 STA 6557 Project
Non-negative Matrix Factorization (NMF)
Lecture 15: Least Square Regression Metric Embeddings
Principal Component Analysis
Non-Negative Matrix Factorization
Unsupervised Learning
Statistics – O. R. 891 Object Oriented Data Analysis
Presentation transcript:

Principal Nested Spheres Analysis Main Goal: Extend Principal Arc Analysis (S2 to Sk) Jung, Dryden & Marron (2012) Impact on Segmentation: PGA Segmentation: used ~20 comp’s PNS Segmentation: only need ~13 Resulted in visually better fits to data

Principal Nested Spheres Analysis Main Goal: Extend Principal Arc Analysis (S2 to Sk) Jung, Dryden & Marron (2012) Important Landmark: This Motivated Backwards PCA

Principal Nested Spheres Analysis Key Idea: Replace usual forwards view of PCA With a backwards approach to PCA

Multiple linear regression: Terminology Multiple linear regression:

Multiple linear regression: Terminology Multiple linear regression: Stepwise approaches: Forwards: Start small, iteratively add variables to model

Multiple linear regression: Terminology Multiple linear regression: Stepwise approaches: Forwards: Start small, iteratively add variables to model Backwards: Start with all, iteratively remove variables from model

Illust’n of PCA View: Recall Raw Data EgView1p1RawData.ps

Illust’n of PCA View: PC1 Projections EgView1p51proj3dPC1.ps

Illust’n of PCA View: PC2 Projections EgView1p52proj3dPC2.ps

Illust’n of PCA View: Projections on PC1,2 plane EgView1p54proj3dPC12.ps

Principal Nested Spheres Analysis Replace usual forwards view of PCA Data  PC1 (1-d approx)

Principal Nested Spheres Analysis Replace usual forwards view of PCA Data  PC1 (1-d approx)  PC2 (1-d approx of Data-PC1)  PC1 U PC2 (2-d approx)

Principal Nested Spheres Analysis Replace usual forwards view of PCA Data  PC1 (1-d approx)  PC2 (1-d approx of Data-PC1)  PC1 U PC2 (2-d approx)  PC1 U … U PCr (r-d approx)

Principal Nested Spheres Analysis With a backwards approach to PCA Data  PC1 U … U PCr (r-d approx)

Principal Nested Spheres Analysis With a backwards approach to PCA Data  PC1 U … U PCr (r-d approx)  PC1 U … U PC(r-1)

Principal Nested Spheres Analysis With a backwards approach to PCA Data  PC1 U … U PCr (r-d approx)  PC1 U … U PC(r-1)  PC1 U PC2 (2-d approx)

Principal Nested Spheres Analysis With a backwards approach to PCA Data  PC1 U … U PCr (r-d approx)  PC1 U … U PC(r-1)  PC1 U PC2 (2-d approx)  PC1 (1-d approx)

Principal Component Analysis Euclidean Settings: Forwards PCA = Backwards PCA (Pythagorean Theorem, ANOVA Decomposition) So Not Interesting But Very Different in Non-Euclidean Settings (Backwards is Better !?!)

Principal Component Analysis Important Property of PCA: Nested Series of Approximations 𝑃𝐶 1 ⊆ 𝑃𝐶 1 ,𝑃𝐶 2 ⊆ 𝑃𝐶 1 ,𝑃𝐶 2 , 𝑃𝐶 3 ⊆⋯ (Often taken for granted) (Desirable in Non-Euclidean Settings)

Desirability of Nesting: Multi-Scale Analysis Makes Sense Non-Euclidean PCA Desirability of Nesting: Multi-Scale Analysis Makes Sense Scores Visualization Makes Sense

Scores Visualization in PCA

Scores Visualization in PCA

Scores Visualization in PCA

An Interesting Question How generally applicable is Backwards approach to PCA?

An Interesting Question How generally applicable is Backwards approach to PCA? Where is this already being done???

An Interesting Question How generally applicable is Backwards approach to PCA? Discussion: Jung et al (2010) Pizer et al (2012)

An Interesting Question How generally applicable is Backwards approach to PCA? Anywhere this is already being done???

An Interesting Question How generally applicable is Backwards approach to PCA? An Application: Nonnegative Matrix Factorization = PCA in Positive Orthant

An Interesting Question How generally applicable is Backwards approach to PCA? An Application: Nonnegative Matrix Factorization = PCA in Positive Orthant Think 𝑋 ≈𝐿𝑆 With ≥ 0 Constraints (on both 𝐿 & 𝑆)

Nonnegative Matrix Factorization Isn’t This Just PCA? In the Nonnegative Orthant? No, Most PC Directions Leave Orthant

Nonnegative Matrix Factorization Isn’t This Just PCA? Data (Near Orthant Faces)

Nonnegative Matrix Factorization Isn’t This Just PCA? Data Mean (Centered Analysis)

Nonnegative Matrix Factorization Isn’t This Just PCA? Data Mean PC1 Projections Leave Orthant!

Nonnegative Matrix Factorization Isn’t This Just PCA? Data Mean PC1 Projections PC1 ⋃ 2 Proj’ns Leave Orthant!

Nonnegative Matrix Factorization Note: Problem not Fixed by SVD (“Uncentered PCA”) Orthant Leaving Gets Worse

Nonnegative Matrix Factorization Standard Approach: Lee et al (1999): Formulate & Solve Optimization

Nonnegative Matrix Factorization Standard Approach: Lee et al (1999): Formulate & Solve Optimization Major Challenge: Not Nested, (𝑘=3 ≉ 𝑘=4)

Nonnegative Matrix Factorization Standard NMF (Projections All Inside Orthant)

Nonnegative Matrix Factorization Standard NMF But Note Not Nested No “Multi-scale” Analysis Possible (Scores Plot?!?)

Nonnegative Matrix Factorization Improved Version: Use Backwards PCA Idea “Nonnegative Nested Cone Analysis” Collaborator: Lingsong Zhang (Purdue) Zhang, Marron, Lu (2013)

Nonnegative Nested Cone Analysis Same Toy Data Set All Projections In Orthant

Nonnegative Nested Cone Analysis Same Toy Data Set Rank 1 Approx. Properly Nested

Nonnegative Nested Cone Analysis Chemical Spectra Data Scores Plot Rainbow Ordering

Nonnegative Nested Cone Analysis Components For Major Reaction

Nonnegative Nested Cone Analysis 3rd Comp. Something Early? (Known Lab Error)

Nonnegative Nested Cone Analysis 4th Comp. Much Less Variation (Random Noise)

Nonnegative Nested Cone Analysis Chemical Spectral Data Loadings Interpretation

Nonnegative Nested Cone Analysis Chemical Spectral Data NNCA 2 Looks very similar?!?

Nonnegative Nested Cone Analysis Chemical Spectral Data Study Difference NNCA2 – NNCA1

Nonnegative Nested Cone Analysis Chemical Spectral Data Rainbow Shows 2 Components Very Close, but Chemically Distinct

Nonnegative Nested Cone Analysis Chemical Spectral Data Zoomed View Is Also Interesting

Nonnegative Nested Cone Analysis Chemical Spectral Data Gives Clearer View

Nonnegative Nested Cone Analysis Chemical Spectral Data Rank 3 Approximation Highlights Lab Early Error

Nonnegative Nested Cone Analysis 5-d Toy Example (Rainbow Colored by Peak Order)

Nonnegative Nested Cone Analysis 5-d Toy Example Rank 1 NNCA Approx.

Nonnegative Nested Cone Analysis 5-d Toy Example Rank 2 NNCA Approx.

Nonnegative Nested Cone Analysis 5-d Toy Example Rank 2 NNCA Approx. Nonneg. Basis Elements (Not Trivial)

Nonnegative Nested Cone Analysis 5-d Toy Example Rank 3 NNCA Approx. Current Research: How Many Nonneg. Basis El’ts Needed?

An Interesting Question How generally applicable is Backwards approach to PCA? Potential Application: Principal Curves Hastie & Stuetzle, (1989) (Foundation of Manifold Learning)

Goal: Find lower dimensional manifold that well approximates data Manifold Learning Goal: Find lower dimensional manifold that well approximates data ISOmap Tennenbaum (2000) Local Linear Embedding Roweis & Saul (2000)

1st Principal Curve Linear Reg’n Usual Smooth

1st Principal Curve Linear Reg’n Proj’s Reg’n Usual Smooth

Linear Reg’n Proj’s Reg’n Usual Smooth Princ’l Curve 1st Principal Curve Linear Reg’n Proj’s Reg’n Usual Smooth Princ’l Curve

An Interesting Question How generally applicable is Backwards approach to PCA? Potential Application: Principal Curves Perceived Major Challenge: How to find 2nd Principal Curve? Backwards approach???

An Interesting Question Key Component: Principal Surfaces LeBlanc & Tibshirani (1996)

An Interesting Question Key Component: Principal Surfaces LeBlanc & Tibshirani (1996) Challenge: Can have any dimensional surface, But how to nest???

An Interesting Question How generally applicable is Backwards approach to PCA? Another Potential Application: Trees as Data (early days)

An Interesting Question How generally applicable is Backwards approach to PCA? An Attractive Answer

An Interesting Question How generally applicable is Backwards approach to PCA? An Attractive Answer: James Damon, UNC Mathematics Geometry Singularity Theory

An Interesting Question How generally applicable is Backwards approach to PCA? An Attractive Answer: James Damon, UNC Mathematics Damon and Marron (2013)

An Interesting Question How generally applicable is Backwards approach to PCA? An Attractive Answer: James Damon, UNC Mathematics Key Idea: Express Backwards PCA as Nested Series of Constraints

General View of Backwards PCA Define Nested Spaces via Constraints Satisfying More Constraints ⇒ ⇒ Smaller Subspaces

General View of Backwards PCA Define Nested Spaces via Constraints E.g. SVD (Singular Value Decomposition = = Not Mean Centered PCA) (notationally very clean)

General View of Backwards PCA Define Nested Spaces via Constraints E.g. SVD Have 𝑘 Nested Subspaces: 𝑆 1 ⊆ 𝑆 2 ⊆⋯⊆ 𝑆 𝑑

General View of Backwards PCA Define Nested Spaces via Constraints E.g. SVD 𝑆 𝑘 = 𝑥 :𝑥= 𝑗=1 𝑘 𝑐 𝑗 𝑢 𝑗 𝑘-th SVD Subspace Scores Loading Vectors

General View of Backwards PCA Define Nested Spaces via Constraints E.g. SVD 𝑆 𝑘 = 𝑥 :𝑥= 𝑗=1 𝑘 𝑐 𝑗 𝑢 𝑗 Now Define: 𝑆 𝑘−1 = 𝑥∈ 𝑆 𝑘 : 𝑥, 𝑢 𝑘 =0

General View of Backwards PCA Define Nested Spaces via Constraints E.g. SVD 𝑆 𝑘 = 𝑥 :𝑥= 𝑗=1 𝑘 𝑐 𝑗 𝑢 𝑗 Now Define: 𝑆 𝑘−1 = 𝑥∈ 𝑆 𝑘 : 𝑥, 𝑢 𝑘 =0 Constraint Gives Nested Reduction of Dim’n

General View of Backwards PCA Define Nested Spaces via Constraints Backwards PCA Reduce Using Affine Constraints

General View of Backwards PCA Define Nested Spaces via Constraints Backwards PCA Principal Nested Spheres Use Affine Constraints (Planar Slices) In Ambient Space

General View of Backwards PCA Define Nested Spaces via Constraints Backwards PCA Principal Nested Spheres Principal Surfaces Spline Constraint Within Previous?

General View of Backwards PCA Define Nested Spaces via Constraints Backwards PCA Principal Nested Spheres Principal Surfaces Spline Constraint Within Previous? {Been Done Already???}

General View of Backwards PCA Define Nested Spaces via Constraints Backwards PCA Principal Nested Spheres Principal Surfaces Other Manifold Data Spaces Sub-Manifold Constraints?? (Algebraic Geometry)

General View of Backwards PCA Define Nested Spaces via Constraints Backwards PCA Principal Nested Spheres Principal Surfaces Other Manifold Data Spaces Tree Spaces Suitable Constraints???