Download presentation
Presentation is loading. Please wait.
1
Functional Data Analysis
Insightful Decomposition Vertical Variation Horizβl Varβn
2
More Data Objects Final Curve Warps:
Warp Each Data Curve, π 1 , β―, π π To Template Mean, π π Denote Warp Functions πΎ 1 , β―, πΎ π Gives (Roughly Speaking): Vertical Components π 1 β πΎ 1 , β―, π π β πΎ π (Aligned Curves) Horizontal Components πΎ 1 , β―, πΎ π Data Objects I
3
More Data Objects Final Curve Warps:
Data Objects II Final Curve Warps: Warp Each Data Curve, π 1 , β―, π π To Template Mean, π π Denote Warp Functions πΎ 1 , β―, πΎ π Gives (Roughly Speaking): Vertical Components π 1 β πΎ 1 , β―, π π β πΎ π (Aligned Curves) Horizontal Components πΎ 1 , β―, πΎ π ~ Kendallβs Shapes
4
More Data Objects Final Curve Warps:
Warp Each Data Curve, π 1 , β―, π π To Template Mean, π π Denote Warp Functions πΎ 1 , β―, πΎ π Gives (Roughly Speaking): Vertical Components π 1 β πΎ 1 , β―, π π β πΎ π (Aligned Curves) Horizontal Components πΎ 1 , β―, πΎ π Data Objects III ~ Changβs Transfoβs
5
Toy Example Conventional PCA Projections Power Spread Across Spectrum
6
Toy Example Conventional PCA Scores Views of 1-d Curve Bending Through 4 Dimβnsβ
7
Toy Example Aligned Curve PCA Projections All Varβn In 1st Component
8
Toy Example Warps, PC Projections Mostly 1st PC, But 2nd Helps Some
9
TIC testbed Special Feature: Answer Key of Known Peaks Goal: Find Warps To Align These
10
TIC testbed Fisher β Rao Alignment
11
PNS on SRVF Sphere Toy Example View As Points Tangent Plane PC 1 PNS 1 Boundary of Nonnegative Orthant
12
PNS on SRVF Sphere Real Data Analysis: Blood Glucose Curves
13
PNS on SRVF Sphere Real Data Analysis: Blood Glucose Curves
14
Juggling Data Clustering In Phase Variation Space:
15
Probability Distributions as Data Objects
Interesting Question: What is βBestβ Representation? (Which Function ~ Distributions?) Density Function? (Very Interpretable) Cumulative Distribution Function Quantile Function (Recall Inverse of CDF)
16
Probability Distributions as Data Objects
Recall Representations of Distributions
17
Probability Distributions as Data Objects
PCA of Random Densities Power Spread Across Spectrum
18
Probability Distributions as Data Objects
Now Try Quantile Representation (Same E.g.)
19
Probability Distributions as Data Objects
PCA of Quantile Repβns Only 2 Modes! Shift Tilt
20
Probability Distributions as Data Objects
Conclusion: Quantile Representation Best for Typical 2 βFirstβ Modes of Variation (Essentially Linear Modes) Density & C. D. F. Generally Much Worse (Natural Modes are Non-Linear)
21
Probability Distributions as Data Objects
Point 1: Mean Changes, Nicely Represented By Quantiles
22
Probability Distributions as Data Objects
Point 1: Mean Changes, Nicely Represented By Quantiles
23
Probability Distributions as Data Objects
Point 2: Spread Changes, Nicely Represented By Quantiles
24
Probability Distributions as Data Objects
Point 2: Spread Changes, Nicely Represented By Quantiles
25
Random Matrix Theory Main Idea:
Pure Noise Distribution of PCA Eigenvalues Usefulness: Interpretation of Scree Plots For Eigenvalues π π of Sample Covariance Ξ£ Plot π π vs. π
26
PCA Redistβn of Energy (Cont.)
Note, have already considered some of these Useful Plots: Power Spectrum (as %s) Cumulative Power Spectrum (%) Common Terminology: Power Spectrum is Called βScree Plotβ Kruskal (1964) Cattell (1966) (all but name βscreeβ) (1st Appearance of name???) 26
27
PCA Redistβn of Energy (Cont.)
Note, have already considered some of these Useful Plots: Power Spectrum (as %s) Cumulative Power Spectrum (%) Large Values Reflect Important Structure 27
28
PCA Redistβn of Energy (Cont.)
Note, have already considered some of these Useful Plots: Power Spectrum (as %s) Cumulative Power Spectrum (%) Zoom In & Characterize Noise 28
29
Random Matrix Theory Pure Noise Data Matrix: π=
Defined as: Entries i.i.d. π(0,1) Thinking of Columns As Data Objects π π
30
Random Matrix Theory Clean Notation Version of Covariance Matrix:
Ξ£ = 1 π π π π‘ Simplified by: No Mean Centering (using π(0,1)) Roughly OK, By Usual Mean Centering Also Standardize by 1 π not 1 πβ1 Easy & Sensible for No Mean Centering Size = πΓπ
31
Random Matrix Theory Eigenvalues are π π , diagonal entries of Ξ in
Ξ£ =πΞ π π‘ (Eigen-analysis) Distribution of π π ?
32
Random Matrix Theory For π=100, π=1000, Eigenvalues β1
But There Is (Chance) Variation
33
Random Matrix Theory Smaller π=500 Boosts Variation (More Uncertainty)
34
Random Matrix Theory Smaller π=200 Boosts Variation (More Uncertainty)
35
Random Matrix Theory Smaller π=100 Boosts Variation
But Canβt Go Negative Although Can Get Large
36
Random Matrix Theory Larger π=10,000 Reduces Variation
37
Random Matrix Theory Larger π=100,000 Reduces Variation
38
Random Matrix Theory Fix π¦= π π , and let π, π grow. Essentially Same
Shape
39
Random Matrix Theory Fix π¦= π π , and let π, π grow. Essentially Same
Shape
40
Random Matrix Theory Fix π¦= π π , and let π, π grow. But Less Sampling
Noise
41
Random Matrix Theory Fix π¦= π π , and let π, π grow. But Less Sampling
Noise
42
Random Matrix Theory Fix π¦= π π , and let π, π grow. What Is
That Shape?
43
Empirical Spectral Density
Random Matrix Theory Shape is Captured by Empirical Spectral Density βDensityβ Of These Eigenvalues
44
(in limit as π, πββ, with π¦= π π )
Random Matrix Theory Limiting Spectral Density (in limit as π, πββ, with π¦= π π ) References: MarΔenko Pastur (1967) Yao et al (2015) Dobriban (2015)
45
Random Matrix Theory Limiting Spectral Density
(in limit as π, πββ, with π¦= π π ) Limit Exists No Closed Form But Can Implicitly Define (Using Integral Equations) And Numerically Approximate
46
Random Matrix Theory Limiting Spectral Density, for given π¦= π π
Convenient Visualization Interface By Hyo Young Choi
47
Random Matrix Theory LSD: Above Case π=200, π=100
48
Random Matrix Theory LSD: Above Case π=200, π=100, π¦=0.5
log 10 π¦ =β0.301
49
Random Matrix Theory LSD Note: These Have Finite Support β(0,β)
50
Random Matrix Theory LSD: Now Try Smaller (More Negative)
Values of π¦= π π
51
Random Matrix Theory LSD: Now Try Smaller (More Negative)
Values of π¦= π π
52
Random Matrix Theory LSD: Now Try Smaller (More Negative)
Values of π¦= π π Note: Support Points β1
53
Random Matrix Theory LSD: Now Try Smaller (More Negative)
Values of π¦= π π
54
Random Matrix Theory LSD: Now Try Smaller (More Negative)
Values of π¦= π π Note: Increasing Symmetry
55
Random Matrix Theory Larger π=100,000 Reduces Variation Recall
Previous Large π Case, LSD is Zooming In On This
56
Random Matrix Theory Limiting Case: lim πββ lim πββ
Called Medium Dimension High Sample Size Resulting Density is βSemi-Circleβ π π₯ = 2 π π
π
2 β π₯ βπ
,π
(π₯) Called βWigner Semi-Circle Distributionβ
57
Random Matrix Theory Summary: Have Studied Data Matrix Shapes
Observed: Convergence to 1 Increasing Symmetry What About Other Direction (Larger π)?
58
Random Matrix Theory Consider Growing π Challenge:
Only π Columns in π (so rank =π) Yet Ξ£ is πΓπ So Have πβπ Eigenvalues =0
59
Random Matrix Theory LSD: Start With π¦= π π =1 Case
60
Random Matrix Theory LSD: Now Try Larger Values of π¦= π π
Proportion of 0 Eigenvalues
61
Random Matrix Theory LSD: Now Try Larger Values of π¦= π π
Spectral Density of Non-0 Eigenvalues
62
Random Matrix Theory LSD: Now Try Larger Values of π¦= π π
63
Random Matrix Theory LSD: Now Try Larger Values of π¦= π π
64
Random Matrix Theory LSD: Now Try Larger Values of π¦= π π Again Heads
Towards Semi-Circle But Small Proportion
65
Shapes Seem Similar to Above
Random Matrix Theory LSD: Now Try Larger Values of π¦= π π Note: Shapes Seem Similar to Above
66
Random Matrix Theory LSD: Dual Covariance Variation
Idea: Replace Ξ£ = 1 π π π π‘ by 1 π π π‘ π Recall: Rows as Data Objects Inner Product of π Different Normalization (π not π) N(0,1) Avoids Messy Centering Issues
67
Random Matrix Theory LSD: Dual Covariance Variation π¦= π π =100 Is
Close to Semi-Circle
68
Random Matrix Theory LSD: Dual Covariance Variation
69
Random Matrix Theory LSD: Dual Covariance Variation Seem to Follow
Similar Pattern
70
Random Matrix Theory LSD: Dual Covariance Variation
71
Random Matrix Theory LSD: Dual Covariance Variation For π<π Now
Get 0 Eignevalues
72
Random Matrix Theory LSD: Dual Covariance Variation
73
Random Matrix Theory LSD: Dual Covariance Variation
74
Random Matrix Theory LSD: Dual Covariance Variation Again Heads To
Semi-Circle
75
Random Matrix Theory LSD: Primal & Dual Overlaid For Direct Comparison
Notes: Area = 1 Area = 1 - Bar
76
Random Matrix Theory LSD: Primal & Dual
77
Random Matrix Theory LSD: Primal & Dual Very Close For πβπ
78
Random Matrix Theory LSD: Primal & Dual Very Close For πβπ
79
Random Matrix Theory LSD: Primal & Dual
80
Random Matrix Theory LSD: Primal & Dual
81
Random Matrix Theory LSD: Primal & Dual
82
Random Matrix Theory LSD: Rescaled Primal & Dual π¦ ΓπΏππ· (underneath)
1 π¦ Γπ·π’ππ πΏππ·
83
Random Matrix Theory LSD: Rescaled Primal & Dual
84
Random Matrix Theory LSD: Rescaled Primal & Dual
85
Random Matrix Theory LSD: Rescaled Primal & Dual
86
Random Matrix Theory LSD: Rescaled Primal & Dual
87
Random Matrix Theory LSD: Rescaled Primal & Dual
88
Random Matrix Theory Conclusion:
Family of Marcenko β Pastur Distributions Has Several Interesting Symmetries
89
Random Matrix Theory Important Parallel Theory:
Distribution of Largest Eigenvalue (Assuming Matrix of i.i.d. N(0,1)s) Tracey Widom (1994) Good Discussion of Statistical Implications Johnstone (2008)
90
Participant Presentations
Zhengling Qi Classification in personalized medicine Zhiyuan Liu CPNS Visualization in Pablo Fuhui Fang DiProPerm Analysis of OsteoArthritis Data
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.