Advances in Random Matrix Theory: Let there be tools

Slides:



Advertisements
Similar presentations
Advanced topics in Financial Econometrics Bas Werker Tilburg University, SAMSI fellow.
Advertisements

Class 6: Hypothesis testing and confidence intervals
Statistical perturbation theory for spectral clustering Harrachov, 2007 A. Spence and Z. Stoyanov.
Eigen Decomposition and Singular Value Decomposition
Chapter 28 – Part II Matrix Operations. Gaussian elimination Gaussian elimination LU factorization LU factorization Gaussian elimination with partial.
Singular Values of the GUE Surprises that we Missed Alan Edelman and Michael LaCroix MIT June 16, 2014 (acknowledging gratefully the help from Bernie Wang)
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Random Matrix Laws & Jacobi Operators Alan Edelman MIT May 19, 2014 joint with Alex Dubbs and Praveen Venkataramana (acknowledging gratefully the help.
An NLA Look at σ min Universality (& the Stoch Diff Operator) Alan Edelman and Po-Ru Loh MIT Applied Mathematics Random Matrices October 10, 2010.
Lecture 3 Probability and Measurement Error, Part 2.
Solving Linear Systems (Numerical Recipes, Chap 2)
Lecture 17 Introduction to Eigenvalue Problems
Modern iterative methods For basic iterative methods, converge linearly Modern iterative methods, converge faster –Krylov subspace method Steepest descent.
1 Quantum Monte Carlo Methods Jian-Sheng Wang Dept of Computational Science, National University of Singapore.
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Started detailed look at PCA Reviewed linear algebra Today: More linear algebra Multivariate.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Maximum likelihood (ML) and likelihood ratio (LR) test
Random Matrices Hieu D. Nguyen Rowan University Rowan Math Seminar
Discrete Event Simulation How to generate RV according to a specified distribution? geometric Poisson etc. Example of a DEVS: repair problem.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Basics of regression analysis
Maximum likelihood (ML)
1cs542g-term Notes  Extra class next week (Oct 12, not this Friday)  To submit your assignment: me the URL of a page containing (links to)
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 6. Eigenvalue problems.
Modern Navigation Thomas Herring
Needle-like Triangles, Matrices, and Lewis Carroll Alan Edelman Mathematics Computer Science & AI Labs Gilbert Strang Mathematics Computer Science & AI.
The Erdös-Rényi models
Random Matrix Theory Numerical Computation and Remarkable Applications Alan Edelman Mathematics Computer Science & AI Labs Computer Science & AI Laboratories.
Random Matrix Theory and Numerical Linear Algebra: A story of communication Alan Edelman Mathematics Computer Science & AI Labs ILAS Meeting June 3, 2013.
Systems of Linear Equation and Matrices
Using Adaptive Methods for Updating/Downdating PageRank Gene H. Golub Stanford University SCCM Joint Work With Sep Kamvar, Taher Haveliwala.
Alan Edelman Oren Mangoubi, Bernie Wang Mathematics
Algorithms for a large sparse nonlinear eigenvalue problem Yusaku Yamamoto Dept. of Computational Science & Engineering Nagoya University.
An Empirical Likelihood Ratio Based Goodness-of-Fit Test for Two-parameter Weibull Distributions Presented by: Ms. Ratchadaporn Meksena Student ID:
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Yung-Kyun Noh and Joo-kyung Kim Biointelligence.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Conjugate Priors Multinomial Gaussian MAP Variance Estimation Example.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Class 27: Question 1 TRUE or FALSE: If P is a projection matrix of the form P=A(A T A) -1 A T then P is a symmetric matrix. 1. TRUE 2. FALSE.
Orthogonalization via Deflation By Achiya Dax Hydrological Service Jerusalem, Israel
A Review of Some Fundamental Mathematical and Statistical Concepts UnB Mestrado em Ciências Contábeis Prof. Otávio Medeiros, MSc, PhD.
A Note on Rectangular Quotients By Achiya Dax Hydrological Service Jerusalem, Israel
Numerical Analysis – Eigenvalue and Eigenvector Hanyang University Jong-Il Park.
Al Parker July 19, 2011 Polynomial Accelerated Iterative Sampling of Normal Distributions.
The Random Matrix Technique of Ghosts and Shadows Alan Edelman Dept of Mathematics Computer Science and AI Laboratories Massachusetts Institute of Technology.
- 1 - Matlab statistics fundamentals Normal distribution % basic functions mew=100; sig=10; x=90; normpdf(x,mew,sig) 1/sig/sqrt(2*pi)*exp(-(x-mew)^2/sig^2/2)
Progress in the method of Ghosts and Shadows for Beta Ensembles Alan Edelman (MIT) Alex Dubbs (MIT) and Plamen Koev (SJS) Aug 10, 2012 IMS Singapore Workshop.
Chapter 31Introduction to Statistical Quality Control, 7th Edition by Douglas C. Montgomery. Copyright (c) 2012 John Wiley & Sons, Inc.
By Josh Zimmer Department of Mathematics and Computer Science The set ℤ p = {0,1,...,p-1} forms a finite field. There are p ⁴ possible 2×2 matrices in.
Stats & Summary. The Woodbury Theorem where the inverses.
Singular Value Decomposition and Numerical Rank. The SVD was established for real square matrices in the 1870’s by Beltrami & Jordan for complex square.
ALGEBRAIC EIGEN VALUE PROBLEMS
Estimating standard error using bootstrap
MA2213 Lecture 8 Eigenvectors.
Chapter 14 Fitting Probability Distributions
Model Inference and Averaging
Stochastic Differential Equations and Random Matrices
Numerical Computations and Random Matrix Theory
Advances in Random Matrix Theory (stochastic eigenanalysis)
Advances in Random Matrix Theory (stochastic eigenanalysis)
Probability & Statistics Probability Theory Mathematical Probability Models Event Relationships Distributions of Random Variables Continuous Random.
Why are random matrix eigenvalues cool?
POINT ESTIMATOR OF PARAMETERS
Singular Value Decomposition SVD
Numerical Computations and Random Matrix Theory
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions July chonbuk national university.
Remainder Theorem.
Berlin Chen Department of Computer Science & Information Engineering
Applied Statistics and Probability for Engineers
Presentation transcript:

Advances in Random Matrix Theory: Let there be tools Alan Edelman Brian Sutton, Plamen Koev, Ioana Dumitriu, Raj Rao and others MIT: Dept of Mathematics, Computer Science AI Laboratories World Congress, Bernoulli Society Barcelona, Spain Wednesday July 28, 2004 1/18/2019

Tools So many applications … Random matrix theory: catalyst for 21st century special functions, analytical techniques, statistical techniques In addition to mathematics and papers Need tools for the novice! Need tools for the engineers! Need tools for the specialists! 1/18/2019

Themes of this talk Tools for general beta What is beta? Think of it as a measure of (inverse) volatility in “classical” random matrices. Tools for complicated derived random matrices Tools for numerical computation and simulation 1/18/2019

Wigner’s Semi-Circle The classical & most famous rand eig theorem Let S = random symmetric Gaussian MATLAB: A=randn(n); S=(A+A’)/2; S known as the Gaussian Orthogonal Ensemble Normalized eigenvalue histogram is a semi-circle Precise statements require n etc. n=20; s=30000; d=.05; %matrix size, samples, sample dist e=[]; %gather up eigenvalues im=1; %imaginary(1) or real(0) for i=1:s, a=randn(n)+im*sqrt(-1)*randn(n);a=(a+a')/(2*sqrt(2*n*(im+1))); v=eig(a)'; e=[e v]; end hold off; [m x]=hist(e,-1.5:d:1.5); bar(x,m*pi/(2*d*n*s)); axis('square'); axis([-1.5 1.5 -1 2]); hold on; t=-1:.01:1; plot(t,sqrt(1-t.^2),'r');

G 6 5 4 3 2 1 sym matrix to tridiagonal form Same eigenvalue distribution as GOE: O(n) storage !! O(n2) compute 1/18/2019

G 6 5 4 3 2  General beta beta: 1: reals 2: complexes 4: quaternions Bidiagonal Version corresponds To Wishart matrices of Statistics 1/18/2019

Tools Motivation: A condition number problem Jack & Hypergeometric of Matrix Argument MOPS: Ioana Dumitriu’s talk The Polynomial Method The tridiagonal numerical 109 trick 1/18/2019

Tools Motivation: A condition number problem Jack & Hypergeometric of Matrix Argument MOPS: Ioana Dumitriu’s talk The Polynomial Method The tridiagonal numerical 109 trick 1/18/2019

Numerical Analysis: Condition Numbers (A) = “condition number of A” If A=UV’ is the svd, then (A) = max/min . Alternatively, (A) =  max (A’A)/ min (A’A) One number that measures digits lost in finite precision and general matrix “badness” Small=good Large=bad The condition of a random matrix??? 1/18/2019

Von Neumann & co. estimates (1947-1951) “For a ‘random matrix’ of order n the expectation value has been shown to be about n” Goldstine, von Neumann “… we choose two different values of , namely n and 10n” Bargmann, Montgomery, vN “With a probability ~1 …  < 10n” X  P(<n) 0.02 P(< 10n)0.44 P(<10n) 0.80 1/18/2019

Random cond numbers, n Distribution of /n Experiment with n=200 1/18/2019

Finite n n=10 n=25 n=50 n=100 1/18/2019

Condition Number Distributions P(κ/n > x) ≈ 2/x P(κ/n2 > x) ≈ 4/x2 Real n x n, n∞ Generalizations: β: 1=real, 2=complex finite matrices rectangular: m x n Complex n x n, n∞ 1/18/2019

Condition Number Distributions P(κ/n > x) ≈ 2/x P(κ/n2 > x) ≈ 4/x2 Square, n∞: P(κ/nβ > x) ≈ (2ββ-1/Γ(β))/xβ (All Betas!!) General Formula: P(κ>x) ~Cµ/x β(n-m+1), where µ= β(n-m+1)/2th moment of the largest eigenvalue of Wm-1,n+1 (β) and C is a known geometrical constant. Density for the largest eig of W is known in terms of 1F1((β/2)(n+1), ((β/2)(n+m-1); -(x/2)Im-1) from which µ is available Tracy-Widom law applies probably all beta for large m,n. Johnstone shows at least beta=1,2. Real n x n, n∞ Complex n x n, n∞ 1/18/2019

Tools Motivation: A condition number problem Jack & Hypergeometric of Matrix Argument MOPS: Ioana Dumitriu’s talk The Polynomial Method The tridiagonal numerical 109 trick 1/18/2019

∫ pκ(x)pλ(x) Δ(x)β ∏i w(xi)dxi = δκλ Multivariate Orthogonal Polynomials & Hypergeometrics of Matrix Argument Ioana Dumitriu’s talk The important special functions of the 21st century Begin with w(x) on I ∫ pκ(x)pλ(x) Δ(x)β ∏i w(xi)dxi = δκλ Jack Polynomials orthogonal for w=1 on the unit circle. Analogs of xm 1/18/2019

Multivariate Hypergeometric Functions 1/18/2019

Multivariate Hypergeometric Functions 1/18/2019

Wishart (Laguerre) Matrices! 1/18/2019

Plamen’s clever idea 1/18/2019

Tools Motivation: A condition number problem Jack & Hypergeometric of Matrix Argument MOPS: Ioana Dumitriu’s talk The Polynomial Method The tridiagonal numerical 109 trick 1/18/2019

Mops (Dumitriu etc.) Symbolic 1/18/2019

Symbolic MOPS applications A=randn(n); S=(A+A’)/2; trace(S^4) det(S^3) 1/18/2019

Symbolic MOPS applications β=3; hist(eig(S)) 1/18/2019

Smallest eigenvalue statistics A=randn(m,n); hist(min(svd(A).^2)) 1/18/2019

Tools Motivation: A condition number problem Jack & Hypergeometric of Matrix Argument MOPS: Ioana Dumitriu’s talk The Polynomial Method -- Raj! The tridiagonal numerical 109 trick 1/18/2019

Courtesy of the Polynomial Method RM Tool – Raj! Courtesy of the Polynomial Method 1/18/2019

1/18/2019

Tools Motivation: A condition number problem Jack & Hypergeometric of Matrix Argument MOPS: Ioana Dumitriu’s talk The Polynomial Method The tridiagonal numerical 109 trick 1/18/2019

Everyone’s Favorite Tridiagonal -2 1 1 n2 … … … … … d2 dx2 1/18/2019

Everyone’s Favorite Tridiagonal -2 1 G 1 (βn)1/2 1 n2 … … … + … … dW β1/2 d2 dx2 + 1/18/2019

Stochastic Operator Limit , dW β 2 x dx d + - , N(0,2) χ n β 2 1 ~ H 2) (n 1) ÷ ø ö ç è æ - … … … , G β 2 H n + » ¥ 1/18/2019

Largest Eigenvalue Plots 1/18/2019

MATLAB beta=1; n=1e9; opts.disp=0;opts.issym=1; alpha=10; k=round(alpha*n^(1/3)); % cutoff parameters d=sqrt(chi2rnd( beta*(n:-1:(n-k-1))))'; H=spdiags( d,1,k,k)+spdiags(randn(k,1),0,k,k); H=(H+H')/sqrt(4*n*beta); eigs(H,1,1,opts) 1/18/2019

Tricks to get O(n9) speedup Sparse matrix storage (Only O(n) storage is used) Tridiagonal Ensemble Formulas (Any beta is available due to the tridiagonal ensemble) The Lanczos Algorithm for Eigenvalue Computation ( This allows the computation of the extreme eigenvalue faster than typical general purpose eigensolvers.) The shift-and-invert accelerator to Lanczos and Arnoldi (Since we know the eigenvalues are near 1, we can accelerate the convergence of the largest eigenvalue) The ARPACK software package as made available seamlessly in MATLAB (The Arnoldi package contains state of the art data structures and numerical choices.) The observation that if k = 10n1/3 , then the largest eigenvalue is determined numerically by the top k × k segment of n. (This is an interesting mathematical statement related to the decay of the Airy function.) 1/18/2019

Random matrix tools! 1/18/2019