M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September 4 2009 1 SVD and LS M.A. Miceli University of Rome I Stats in the Château Jouy-en-Josas.

Slides:



Advertisements
Similar presentations
Vector Spaces A set V is called a vector space over a set K denoted V(K) if is an Abelian group, is a field, and For every element vV and K there exists.
Advertisements

4.1 Introduction to Matrices
Eigen Decomposition and Singular Value Decomposition
3.3 Hypothesis Testing in Multiple Linear Regression
General Linear Model With correlated error terms  =  2 V ≠  2 I.
A. The Basic Principle We consider the multivariate extension of multiple linear regression – modeling the relationship between m responses Y 1,…,Y m and.
Chapter 28 – Part II Matrix Operations. Gaussian elimination Gaussian elimination LU factorization LU factorization Gaussian elimination with partial.
Covariance Matrix Applications
Lecture 3: A brief background to multivariate statistics
Financial Applications of RMT Max Timmons May 13, 2013 Main Application: Improving Estimates from Empirical Covariance Matricies Overview of optimized.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
An introduction to Principal Component Analysis (PCA)
Principal component analysis (PCA)
Linear Regression Models Based on Chapter 3 of Hastie, Tibshirani and Friedman Slides by David Madigan.
Canonical Correlation: Equations Psy 524 Andrew Ainsworth.
Data mining and statistical learning, lecture 4 Outline Regression on a large number of correlated inputs  A few comments about shrinkage methods, such.
Canonical correlations
TFIDF-space  An obvious way to combine TF-IDF: the coordinate of document in axis is given by  General form of consists of three parts: Local weight.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Lecture 20 SVD and Its Applications Shang-Hua Teng.
Ordinary least squares regression (OLS)
Principal component analysis (PCA)
Boot Camp in Linear Algebra Joel Barajas Karla L Caballero University of California Silicon Valley Center October 8th, 2008.
Statistics 200b. Chapter 5. Chapter 4: inference via likelihood now Chapter 5: applications to particular situations.
Correlation. The sample covariance matrix: where.
Linear regression models in matrix terms. The regression function in matrix terms.
Matrices CS485/685 Computer Vision Dr. George Bebis.
Matrix Approach to Simple Linear Regression KNNL – Chapter 5.
Principle Component Analysis Presented by: Sabbir Ahmed Roll: FH-227.
Chapter 2 Dimensionality Reduction. Linear Methods
Presented By Wanchen Lu 2/25/2013
Multiple Linear Regression - Matrix Formulation Let x = (x 1, x 2, …, x n )′ be a n  1 column vector and let g(x) be a scalar function of x. Then, by.
Next. A Big Thanks Again Prof. Jason Bohland Quantitative Neuroscience Laboratory Boston University.
Some matrix stuff.
Basic Statistics Correlation Var Relationships Associations.
Least SquaresELE Adaptive Signal Processing 1 Method of Least Squares.
Method of Least Squares. Least Squares Method of Least Squares:  Deterministic approach The inputs u(1), u(2),..., u(N) are applied to the system The.
A Review of Some Fundamental Mathematical and Statistical Concepts UnB Mestrado em Ciências Contábeis Prof. Otávio Medeiros, MSc, PhD.
4.6: Rank. Definition: Let A be an mxn matrix. Then each row of A has n entries and can therefore be associated with a vector in The set of all linear.
Introduction to Matrices and Matrix Approach to Simple Linear Regression.
Trees Example More than one variable. The residual plot suggests that the linear model is satisfactory. The R squared value seems quite low though,
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
Université d’Ottawa / University of Ottawa 2001 Bio 8100s Applied Multivariate Biostatistics L11.1 Lecture 11: Canonical correlation analysis (CANCOR)
Singular Value Decomposition and Numerical Rank. The SVD was established for real square matrices in the 1870’s by Beltrami & Jordan for complex square.
The Principal Components Regression Method David C. Garen, Ph.D. Hydrologist USDA Natural Resources Conservation Service National Water and Climate Center.
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
Central limit theorem - go to web applet. Correlation maps vs. regression maps PNA is a time series of fluctuations in 500 mb heights PNA = 0.25 *
Boot Camp in Linear Algebra TIM 209 Prof. Ram Akella.
Reduced echelon form Matrix equations Null space Range Determinant Invertibility Similar matrices Eigenvalues Eigenvectors Diagonabilty Power.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Estimating standard error using bootstrap
Introduction to Vectors and Matrices
Linear Algebra review (optional)
CH 5: Multivariate Methods
Regression.
Singular Value Decomposition
Matrices Definition: A matrix is a rectangular array of numbers or symbolic elements In many applications, the rows of a matrix will represent individuals.
The regression model in matrix form
4.6: Rank.
OVERVIEW OF LINEAR MODELS
Parallelization of Sparse Coding & Dictionary Learning
Matrix Algebra and Random Vectors
Further Matrix Algebra
SVD, PCA, AND THE NFL By: Andrew Zachary.
OVERVIEW OF LINEAR MODELS
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 – 14, Tuesday 8th November 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR)
Lecture 13: Singular Value Decomposition (SVD)
Introduction to Vectors and Matrices
Principal Component Analysis
Presentation transcript:

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September SVD and LS M.A. Miceli University of Rome I Stats in the Château Jouy-en-Josas August 31 - September

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Motivations Problems of high dimensionality in estimation: –Rank < actual dimension of the data sets  inverse problems –Threholds in accepting variables eases on every dimension, as the number of variables/dimensions increases (ex. Wald test). How the SVD helps in extracting robust correlations between dependent and independent variables: automatic choice of “model”. Why Some evidence in predicting US CPIs indexes Some issues about normalizations

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Motivations Given a simultaneous linear system of equations 1.Collapsing dimensionality of the system to its min rank = min [rank(Y), rank (X)], 2.Advantages of SVD w.r.t. Principal Components: PC requires a sqare matrix, e.g. autocorrelation matrix, and ranks the dimensions within that single matrix; SVD ranks the correlations between X and Y dimensions 3.Discretionary possibility of getting rid of some - believed negligible – dimensions: we are interested in getting rid of those dimensions that can be generated by a totally random system of same dimensions (Marchenko-Pastur conditions adapted to a rectangular matrix).

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Definition of SVD of a matrix product SVD definition Having two matrices one can write and therefore If T << max(M,N)? No problems

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Diagonalizing the LS estimator Consider regressing every column y over the set of explanatory variables X: we write We diagonalize both matrices: (X’X) and (X’Y): –X’X –X’Y rectangular –NB. The SVD of a square matrix IS the same as the diagonalisation. We will write

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September

7 (X’ Y) Uxy 0 Sxy Vxy SVD of the covariance matrix

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September X’Y Vxy Uxy Sxy 0 SVD mapping from column basis to row basis

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Y Vxy X Uxy Sxy Y linear combin X linear combin SVD: splitting the product X’Y

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Adding diagonalisation of both X and Y matrices

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September YXUxx Uxy Inv(Dxx)Sxy Vxy ‘Vyy ’ Returning to the original variables Replacing the old “B”: any advantage??!! We may cancel factors: any criterium?

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September RMT 1.Marcenko-Pastur conditions compute singular values density and interval limits for square matrices. Bouchaud, Miceli et al (2005) derive them for rectangular matrices. 2.We run exactly the same experiment with purely random generated matrices for “many times”: limits and densities reply the theory

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Marcenko-Pastur limits and density

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September RMT 1.Density and limits do change if we use raw or already diagonalized data. 2.Is this “double diagonalization” worthwhile? singular values are HD0 in standardization, eigenvectors are NOT.

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Diagonalized “LS estimator” We may approach the same problem in different ways 1.raw data 2.normalized factors 3.non normalized factors “unfortunately” 3. works best. Why? … Is it because factor normalization changes the ranking of the SVD singular values and this affect eventually the factor selection? NO! Answer at the end …. Very disturbing

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Example: Forecasting US CPIs Indexes Time series are mom % changes: Y:= 9 CPIs Indexes, aug83 – apr07 X:= 77 macroeconomic series nov83-apr07 including 3 lags of the Ys. T=282, N=9, M=77, rolling window W=100 or else. n= N/W, m=M/W.

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September CPIs

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Xs

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Estimation by Model III

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Singular values: Model I – Random generated DATA

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Singular values for SVD on raw and random DATA

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Estimation by Model II Factors are divided by their own eigenvalue

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Singular values: Model II – Data NORMALIZED FACTORS lambda max = Lambda min =0.608

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September lambda max = Lambda min =0.608 Singular values: Model II – Random generated NORMALIZED FACTORS Random generated singular values don’t look very differently ….

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Singular values for SVD on raw and random FACTORS

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Let’s see estimations by Model III

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September P&L Model III - Factors on raw data

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September P&L Model III - CPI Indexes (Model of Non Normalized Factors) – In sample With ALL svd factors2 svd factors

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Let’s see estimations by Model II (normalized factors)

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September P&L Model II (Normalized factors) - Factors

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September P&L Model II (Normalized factors) – CPI’s

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Normalized factorsNon normalized factors Example of CPI_comdty estimation

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September OUT OF SAMPLE Estimation on t=1,…,120 Forecast at fixed coefficients for t= 121, … 282

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September P&L: Factors (Model II)

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Forecast on CPI’s All factors 2 factors only Easier to predict: 1. medical care (since stable), 2. commodities (oil), 3. Transports

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Forecasts on Cpi’s Comdty

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Conclusions 1

M. A. Miceli “SVD and LS” - Stats in the Château - August 31 - September Conclusions on the example