Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction

Slides:



Advertisements
Similar presentations
Text mining Gergely Kótyuk Laboratory of Cryptography and System Security (CrySyS) Budapest University of Technology and Economics
Advertisements

Eigen Decomposition and Singular Value Decomposition
Eigen Decomposition and Singular Value Decomposition
3D Geometry for Computer Graphics
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
EigenFaces.
Dimensionality Reduction PCA -- SVD
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
Lecture 19 Singular Value Decomposition
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #20.
Symmetric Matrices and Quadratic Forms
© 2003 by Davi GeigerComputer Vision September 2003 L1.1 Face Recognition Recognized Person Face Recognition.
Principal Component Analysis
Lecture 16 Graphs and Matrices in Practice Eigenvalue and Eigenvector Shang-Hua Teng.
Lecture 19 Quadratic Shapes and Symmetric Positive Definite Matrices Shang-Hua Teng.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
TFIDF-space  An obvious way to combine TF-IDF: the coordinate of document in axis is given by  General form of consists of three parts: Local weight.
Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction
Singular Value Decomposition
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
SLIDE 1IS 240 – Spring 2007 Prof. Ray Larson University of California, Berkeley School of Information Tuesday and Thursday 10:30 am - 12:00.
3D Geometry for Computer Graphics
Lecture 20 SVD and Its Applications Shang-Hua Teng.
DATA MINING LECTURE 7 Dimensionality Reduction PCA – SVD
Lecture 16 Cramer’s Rule, Eigenvalue and Eigenvector Shang-Hua Teng.
Subspaces, Basis, Dimension, Rank
SVD(Singular Value Decomposition) and Its Applications
Machine Learning CS 165B Spring Course outline Introduction (Ch. 1) Concept learning (Ch. 2) Decision trees (Ch. 3) Ensemble learning Neural Networks.
Chapter 2 Dimensionality Reduction. Linear Methods
Recognition Part II Ali Farhadi CSE 455.
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
AN ORTHOGONAL PROJECTION
An Introduction to Latent Semantic Analysis. 2 Matrix Decompositions Definition: The factorization of a matrix M into two or more matrices M 1, M 2,…,
SINGULAR VALUE DECOMPOSITION (SVD)
Face Recognition: An Introduction
CSE 185 Introduction to Computer Vision Face Recognition.
1 Latent Concepts and the Number Orthogonal Factors in Latent Semantic Analysis Georges Dupret
A Convergent Solution to Tensor Subspace Learning.
1. Systems of Linear Equations and Matrices (8 Lectures) 1.1 Introduction to Systems of Linear Equations 1.2 Gaussian Elimination 1.3 Matrices and Matrix.
2D-LDA: A statistical linear discriminant analysis for image matrix
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Eigen & Singular Value Decomposition
CS276A Text Information Retrieval, Mining, and Exploitation
Document Clustering Based on Non-negative Matrix Factorization
University of Ioannina
LECTURE 10: DISCRIMINANT ANALYSIS
Recognition with Expression Variations
Lecture 8:Eigenfaces and Shared Features
Face Recognition and Feature Subspaces
Lecture: Face Recognition and Feature Reduction
Recognition: Face Recognition
LSI, SVD and Data Management
PCA vs ICA vs LDA.
Packing to fewer dimensions
Singular Value Decomposition
Lecture 16 Cramer’s Rule, Eigenvalue and Eigenvector
Object Modeling with Layers
SVD: Physical Interpretation and Applications
Recitation: SVD and dimensionality reduction
Concept Decomposition for Large Sparse Text Data Using Clustering
Outline Singular Value Decomposition Example of PCA: Eigenfaces.
FACE RECOGNITION USING LAPLACIANFACES
Symmetric Matrices and Quadratic Forms
CS4670: Intro to Computer Vision
LECTURE 09: DISCRIMINANT ANALYSIS
Lecture 13: Singular Value Decomposition (SVD)
Packing to fewer dimensions
Lecture 20 SVD and Its Applications
Symmetric Matrices and Quadratic Forms
Presentation transcript:

Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction Shang-Hua Teng

Singular Value Decomposition where u1 …ur are the r orthonormal vectors that are basis of C(A) and v1 …vr are the r orthonormal vectors that are basis of C(AT )

Low Rank Approximation and Reduction

The Singular Value Decomposition · · A U S VT = m x n m x m m x n n x n A U S VT = m x n m x r r x r r x n

The Singular Value Reduction · · A U VT m x n m x r r x r r x n = S Ak Uk S VkT = m x n m x k k x k k x n

How Much Information Lost?

Distance between Two Matrices Frobenius Norm of a matrix A. Distance between two matrices A and B

How Much Information Lost?

Approximation Theorem [Schmidt 1907; Eckart and Young 1939] Among all m by n matrices B of rank at most k, Ak is the one that minimizes

Application: Image Compression Uncompressed m by n pixel image: m×n numbers Rank k approximation of image: k singular values The first k columns of U (m-vectors) The first k columns of V (n-vectors) Total: k × (m + n + 1) numbers

Example: Yogi (Uncompressed) Source: [Will] Yogi: Rock photographed by Sojourner Mars mission. 256 × 264 grayscale bitmap  256 × 264 matrix M Pixel values  [0,1] ~ 67584 numbers

Example: Yogi (Compressed) M has 256 singular values Rank 81 approximation of M: 81 × (256 + 264 + 1) = ~  42201 numbers

Example: Yogi (Both)

Eigenface Patented by MIT Utilizes two dimentional, global grayscale images Face is mapped to numbers Create an image subspace(face space) which best discriminated between faces Can be used in properly lit and only frontal images

The Face Database Set of normalized face images Used ORL Face DB

Two-dimensional Embedding We also conducted the locality preserving projections on the face manifold, which are based on the face images of one person with different poses and expressions. The horizontal axis demonstrates the expression variation; The vertical axis shows the pose variation;

EigenFaces Eigenface (PCA) 2’ In this slide, the small images are the first 10 Eigenfaces, Fisherfaces and Laplacianfaces calculated from the face images in the YALE database. We can see the eigenfaces are somewhat like a face since they are derived in the sense of reconstruction; while the fisherfaces and laplacianfaces are not so clear as the eigenfaces. We conducted different experiments to evaluate our algorithm compared with the traditional linear subspace algorithms Eigenfaces, Fisherfaces. There are three databases are evaluated: YALE, PIE and the database collected at Microsoft Research Asia. For the PIE database, we only used the 5 poses near to the frontal view. On the Yale database, we conducted two types of experiments, Leave one out ant the experiment using 6 images each person for training and the other 5 for testing. we use the nearest neighbor classifier for classification. Since Laplacianface method preserves local neighborhood structure. We can see that in all the experiments, Fisherfaces outperform Eigenfaces, and Laplacianfaces perform better than Fisherfaces especially on the MSRA database. I think it is because that there are sufficient samples for each person in the database and manifold structure can be well represented by these samples.

Latent Semantic Analysis (LSA) Latent Semantic Indexing (LSI) Principal Component Analysis (PCA)

Term-Document Matrix Index each document (by human or by computer) fij counts, frequencies, weights, etc Each document can be regarded as a point in m dimensions

Document-Term Matrix Index each document (by human or by computer) fij counts, frequencies, weights, etc Each document can be regarded as a point in n dimensions

Term Occurrence Matrix

c1 c2 c3 c4 c5 m1 m2 m3 m4 human 1 1 interface 1 1 computer 1 1 user 1 1 1 system 1 1 2 response 1 1 time 1 1 EPS 1 1 survey 1 1 trees 1 1 1 graph 1 1 1 minors 1 1

Another Example

Term Document Matrix

LSI using k=2… T D “applications & algorithms” LSI Factor 2 LSI Factor 1 “differential equations” T Each term’s coordinates specified in first K values of its row. Each doc’s coordinates specified in first K values of its column. D

Positive Definite Matrices and Quadratic Shapes

Positive Definite Matrices and Quadratic Shapes For any m x n matrix A, all eigenvalues of AAT and ATA are non-negative Symmetric matrices that have positive eigenvalues are called Positive Definite matrices Symmetric matrices that have non-negative eigenvalues are called Positive semi-definite matrices