1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 劉冠成 韓仁智

Slides:



Advertisements
Similar presentations
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Advertisements

Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning,
Dimension reduction (1)
Face Recognition By Sunny Tang.
1 Fast Asymmetric Learning for Cascade Face Detection Jiaxin Wu, and Charles Brubaker IEEE PAMI, 2008 Chun-Hao Chang 張峻豪 2009/12/01.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
One-Shot Multi-Set Non-rigid Feature-Spatial Matching
Pattern Recognition and Machine Learning
Dimensionality Reduction Chapter 3 (Duda et al.) – Section 3.8
Principal Component Analysis
CS 790Q Biometrics Face Recognition Using Dimensionality Reduction PCA and LDA M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Eigenfaces As we discussed last time, we can reduce the computation by dimension reduction using PCA –Suppose we have a set of N images and there are c.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Comparison and Combination of Ear and Face Images in Appearance-Based Biometrics IEEE Trans on PAMI, VOL. 25, NO.9, 2003 Kyong Chang, Kevin W. Bowyer,
Face Recognition: An Introduction
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
8/10/ RBF NetworksM.W. Mak Radial Basis Function Networks 1. Introduction 2. Finding RBF Parameters 3. Decision Surface of RBF Networks 4. Comparison.
Dimensionality reduction Usman Roshan CS 675. Supervised dim reduction: Linear discriminant analysis Fisher linear discriminant: –Maximize ratio of difference.
Representative Previous Work
Summarized by Soo-Jin Kim
Enhancing Tensor Subspace Learning by Element Rearrangement
BDPCA Plus LDA: A Novel Fast Feature Extraction Technique for Face Recognition 授課教授 : 連震杰 老師 組員 : 黃彥綸 何域禎 W. Zuo, D. Zhang, J. Yang, K. Wang, “BBPCA plus.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Probability of Error Feature vectors typically have dimensions greater than 50. Classification accuracy depends upon the dimensionality and the amount.
Recognition Part II Ali Farhadi CSE 455.
Face Recognition and Feature Subspaces
Face Recognition and Feature Subspaces
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
Face Recognition: An Introduction
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
CSE 185 Introduction to Computer Vision Face Recognition.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Linear Models for Classification
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
Feature extraction using fuzzy complete linear discriminant analysis The reporter : Cui Yan
A Convergent Solution to Tensor Subspace Learning.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Principal Component Analysis and Linear Discriminant Analysis for Feature Reduction Jieping Ye Department of Computer Science and Engineering Arizona State.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
2D-LDA: A statistical linear discriminant analysis for image matrix
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
Feature Extraction 主講人:虞台文.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Dimension reduction (2) EDR space Sliced inverse regression Multi-dimensional LDA Partial Least Squares Network Component analysis.
Spectral Methods for Dimensionality
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Recognition with Expression Variations
Face Recognition and Feature Subspaces
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
PCA vs ICA vs LDA.
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Learning with information of features
Principal Component Analysis
Presented by: Chang Jia As for: Pattern Recognition
FACE RECOGNITION USING LAPLACIANFACES
Feature space tansformation methods
Lecture 16. Classification (II): Practical Considerations
Presentation transcript:

1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 劉冠成 韓仁智

2 Outline 1.Introduction 2.System Flowchart 3.Dimensionality Reduction- Graph Embedding 3.1 Cost Function -Intrinsic Graph/Penalty Graph 3.2 Linearization 3.3 Example: LDA 4.Marginal Fisher Analysis 5.Experiments Result

3 Outline 1.Introduction 2.System Flowchart 3.Dimensionality Reduction- Graph Embedding 3.1 Cost Function -Intrinsic Graph/Penalty Graph 3.2 Linearization 3.3 Example: LDA 4.Marginal Fisher Analysis 5.Experiments Result

4 1. Introduction We present a general framework called Graph Embedding (GE). In graph embedding, the underlying merits and shortcomings of different dimensionality reduction schemes, existing or new, are revealed by differences in the design of their intrinsic and penalty graphs and their types of embedding. A novel dimensionality reduction algorithm, Marginal Fisher Analysis (MFA).

5 Outline 1.Introduction 2.System Flowchart 3.Dimensionality Reduction- Graph Embedding 3.1 Cost Function -Intrinsic Graph/Penalty Graph 3.2 Linearization 3.3 Example: LDA 4.Marginal Fisher Analysis 5.Experiments Result

6 2. Face Recognition Flowchart N : # I mage (200, 20 image pre person) : # People (10) m : Image Size (24x24) w : Unitary Linear Projection Vector k : k nearest neighbors 1. Training Image Set: X 2.1 MFA Space Creation: w 2.2 Projection to MFA Space: Y 3. k-NN Classification 1. Test image: x test Classification Result Training ProcessTest Process 2. Projection to MFA Space: y test influence k1 : k -NN for intrinsic graph k2 : k -NN for penalty graph

7 Outline 1.Introduction 2.System Flowchart 3.Dimensionality Reduction- Graph Embedding 3.1 Cost Function -Intrinsic Graph/Penalty Graph 3.2 Linearization 3.3 Example: LDA 4.Marginal Fisher Analysis 5.Experiments Result

8 3. Graph Embedding For a dimensionality reduction problem, we require an intrinsic graph G and, optionally, a penalty graph as input. We now introduce the dimensionality reduction problem from the new point of view of graph embedding. Let G={X,W} be an undirected weighted graph (two-way direction) with vertex set X (N nodes) and similarity (or weighted) matrix. W : 1.Symmetric matrix 2.May be negative

9 3. Graph Embedding: Laplacian Matrix L=Degree-Adjacent=> (2) W is weighted matrix,also call similarity matrix G={X,W} :

Cost Function (1/2) Our graph-preserving criterion is given as follows: For larger (positive) similarity samples and : For smaller (negative) similarity samples and : B typically is the Laplacian matrix of a penalty graph. Use Lagrange multipliers to solve: (3) Y must be an eigenvector of the Intrinsic Graph Penalty Graph unknownKnown

Cost Function: Intrinsic Graph in LDA If we have 6 images for 2 people. is not important, because in GE i always not equal to j.

Cost Function: Penalty Graph We define an intrinsic graph to be the graph G itself Penalty graph : 1.As a graph whose vertices X are the same as those of G. 2.Whose edge weight matrix corresponds to the similarity characteristic that is to be suppressed in the dimension- reduced feature space. 3.Penalty graph = constraint

Cost Function: Penalty Graph in LDA If we have 6 images for 2 people. Maximize class covariance is equal to maximize data covariance. Here, LDA penalty graph is as PCA intrinsic graph: Consider only btw-class scatter.

Linearization Assuming that the low-dimensional vector representations of the vertices can be obtained from a linear projection as, where w is the unitary projection vector, the objective function in (3) becomes : (4) Solution: w can be solved by singular value decomposition (SVD)

15 Outline 1.Introduction 2.System Flowchart 3.Dimensionality Reduction- Graph Embedding 3.1 Cost Function -Intrinsic Graph/Penalty Graph 3.2 Linearization 3.3 Example: LDA 4.Marginal Fisher Analysis (MFA) 5.Experiments Result

Marginal Fisher Analysis (MFA) (1/3) Fig. 4. The adjacency relationships of the intrinsic and penalty graphs for the Marginal Fisher Analysis algorithm. k-NN : k 1 =5 In same class k-NN : k 2 =2 In difference class Marginal k-NN (Within-Class) (Btw-Class)

17 (15) Cost function: Minimize within-class (intrinsic graph) Maximize between- class (penalty graph) (13) Intrinsic graph 4.1 Marginal Fisher Analysis (MFA) (2/3) indicates the index set of the k 1 nearest neighbors of the sample x i in the same class. L k-NN : k 1 =5 In same class

Marginal Fisher Analysis (MFA) (3/3) (14) Penalty graph How to decide k 1 and k 2 -nearest neighbor : k-nearest neighbor (k-NN) is a set of data pairs that are the k 2 nearest pairs among the set denote the index set belonging to the c th class B

Q&A (1/2) Q1: MFA: How to Decide k 1 k 2 A1:  k 1 : (Within Class) Sampled five values between two and {min c (n c -1)}, and chose the value with the best MFA performance. n c : # of images per class (subject) (We direct use 5)  k 2 : (Btw-Class) Choose the best k 2 between 20 and 8N c at sampled intervals of 20. N c : # of classes (subjects) (We direct use 20)

20 MFA: Comparison with LDA Advantages : 1.The number of available projection directions (axes) is much larger than that of LDA (MFA finds more significant axes and has better classification results. ). MFA: Rank (B -1 L) LDA: N c -1 2.There is no assumption on the data distribution, thus it is more general for discriminant analysis, LDA assumption data is approximately Gaussian distributed. Data distribution: MFA: Non-linear LDA: Linear 3.The inter-class margin can better characterize the separability of different classes than the inter-class scatter in LDA. MFA: Maximize margin LDA: Difference between means Disadvantage: LDA-> incremental LDA, MFA->? 4.1 Q&A (2/2) margin Positive Negative

21 Outline 1.Introduction 2.System Flowchart 3.Dimensionality Reduction- Graph Embedding 3.1 Cost Function -Intrinsic Graph/Penalty Graph 3.2 Linearization 3.3 Example: LDA 4.Marginal Fisher Analysis 5.Experiments Result

22 5. Experiments: Database (1/2) DadabaseYale B People10 #Image 30 per person, random select 20 image for training, and remain 10 image for test (G30/P20) Image size24x24 VariationsVariable illumination, cropped face 24x24

23 5. Experiments Result k-NNk=1k=3k=5 ErrorRate11.72%± %± ±3.27 For each k run 100 times, and calculate mean and standard deviation.

24 Reference 1.P. Belhumeur, J. Hespanha, and D. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,” IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol. 19, No. 7, pp. 711–720, T.K. Kim, S.F. Wong, B. Stenger, J. Kittler and R. Cipolla, “Incremental Linear Discriminant Analysis Using Sufficient Spanning Set Approximations”, CVPR, pp. 1-8, S. Yan, D. Xu, B. Zhang, H. Zhang, Q. Yang, and S. Lin, “Graph Embedding and Extensions: A General Framework for Dimensionality Reduction,” IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol. 29, No. 1, pp. 40–51, 2007.