Representative Previous Work

Slides:



Advertisements
Similar presentations
Active Appearance Models
Advertisements

1 Manifold Alignment for Multitemporal Hyperspectral Image Classification H. Lexie Yang 1, Melba M. Crawford 2 School of Civil Engineering, Purdue University.
DONG XU, MEMBER, IEEE, AND SHIH-FU CHANG, FELLOW, IEEE Video Event Recognition Using Kernel Methods with Multilevel Temporal Alignment.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Human Identity Recognition in Aerial Images Omar Oreifej Ramin Mehran Mubarak Shah CVPR 2010, June Computer Vision Lab of UCF.
A Geometric Perspective on Machine Learning 何晓飞 浙江大学计算机学院 1.
Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning,
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
Discriminative and generative methods for bags of features
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
One-Shot Multi-Set Non-rigid Feature-Spatial Matching
“Random Projections on Smooth Manifolds” -A short summary
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Graph Based Semi- Supervised Learning Fei Wang Department of Statistical Science Cornell University.
Principal Component Analysis
LLE and ISOMAP Analysis of Robot Images Rong Xu. Background Intuition of Dimensionality Reduction Linear Approach –PCA(Principal Component Analysis) Nonlinear.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
DVMM Lab, Columbia UniversityVideo Event Recognition Video Event Recognition: Multilevel Pyramid Matching Dong Xu and Shih-Fu Chang Digital Video and Multimedia.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Enhancing Tensor Subspace Learning by Element Rearrangement
Problem Statement A pair of images or videos in which one is close to the exact duplicate of the other, but different in conditions related to capture,
1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 劉冠成 韓仁智
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Line detection Assume there is a binary image, we use F(ά,X)=0 as the parametric equation of a curve with a vector of parameters ά=[α 1, …, α m ] and X=[x.
1 Learning from Shadows Dimensionality Reduction and its Application in Artificial Intelligence, Signal Processing and Robotics Ali Ghodsi Department of.
Using Support Vector Machines to Enhance the Performance of Bayesian Face Recognition IEEE Transaction on Information Forensics and Security Zhifeng Li,
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Computer examples Tenenbaum, de Silva, Langford “A Global Geometric Framework for Nonlinear Dimensionality Reduction”
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Presented by Xianwang Wang Masashi Sugiyama.
Character Identification in Feature-Length Films Using Global Face-Name Matching IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 11, NO. 7, NOVEMBER 2009 Yi-Fan.
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
Manifold learning: MDS and Isomap
Nonlinear Dimensionality Reduction Approach (ISOMAP)
H. Lexie Yang1, Dr. Melba M. Crawford2
Non-Linear Dimensionality Reduction
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
A Convergent Solution to Tensor Subspace Learning.
Supervisor: Nakhmani Arie Semester: Winter 2007 Target Recognition Harmatz Isca.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
June 25-29, 2006ICML2006, Pittsburgh, USA Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Masashi Sugiyama Tokyo Institute of.
Principal Component Analysis and Linear Discriminant Analysis for Feature Reduction Jieping Ye Department of Computer Science and Engineering Arizona State.
2D-LDA: A statistical linear discriminant analysis for image matrix
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
Spectral Methods for Dimensionality
Nonlinear Dimensionality Reduction
Deeply learned face representations are sparse, selective, and robust
Shuang Hong Yang College of Computing, Georgia Tech, USA Hongyuan Zha
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
Unsupervised Riemannian Clustering of Probability Density Functions
CS 2750: Machine Learning Dimensionality Reduction
Recognition: Face Recognition
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Image Segmentation Techniques
Learning with information of features
Paper Reading Dalong Du April.08, 2011.
Nonlinear Dimension Reduction:
NonLinear Dimensionality Reduction or Unfolding Manifolds
Presentation transcript:

Representative Previous Work PCA LDA ISOMAP: Geodesic Distance Preserving J. Tenenbaum et al., 2000 LLE: Local Neighborhood Relationship Preserving S. Roweis & L. Saul, 2000 LE/LPP: Local Similarity Preserving, M. Belkin, P. Niyogi et al., 2001, 2003

Dimensionality Reduction Algorithms Statistics-based Geometry-based Hundreds … PCA/KPCA LDA/KDA ISOMAP LLE LE/LPP … Matrix Tensor Any common perspective to understand and explain these dimensionality reduction algorithms? Or any unified formulation that is shared by them? Any general tool to guide developing new algorithms for dimensionality reduction?

Our Answers Direct Graph Embedding Linearization Kernelization Original PCA & LDA, ISOMAP, LLE, Laplacian Eigenmap PCA, LDA, LPP KPCA, KDA Tensorization Type Formulation CSA, DATER Example S. Yan, D. Xu, H. Zhang and et al., CVPR, 2005, T-PAMI,2007

Direct Graph Embedding Intrinsic Graph: S, SP: Similarity matrix (graph edge) Similarity in high dimensional space L, B: Laplacian matrix from S, SP; Data in high-dimensional space and low-dimensional space (assumed as 1D space here): Penalty Graph

Direct Graph Embedding -- Continued Intrinsic Graph: S, SP: Similarity matrix (graph edge) L, B: Laplacian matrix from S, SP; Similarity in high dimensional space Data in high-dimensional space and low-dimensional space (assumed as 1D space here): Criterion to Preserve Graph Similarity: Penalty Graph Special case B is Identity matrix (Scale normalization) Problem: It cannot handle new test data.

Linearization Objective function in Linearization Intrinsic Graph Linear mapping function Penalty Graph Objective function in Linearization Problem: linear mapping function is not enough to preserve the real nonlinear structure?

Kernelization Objective function in Kernelization Nonlinear mapping: Intrinsic Graph Nonlinear mapping: the original input space to another higher dimensional Hilbert space. Penalty Graph Constraint: Kernel matrix: Objective function in Kernelization

Tensorization Objective function in Tensorization Intrinsic Graph Low dimensional representation is obtained as: Penalty Graph Objective function in Tensorization where

Common Formulation Direct Graph Embedding Linearization Kernelization Intrinsic graph S, SP: Similarity matrix L, B: Laplacian matrix from S, SP; Penalty graph Direct Graph Embedding Linearization Kernelization Tensorization where

A General Framework for Dimensionality Reduction Algorithm S & B Definition Embedding Type PCA/KPCA/CSA L/K/T LDA/KDA/DATER ISOMAP D LLE LE/LPP if ; B=D D/L D: Direct Graph Embedding L: Linearization K: Kernelization T: Tensorization

New Dimensionality Reduction Algorithm: Marginal Fisher Analysis Important Information for face recognition: 1) Label information 2) Local manifold structure (neighborhood or margin) 1: if xi is among the k1-nearest neighbors of xj in the same class; 0 : otherwise 1: if the pair (i,j) is among the k2 shortest pairs among the data set; 0: otherwise

Marginal Fisher Analysis: Advantage No Gaussian distribution assumption

Experiments: Face Recognition ORL G3/P7 G4/P6 PCA+LDA (Linearization) 87.9% 88.3% PCA+MFA (Ours) 89.3% 91.3% KDA (Kernelization) 87.5% 91.7% KMFA (Ours) 88.6% 93.8% DATER-2 (Tensorization) 92.0% TMFA-2 (Ours) 95.0% 96.3% PIE-1 G3/P7 G4/P6 PCA+LDA (Linearization) 65.8% 80.2% PCA+MFA (Ours) 71.0% 84.9% KDA (Kernelization) 70.0% 81.0% KMFA (Ours) 72.3% 85.2% DATER-2 (Tensorization) 80.0% 82.3% TMFA-2 (Ours) 82.1%

Summary Optimization framework that unifies previous dimensionality reduction algorithms as special cases. A new dimensionality reduction algorithm: Marginal Fisher Analysis.

Event Recognition in News Video Online and offline video search 56 events are defined in LSCOM Airplane Flying Existing Car Riot Geometric and photometric variances Clutter background Complex camera motion and object motion More diverse !

Earth Mover’s Distance in Temporal Domain (T-MM, Under Review) Key Frames of two video clips in class “riot” EMD can efficiently utilize the information from multiple frames.

Multi-level Pyramid Matching (CVPR 2007, Under Review) One Clip = several subclips (stages of event evolution) . No prior knowledge about the number of stages in an event, and videos of the same event may include a subset of stage only. Smoke Fire Level-1 Level-1 Level-0 Level-0 Fire Smoke Level-1 Level-1 Solution: Multi-level Pyramid Matching in Temporal Domain

Other Publications & Professional Activities Kernel based Learning: Coupled Kernel-based Subspace Analysis: CVPR 2005 Fisher+Kernel Criterion for Discriminant Analysis: CVPR 2005 Manifold Learning: Nonlinear Discriminant Analysis on Embedding Manifold : T-CSVT (Accepted) Face Verification: Face Verification with Balanced Thresholds: T-IP (Accepted) Multimedia: Insignificant Shadow Detection for Video Segmentation: T-CSVT 2005 Anchorperson extraction for Picture in Picture News Video: PRL 2005 Guest Editor: Special issue on Video Analysis, Computer Vision and Image Understanding Special issue on Video-based Object and Event Analysis, Pattern Recognition Letters Book Editor: Semantic Mining Technologies for Multimedia Databases Publisher: Idea Group Inc. (www.idea-group.com)

Future Work Machine Learning Computer Vision Pattern Recognition Event Recognition Biometric Computer Vision Pattern Recognition Multimedia Content Analysis Web Search Multimedia

Acknowledgement Shuicheng Yan UIUC Steve Lin Microsoft Lei Zhang Hong-Jiang Zhang Microsoft Zhengkai Liu, USTC Xuelong Li UK Shih-Fu Chang Columbia Xiaoou Tang Hong Kong

Thank You very much!

What is Gabor Features? … Gabor features can improve recognition performance in comparison to grayscale features. Chengjun Liu T-IP, 2002 Five Scales … Input: Grayscale Image Eight Orientations Output: 40 Gabor-filtered Images Gabor Wavelet Kernels

How to Utilize More Correlations? Potential Assumption in Previous Tensor-based Subspace Learning: Intra-tensor correlations: Correlations among the features within certain tensor dimensions, such as rows, columns and Gabor features… Sets of highly correlated pixels Columns of highly Pixel Rearrangement Pixel Rearrangement

Tensor Representation: Advantages Enhanced Learnability 2. Appreciable reductions in computational costs 3. Large number of available projection directions 4. Utilize the structure information This figure shows the procedure of the first step of this algorithm PCA CSA Feature Dimension Sample Number Computation Complexity

Connection to Previous Work –Tensorface (M. Vasilescu and D Connection to Previous Work –Tensorface (M. Vasilescu and D. Terzopoulos, 2002) From an algorithmic view or mathematics view, CSA and Tensorface are both variants of Rank-(R1,R2,…,Rn) decomposition. Tensorface CSA Motivation Characterize external factors Characterize internal factors Input: Gray-level Image Vector Matrix Input: Gabor-filtered Image (Video Sequence ) Not address 3rd-order tensor When equal to PCA The number of images per person are only one or are a prime number Never Number of Images per Person for Training Lots of images per person One image per person This figure shows the procedure of the first step of this algorithm