کاربرد نگاشت با حفظ تنکی در شناسایی چهره

Slides:



Advertisements
Similar presentations
ECG Signal processing (2)
Advertisements

Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
A Geometric Perspective on Machine Learning 何晓飞 浙江大学计算机学院 1.
Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning,
« هو اللطیف » By : Atefe Malek. khatabi Spring 90.
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
A novel supervised feature extraction and classification framework for land cover recognition of the off-land scenario Yan Cui
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
“Random Projections on Smooth Manifolds” -A short summary
Chapter 2: Pattern Recognition
Lecture 4 Unsupervised Learning Clustering & Dimensionality Reduction
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Unsupervised Learning
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
CS Instance Based Learning1 Instance Based Learning.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Radial Basis Function Networks
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Boris Babenko Department of Computer Science and Engineering University of California, San Diego Semi-supervised and Unsupervised Feature Scaling.
Graph Embedding: A General Framework for Dimensionality Reduction Dong XU School of Computer Engineering Nanyang Technological University
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Transductive Regression Piloted by Inter-Manifold Relations.
CSE 185 Introduction to Computer Vision Face Recognition.
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Principal Manifolds and Probabilistic Subspaces for Visual Recognition Baback Moghaddam TPAMI, June John Galeotti Advanced Perception February 12,
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
June 25-29, 2006ICML2006, Pittsburgh, USA Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Masashi Sugiyama Tokyo Institute of.
Irfan Ullah Department of Information and Communication Engineering Myongji university, Yongin, South Korea Copyright © solarlits.com.
2D-LDA: A statistical linear discriminant analysis for image matrix
CSC321: Lecture 25: Non-linear dimensionality reduction Geoffrey Hinton.
Date of download: 9/18/2016 Copyright © 2016 SPIE. All rights reserved. Key observation of geometric analysis of two different time series construction.
Spectral Methods for Dimensionality
Support Vector Machines
Nonlinear Dimensionality Reduction
Support Feature Machine for DNA microarray data
PREDICT 422: Practical Machine Learning
DEEP LEARNING BOOK CHAPTER to CHAPTER 6
Ch 12. Continuous Latent Variables ~ 12
LECTURE 11: Advanced Discriminant Analysis
Unsupervised Riemannian Clustering of Probability Density Functions
Dimensionality reduction
Recognition: Face Recognition
Principal Component Analysis (PCA)
Machine Learning Basics
Dynamical Statistical Shape Priors for Level Set Based Tracking
Machine Learning Dimensionality Reduction
In summary C1={skin} C2={~skin} Given x=[R,G,B], is it skin or ~skin?
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Image Retrieval Longin Jan Latecki.
Learning with information of features
Goodfellow: Chapter 14 Autoencoders
COSC 4335: Other Classification Techniques
Feature space tansformation methods
Generalized Locality Preserving Projections
CS4670: Intro to Computer Vision
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
NonLinear Dimensionality Reduction or Unfolding Manifolds
Support Vector Machines 2
Marios Mattheakis and Pavlos Protopapas
Random Neural Network Texture Model
Presentation transcript:

کاربرد نگاشت با حفظ تنکی در شناسایی چهره کاربرد نگاشت با حفظ تنکی در شناسایی چهره ارائه دهنده: اسفند 91

Introduction to DRs روش های کاهش ابعاد یا DRs یک راه حل اصلی برای انجام عملیات پردازشی بر روی داده هایی که دارای ابعاد بالایی هستند (از جمله تصاویر چهره) می باشند. روش های زیادی برای کاهش ابعاد تا به حال معرفی شده اند که از بین آنها می توان به روش LPP و روش NPE اشاره کرد. یکی از این روش ها روش SPP است. در این روش با کمک ذخیره رابطه بین داده ها و نمایش تنک آنها می توان علاوه بر کاهش ابعاد که به دنبال آن پیچیدگی محاسباتی نیز کاهش می یابد، اطلاعات جداکننده گروه ها بدون هیچ برچسبی نیز حفظ شود. یکی از ویژگی های ممتاز در این روش، مقاوم بودن آن در برابر چرخش ، تغییر مقیاس و تبدیل داده است.

DRs methods The supervised DRs The unsupervised DRs Linear Discriminant Analysis(LDA) Marginal Fisher Analysis(MFA) Maximum Margin Criterion(MMC) The unsupervised DRs Principal Component Analysis(PCA) Locality Preserving Projections(LPP) The semi-supervised DRs Semi-supervised Dimensionality Reduction(SSDR) Semi-supervised Discriminant Analysis(SDA) In this paper, we only focus on unsupervised scenario

Manifold Learning In the unsupervised DRs, PCA seems to be the most popular one. PCA may fail to discover essential data structures that are nonlinear. Although the kernel-based techniques such as KPCA can implicitly deal with nonlinear DR problems.How to select kernel and assign optimal kernel parameter is generally difficult and unsolved fully in many practical applications. Some desirable virtues the traditional PCA possesses are not inherited. 1) a research has shown that nonlinear techniques perform well on some artificial data sets, but do not necessarily outperform the traditional PCA for real-world tasks yet 2) it is generally difficult to select suitable values for the hyper-parameters (e.g., the neighborhood size) in such models. One effective approach to overcome the above limitations is approximating the nonlinear DRs using linear ones.

Characteristics of SPP SPP shares some advantages of both LPP and many other linear DRs. For example, it is linear and defined everywhere, thus the “out-of-sample” problem is naturally solved. In addition, the weight matrix is kept sparse like in most locality preserving algorithms, which is beneficial to computational tractability. SPP does not have to encounter model parameters such as the neighborhood size and heat kernel width incurred in LPP and NPE, etc, which are generally difficult to set in practice. Although cross-validation technique can be used in these cases, it is very time-consuming and tends to waste the limited training data. In contrast, SPP does not need to deal with such parameters, which makes it very simple to use in practice. Although SPP belongs to global methods in nature, it owns some local properties due to the sparse representation procedure. This technique proposed here can be easily extended to supervised and semi-supervised scenarios based on the existing dimensionality reduction framework.

PCA

LPP

NPE

SPP نگاشت با حفظ تنکی:(SPP) ماتریس دیکشنری: داده های آموزشی هدف: نمایش بردار ورودی دلخواه با استفاده از تعداد کمی از ستون های ماتریس دیکشنری یا همان داده های آموزشی است. شیوه تعیین ماتریس نگاشت:

SPP نگاشت با حفظ تنکی: (SPP)

SPP نگاشت با حفظ تنکی: (SPP)

Implementation

Comparison with Related Works PCA PCA can be seen as a globality preserving DR method in that a single hyper-plane is used to represent the data, hence not facing problem of selecting appropriate neighborhood size. SPP also doesn’t need to worry about this since it actually uses all the training samples to construct the weight matrix without explicitly setting the neighborhood size. But compared to PCA, SPP has an extra advantage that it is capable of implicitly and naturally employing the “local” structure of the data by imposing the sparsity prior

Comparison with Related Works NPE and other Locality preserving DRs SPP has a similar objective function to NPE. Both of them can be closely related to LLE. In fact, NPE is a directly linearized version of LLE, while SPP constructs the “affinity” weight matrix in a completely different manner from LLE. In particular, SPP constructs the weight matrix using all the training samples instead of k nearest neighbors, preventing it from suffering from the difficulty of parameter selection as in the case of NPE and other locality preserving DRs. Despite of such difference, SPP can actually be thought as a regularized extension of NPE through the modified L1-regularization problem. From the Bayesian view, such regularization essentially encodes prior knowledge of sparsity, allowing it to extract more discriminating information from the data than NPE does

toy dataset

toy dataset For PCA, LPP and NPE, the samples from two classes are overlapped together, but the degree of overlapping is different for these methods respectively. The PCA suffers most since it tries to use a single hyper-plane to model the data according to the directions of large sample variance. Both LPP and NPE improve over this by explicitly taking the local structure of data into account. But the Euclidean distance measure and the predefined neighborhood size in these methods fail to identify the real local structure they are supposed. On the contrary, one can see that, with SPP, the two classes can be perfectly separated in the low-dimensional subspace. This can be explained from the angle of sparse prior, which assumes that a point should be best explained by a set of samples as small as possible. Further, this illustrates that SPP can effectively and implicitly use the subspace assumption, even when the two classes are close to each other and the data are noisy.

Wine dataset from UCI 1) The leading projection directions of PCA are decided by the 13th feature. Thus the 2D data distribute in a large range along the direction dominated by the last feature, which makes the projected data mixed up. 2) The locality preserving methods such as LPP and NPE suffer from the same problem as in PCA, even though the local information is considered. This is because the neighborhood of a certain sample point is still dominated by the 13th feature due to its large variance and range. 3) The data projected by SPP form a point-cloud distribution instead of a “linear” one. This reflects that the SPP gives other features besides the 13th one a chance to play a role in capturing a reasonable structure of the data.

Resulting From the two illustrative examples, we can see that sparsity actually works, especially when the data from each class lie on a subspace. Furthermore, SPP is not quite sensitive to the imbalance of the feature distribution which incurs the failure of LPP and NPE, due to the fact that neighborhood is mainly decided by the features with large range and variance.

Face representation and recognition

Face representation and recognition

Face representation and recognition

Face representation and recognition

با تشکر

نگاشت با حفظ تنکی: (SPP)