Proceedings of the 2007 SIAM International Conference on Data Mining.

Slides:



Advertisements
Similar presentations
Text mining Gergely Kótyuk Laboratory of Cryptography and System Security (CrySyS) Budapest University of Technology and Economics
Advertisements

Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
CS Statistical Machine learning Lecture 13 Yuan (Alan) Qi Purdue CS Oct
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
The UNIVERSITY of Kansas EECS 800 Research Seminar Mining Biological Data Instructor: Luke Huan Fall, 2006.
A Probabilistic Framework for Semi-Supervised Clustering
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
© 2003 by Davi GeigerComputer Vision September 2003 L1.1 Face Recognition Recognized Person Face Recognition.
Principal Component Analysis
An Introduction to Kernel-Based Learning Algorithms K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda and B. Scholkopf Presented by: Joanna Giforos CS8980: Topics.
Unsupervised Learning: Clustering Some material adapted from slides by Andrew Moore, CMU. Visit for
Principal Component Analysis IML Outline Max the variance of the output coordinates Optimal reconstruction Generating data Limitations of PCA.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Spatial Semi- supervised Image Classification Stuart Ness G07 - Csci 8701 Final Project 1.
Implementing a reliable neuro-classifier
FACE RECOGNITION, EXPERIMENTS WITH RANDOM PROJECTION
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Lecture outline Support vector machines. Support Vector Machines Find a linear hyperplane (decision boundary) that will separate the data.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
Dimensionality reduction Usman Roshan CS 675. Supervised dim reduction: Linear discriminant analysis Fisher linear discriminant: –Maximize ratio of difference.
Summarized by Soo-Jin Kim
Recognition Part II Ali Farhadi CSE 455.
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
Xiangnan Kong,Philip S. Yu Department of Computer Science University of Illinois at Chicago KDD 2010.
Domain Range definition: T is a linear transformation, EIGENVECTOR EIGENVALUE.
Face Recognition: An Introduction
Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Presented by Xianwang Wang Masashi Sugiyama.
Chapter 7 Multivariate techniques with text Parallel embedded system design lab 이청용.
CSE 185 Introduction to Computer Vision Face Recognition.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Reduces time complexity: Less computation Reduces space complexity: Less parameters Simpler models are more robust on small datasets More interpretable;
Dimensionality reduction
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
June 25-29, 2006ICML2006, Pittsburgh, USA Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Masashi Sugiyama Tokyo Institute of.
Irfan Ullah Department of Information and Communication Engineering Myongji university, Yongin, South Korea Copyright © solarlits.com.
Principal Component Analysis and Linear Discriminant Analysis for Feature Reduction Jieping Ye Department of Computer Science and Engineering Arizona State.
Clustering With Constraints Feasibility Issues and the k-Means Algorithm 報告者:林俞均 日期: 2014/8/27.
Ultra-high dimensional feature selection Yun Li
Advanced Artificial Intelligence Lecture 8: Advance machine learning.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
Document Clustering with Prior Knowledge Xiang Ji et al. Document Clustering with Prior Knowledge. SIGIR 2006 Presenter: Suhan Yu.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Septian Adi Wijaya – Informatics Brawijaya University
Principal Component Analysis (PCA)
PREDICT 422: Practical Machine Learning
Semi-Supervised Clustering
Dimensionality Reduction
Background on Classification
Recognition with Expression Variations
کاربرد نگاشت با حفظ تنکی در شناسایی چهره
Dimensionality reduction
Face Recognition and Feature Subspaces
Recognition: Face Recognition
Support Vector Machines Introduction to Data Mining, 2nd Edition by
Techniques for studying correlation and covariance structure
Principal Component Analysis
Introduction PCA (Principal Component Analysis) Characteristics:
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Dimensionality Reduction
Feature space tansformation methods
Principal Component Analysis
Nonlinear Dimension Reduction:
INTRODUCTION TO Machine Learning
Presentation transcript:

Proceedings of the 2007 SIAM International Conference on Data Mining

Abstract The paper studies semi-supervised dimensionality reduction. Besides unlabeled samples, must-link and cannot-link constraints are incorporated as domain knowledge. SSDR algorithm: preserves structure of data as well as constraints in the projected low-dimension space.

Introduction There exist supervised and unsupervised dimensionality reduction methods FLD (Fisher Linear Discriminant): extracts discriminant vectors when class labels are available cFLD (Constrained FLD): dimensionality reduction from equivalence constraints PCA (Principal Component Analysis): preserves the global covariance structure of data when class labels are not available

Introduction (cont) SSDR: Must-link constraints: pairs of instances belonging to the same class Cannot-link constraints: pairs of instances belonging to different classes Structure of data SSDR: simultaneously preserves the structure of data and pairwise constraints specified by users

SSDR Algorithm Maximizing objective function: Find project vector W: Subject to: w T w = 1 ???

SSDR Algorithm (cont) Extended objective function: Final form of extended objective function: (2.5) is a typical eigen-problem, which can be solved by computing the eigenvectors of XLX T corresponding to the largest eigenvalues.

Experiments Data sets: 6 UCI data sets, YaleB facial image data set, 20-Newsgroup. Results are averaged over 100 runs with different generation of constraints. Parameters: α = 1, β = 20.

Results on UCI Data Sets

Results on UCI Data Sets (cont)