Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.

Slides:



Advertisements
Similar presentations
EigenFaces and EigenPatches Useful model of variation in a region –Region must be fixed shape (eg rectangle) Developed for face recognition Generalised.
Advertisements

Component Analysis (Review)
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Face Recognition Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Machine Learning Lecture 8 Data Processing and Representation
Face Recognition By Sunny Tang.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #20.
Dimensionality Reduction Chapter 3 (Duda et al.) – Section 3.8
© 2003 by Davi GeigerComputer Vision September 2003 L1.1 Face Recognition Recognized Person Face Recognition.
Principal Component Analysis
CS 790Q Biometrics Face Recognition Using Dimensionality Reduction PCA and LDA M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Eigenfaces As we discussed last time, we can reduce the computation by dimension reduction using PCA –Suppose we have a set of N images and there are c.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Project 4 out today –help session today –photo session today Project 2 winners Announcements.
Face Recognition Jeremy Wyatt.
Computer Vision I Instructor: Prof. Ko Nishino. Today How do we recognize objects in images?
Face Detection and Recognition
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Dimensionality reduction Usman Roshan CS 675. Supervised dim reduction: Linear discriminant analysis Fisher linear discriminant: –Maximize ratio of difference.
Summarized by Soo-Jin Kim
Machine Learning CS 165B Spring Course outline Introduction (Ch. 1) Concept learning (Ch. 2) Decision trees (Ch. 3) Ensemble learning Neural Networks.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Probability of Error Feature vectors typically have dimensions greater than 50. Classification accuracy depends upon the dimensionality and the amount.
Recognition Part II Ali Farhadi CSE 455.
Presented By Wanchen Lu 2/25/2013
Face Recognition and Feature Subspaces
This week: overview on pattern recognition (related to machine learning)
Face Recognition and Feature Subspaces
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
Face Recognition: An Introduction
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
CSE 185 Introduction to Computer Vision Face Recognition.
Final Review Course web page: vision.cis.udel.edu/~cv May 21, 2003  Lecture 37.
Lecture 10 Pattern Recognition and Classification II
Dimensionality reduction
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Feature Selection and Dimensionality Reduction. “Curse of dimensionality” – The higher the dimensionality of the data, the more data is needed to learn.
Principal Component Analysis and Linear Discriminant Analysis for Feature Reduction Jieping Ye Department of Computer Science and Engineering Arizona State.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
2D-LDA: A statistical linear discriminant analysis for image matrix
Face Recognition and Feature Subspaces Devi Parikh Virginia Tech 11/05/15 Slides borrowed from Derek Hoiem, who borrowed some slides from Lana Lazebnik,
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
3D Face Recognition Using Range Images Literature Survey Joonsoo Lee 3/10/05.
Irena Váňová. B A1A1. A2A2. A3A3. repeat until no sample is misclassified … labels of classes Perceptron algorithm for i=1...N if then end * * * * *
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Background on Classification
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
University of Ioannina
Dimensionality Reduction
Lecture 8:Eigenfaces and Shared Features
CS 2750: Machine Learning Dimensionality Reduction
Face Recognition and Feature Subspaces
Recognition: Face Recognition
Machine Learning Dimensionality Reduction
Principal Component Analysis
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
In summary C1={skin} C2={~skin} Given x=[R,G,B], is it skin or ~skin?
Techniques for studying correlation and covariance structure
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Feature space tansformation methods
CS4670: Intro to Computer Vision
Announcements Project 2 artifacts Project 3 due Thursday night
Announcements Project 4 out today Project 2 winners help session today
Announcements Artifact due Thursday
Announcements Artifact due Thursday
The “Margaret Thatcher Illusion”, by Peter Thompson
Presentation transcript:

Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33

Announcements Read Forsyth & Ponce Chapter on neural networks for Wednesday HW 6 assigned today is due on Wednesday, May 21 Prof. Kambhamettu will give a guest lecture on Friday on deformable contours (aka snakes) –Read pp of Trucco & Verri (PDF file on course page)

Outline Performance measurement Dimensionality reduction Face recognition –Nearest neighbor –Eigenfaces Homework 6

Supervised Learning: Measuring Performance Testing on the training data is not a reliable indicator of classifier performance –“Trusting” training data too much leads to bias (aka overfitting) Want some indicator of classifier generality Validation: Split data into training and test set –Training set: Labeled data points used to guide parametrization of classifier % misclassified guides learning –Test set: Labeled data points left out of training procedure % misclassified taken to be overall classifier error

Supervised Learning: Cross-Validation m -fold cross-validation –Randomly split data into m equal-sized subsets –Train m times on m - 1 subsets, test on left-out subset –Error is mean test error over left-out subsets Leave-one-out: Cross-validation with 1-point subsets –Very accurate but expensive; variance allows confidence measuring

Feature Selection What features to use? How do we extract them from the image? –Using images themselves as feature vectors is easy, but has problem of high dimensionality A 128 x 128 image = 16,384-dimensional feature space! What do we know about the structure of the categories in feature space? –Intuitively, we want features that result in well- separated classes

Dimensionality Reduction Functions y i = y i (x) can reduce dimensionality of feature space  More efficient classification If chosen intelligently, we won’t lose much information and classification is easier Common methods –Principal components analysis (PCA): Projection maximizing total variance of data –Fisher’s Linear Discriminant (FLD): Maximize ratio of between-class variance to within-class variance Closely related to canonical variates

Geometric Interpretation of Covariance adapted from Z. Dodds Covariance C = X X T can be thought of as linear transform that redistributes variance of unit normal distribution, where zero-mean X is n (number of dimensions) x d (number of points) C

Geometric Factorization of Covariance SVD of covariance matrix C = R T D R describes geometric components of transform by extracting: –Diagonal scaling matrix D –Rotation matrix R E.g., given points, the covariance factors as X = X X T = = cos, sin of 70  major, minor axis lengths “best” axis 2 nd -best axis adapted from Z. Dodds

PCA for Dimensionality Reduction Any point in n -dimensional original space can thus be expressed as a linear combination of the n eigenvectors (the rows of R ) via a set of weights [  1,  2, …,  n ] By projecting points onto only the first k << n principal components (eigenvectors with the largest eigenvalues), we are essentially throwing away the least important feature information

Projection onto Principal Components Full n -dimensional space (here n = 2) k -dimensional subspace (here k = 1) adapted from Z. Dodds

Dimensionality Reduction: PCA vs. FLD from Belhumeur et al., 1996

Face Recognition 20 faces (i.e., classes), 9 examples (i.e., training data) of each ?

Simple Face Recognition Idea: Search over training set for most similar image (e.g., in SSD sense) and choose its class This is the same as a 1-nearest neighbor classifier when feature space = image space Issues –Large storage requirements ( nd, where n = image space dimensionality and d = number of faces in training set) –Correlation is computationally expensive

Eigenfaces Idea: Compress image space to “face space” by projecting onto principal components (“eigenfaces” = eigenvectors of image space) –Represent each face as a low-dimensional vector (weights on eigenfaces) –Measure similarity in face space for classification Advantage: Storage requirements are (n + d) k instead of nd

Eigenfaces: Initialization Calculate eigenfaces –Compute n –dimensional mean face  –Compute difference of every face from mean face  j =  j -  –Form covariance matrix of these C = AA T, where A = [  1,  2, …,  d ] –Extract eigenvectors u i from C such that Cu i = i u i –Eigenfaces are k eigenvectors with largest eigenvalues Example eigenfaces

Eigenfaces: Initialization Project faces into face space –Get eigenface weights for every face in the training set –The weights [  j1,  j2, …,  jk ] for face j are computed via dot products  ji = u i T  j

Calculating Eigenfaces Obvious way is to perform SVD of covariance matrix, but this is often prohibitively expensive –E.g., for 128 x 128 images, C is 16,384 x 16,384 Consider eigenvector decomposition of d x d matrix A T A : A T Av i = i v i. Multiplying both sides on the left by A, we have AA T Av i = i Av i So u i = Av i are the eigenvectors of C = AA T C

Eigenfaces: Recognition Project new face into face space Classify –Assign class of nearest face from training set –Or, precalculate class means over training set and find nearest mean class face adapted from Z. Dodds 8 eigenfaces Original face Weights [  1,  2, …,  8 ]

Homework 6 Eigenfaces for face recognition Graduate students: Hough transform