CSSE463: Image Recognition Day 26

Slides:



Advertisements
Similar presentations
Face Recognition Sumitha Balasuriya.
Advertisements

EigenFaces and EigenPatches Useful model of variation in a region –Region must be fixed shape (eg rectangle) Developed for face recognition Generalised.
EigenFaces.
Machine Learning Lecture 8 Data Processing and Representation
As applied to face recognition.  Detection vs. Recognition.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #20.
Principal Component Analysis
Region labelling Giving a region a name. Image Processing and Computer Vision: 62 Introduction Region detection isolated regions Region description properties.
Dimensionality Reduction Chapter 3 (Duda et al.) – Section 3.8
© 2003 by Davi GeigerComputer Vision September 2003 L1.1 Face Recognition Recognized Person Face Recognition.
Data-driven Methods: Faces : Computational Photography Alexei Efros, CMU, Fall 2007 Portrait of Piotr Gibas © Joaquin Rosales Gomez.
Principal Component Analysis
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
CS 790Q Biometrics Face Recognition Using Dimensionality Reduction PCA and LDA M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Face Recognition Jeremy Wyatt.
Face Recognition Using Eigenfaces
FACE RECOGNITION, EXPERIMENTS WITH RANDOM PROJECTION
EECE 279: Real-Time Systems Design Vanderbilt University Ames Brown & Jason Cherry MATCH! Real-Time Facial Recognition.
Face Collections : Rendering and Image Processing Alexei Efros.
Face Detection and Recognition
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Face Recognition Using EigenFaces Presentation by: Zia Ahmed Shaikh (P/IT/2K15/07) Authors: Matthew A. Turk and Alex P. Pentland Vision and Modeling Group,
PCA & LDA for Face Recognition
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Chapter 2 Dimensionality Reduction. Linear Methods
Face Recognition and Feature Subspaces
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Face Recognition and Feature Subspaces
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
CSE 185 Introduction to Computer Vision Face Recognition.
CSSE463: Image Recognition Day 27 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
CS 691B Computational Photography Instructor: Gianfranco Doretto Data Driven Methods: Faces.
CSSE463: Image Recognition Day 10 Lab 3 due Weds Lab 3 due Weds Today: Today: finish circularity finish circularity region orientation: principal axes.
Obama and Biden, McCain and Palin Face Recognition Using Eigenfaces Justin Li.
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
CSSE463: Image Recognition Day 25 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
CSSE463: Image Recognition Day 10 Lab 3 due Weds, 11:59pm Lab 3 due Weds, 11:59pm Take-home quiz due Friday, 4:00 pm Take-home quiz due Friday, 4:00 pm.
Principal Components Analysis ( PCA)
CSSE463: Image Recognition Day 10 Lab 3 due Weds, 3:25 pm Lab 3 due Weds, 3:25 pm Take-home quiz due Friday, 4:00 pm Take-home quiz due Friday, 4:00 pm.
CSSE463: Image Recognition Day 27
Dimensionality Reduction
University of Ioannina
9.3 Filtered delay embeddings
Recognition with Expression Variations
Lecture 8:Eigenfaces and Shared Features
Face Recognition and Feature Subspaces
Lecture: Face Recognition and Feature Reduction
Principal Component Analysis (PCA)
Dimension Reduction via PCA (Principal Component Analysis)
Principal Component Analysis
Data-driven Methods: Faces
CS 332 Visual Processing in Computer and Biological Vision Systems
Object Modeling with Layers
Face Recognition and Detection Using Eigenfaces
Computational Photography
Principal Component Analysis
PCA is “an orthogonal linear transformation that transfers the data to a new coordinate system such that the greatest variance by any projection of the.
Eigenfaces for recognition (Turk & Pentland)
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
Introduction PCA (Principal Component Analysis) Characteristics:
Recitation: SVD and dimensionality reduction
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
X.1 Principal component analysis
CSSE463: Image Recognition Day 25
CSSE463: Image Recognition Day 26
CSSE463: Image Recognition Day 25
Presentation transcript:

CSSE463: Image Recognition Day 26 This week Today: Finding lines and circles using the Hough transform Tomorrow: Applications of PCA Weds night: k-means lab due. Thursday: template matching for object recognition Sunday night (new): project plans and prelim work due Questions?

Principal Components Analysis Given a set of samples, find the direction(s) of greatest variance. We’ve done this! Spatial moments Principal axes are eigenvectors of covariance matrix Eigenvalues gave relative importance of each dimension Note that each point can be represented in 2D using the new coordinate system defined by the eigenvectors The 1D representation obtained by projecting the point onto the principal axis is a reasonably-good approximation size weight girth height

Projecting a point onto a unit vector Assuming the axis is represented by a unit vector, we can just take the dot-product of the point and the vector. u*p = uTp (which is 1D) Example: Project (5,2) onto line y=x. If we want to project onto two vectors, u and v simultaneously: Create w = [u v], then compute wTp, which is 2D. Result: p is now in terms of u and v. This generalizes to arbitrary dimensions. Do example with unit vector, r= 1, theta = 30deg, and the point (0,1) or the point (1,0) Q1

Two applications Eigenfaces Time-elapsed photography

“Eigenfaces” Question: what are the primary ways in which faces vary? What happens when we apply PCA? For each face, create a column vector that contains all the pixels from that face This is a point in a high dimensional space (e.g., 65536 for a 256x256 pixel image) Create a matrix F of all M faces in the training set. Subtract off the “average face”, m, to get N Compute the covariance matrix C = N*NT . Contrast with doing in 2D. This is 65536-D. M. Turk and A. Pentland, Eigenfaces for Recognition, J Cog Neurosci, 3(1) Q2

“Eigenfaces” Question: what are the primary ways in which faces vary? What happens when we apply PCA? The eigenvectors are the directions of greatest variability Note that these are in 65536-D; thus form a face. This is an “eigenface” Here’s the first 4 from the ORL face dataset. Contrast with doing in 2D. This is 65536-D. Q3

“Eigenfaces” Question: what are the primary ways in which faces vary? What happens when we apply PCA? The eigenvectors are the directions of greatest variability Note that these are in 65536-D; thus form a face. This is an “eigenface” Here are the first 4 from the ORL face dataset. Contrast with doing in 2D. This is 65536-D. http://upload.wikimedia.org/wikipedia/commons/6/67/Eigenfaces.png; from the ORL face database, AT&T Laboratories Cambridge

“Eigenfaces” Question: what are the primary ways in which faces vary? What happens when we apply PCA? The top M eigenfaces for “face space”. We can project any face onto these eigenvectors (right?). Thus, any face is a linear combination of the eigenfaces. Can classify faces in this lower-D space. There are computational tricks to make the computation feasible

Time-elapsed photography Question: what are the ways that outdoor images vary over time? Form a matrix in which each column is an image Find eigs of covariance matrix See example images on Dr. B’s laptop or at the link below. N Jacobs, N Roman, R Pless, Consistent Temporal Variations in Many Outdoor Scenes. IEEE Computer Vision and Pattern Recognition, Minneapolis, MN, June 2007.

Time-elapsed photography Question: what are the ways that outdoor images vary over time? The mean and top 3 eigenvectors (scaled): Interpretation? N Jacobs, N Roman, R Pless, Consistent Temporal Variations in Many Outdoor Scenes. IEEE Computer Vision and Pattern Recognition, Minneapolis, MN, June 2007. Q4-5

Time-elapsed photography Recall that each image in the dataset is a linear combination of the eigenimages. mean PC1 PC2 PC3 = + 4912* - 217* +393* = - 2472* + 308* +885* N Jacobs, N Roman, R Pless, Consistent Temporal Variations in Many Outdoor Scenes. IEEE Computer Vision and Pattern Recognition, Minneapolis, MN, June 2007.

Time-elapsed photography Every image’s projection onto the first eigenvector N Jacobs, N Roman, R Pless, Consistent Temporal Variations in Many Outdoor Scenes. IEEE Computer Vision and Pattern Recognition, Minneapolis, MN, June 2007.

Research idea Done: Yet to do: Finding the PCs Using to detect latitude and longitude given images from camera Yet to do: Classifying images based on their projection into this space, as was done for eigenfaces