Approximate Nearest Subspace Search with Applications to Pattern Recognition Ronen Basri, Tal Hassner, Lihi Zelnik-Manor presented by Andrew Guillory and.

Slides:



Advertisements
Similar presentations
Linear Subspaces - Geometry. No Invariants, so Capture Variation Each image = a pt. in a high-dimensional space. –Image: Each pixel a dimension. –Point.
Advertisements

k-Nearest Neighbors Search in High Dimensions
Nearest Neighbor Search in High Dimensions Seminar in Algorithms and Geometry Mica Arie-Nachimson and Daniel Glasner April 2009.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Sparse Modeling for Finding Representative Objects Ehsan Elhamifar Guillermo Sapiro Ren´e Vidal Johns Hopkins University University of Minnesota Johns.
Searching on Multi-Dimensional Data
Cse 521: design and analysis of algorithms Time & place T, Th pm in CSE 203 People Prof: James Lee TA: Thach Nguyen Book.
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
A novel supervised feature extraction and classification framework for land cover recognition of the off-land scenario Yan Cui
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Coherency Sensitive Hashing (CSH) Simon Korman and Shai Avidan Dept. of Electrical Engineering Tel Aviv University ICCV2011 | 13th International Conference.
Dimensionality reduction. Outline From distances to points : – MultiDimensional Scaling (MDS) – FastMap Dimensionality Reductions or data projections.
Computational Support for RRTs David Johnson. Basic Extend.
Principal Component Analysis
3D Geometry for Computer Graphics. 2 The plan today Least squares approach  General / Polynomial fitting  Linear systems of equations  Local polynomial.
Dimensionality Reduction and Embeddings
Instance Based Learning
Dimensionality Reduction
CS 790Q Biometrics Face Recognition Using Dimensionality Reduction PCA and LDA M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Approximate Nearest Neighbors and the Fast Johnson-Lindenstrauss Transform Nir Ailon, Bernard Chazelle (Princeton University)
Scalable Data Mining The Auton Lab, Carnegie Mellon University Brigham Anderson, Andrew Moore, Dan Pelleg, Alex Gray, Bob Nichols, Andy.
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
3D Geometry for Computer Graphics
Lecture 7 Hyper-planes, Matrices, and Linear Systems Shang-Hua Teng.
Dimensionality Reduction
Approximate Nearest Subspace Search with applications to pattern recognition Ronen Basri Tal Hassner Lihi Zelnik-Manor Weizmann Institute Caltech.
Stockman MSU/CSE Math models 3D to 2D Affine transformations in 3D; Projections 3D to 2D; Derivation of camera matrix form.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
1cs542g-term Notes  Extra class next week (Oct 12, not this Friday)  To submit your assignment: me the URL of a page containing (links to)
FLANN Fast Library for Approximate Nearest Neighbors
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Length and Dot Product in R n Notes: is called a unit vector. Notes: The length of a vector is also called its norm. Chapter 5 Inner Product Spaces.
SVD(Singular Value Decomposition) and Its Applications
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Module 04: Algorithms Topic 07: Instance-Based Learning
Summarized by Soo-Jin Kim
Linear Least Squares Approximation. 2 Definition (point set case) Given a point set x 1, x 2, …, x n  R d, linear least squares fitting amounts to find.
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
 Karthik Gurumoorthy  Ajit Rajwade  Arunava Banerjee  Anand Rangarajan Department of CISE University of Florida 1.
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
AN ORTHOGONAL PROJECTION
Scale-less Dense Correspondences Tal Hassner The Open University of Israel ICCV’13 Tutorial on Dense Image Correspondences for Computer Vision.
Pattern Recognition April 19, 2007 Suggested Reading: Horn Chapter 14.
CSE 185 Introduction to Computer Vision Face Recognition.
Optimal Component Analysis Optimal Linear Representations of Images for Object Recognition X. Liu, A. Srivastava, and Kyle Gallivan, “Optimal linear representations.
An Introduction to Support Vector Machine (SVM)
Geometric Problems in High Dimensions: Sketching Piotr Indyk.
Approximate Nearest Neighbors: Towards Removing the Curse of Dimensionality Piotr Indyk, Rajeev Motwani The 30 th annual ACM symposium on theory of computing.
Distinctive Image Features from Scale-Invariant Keypoints
Nearest-Neighbor Searching Under Uncertainty Wuzhou Zhang Joint work with Pankaj K. Agarwal, Alon Efrat, and Swaminathan Sankararaman. To appear in PODS.
Unsupervised Learning II Feature Extraction
CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning.
Support Vector Machines (SVM)
Robust and Fast Collaborative Tracking with Two Stage Sparse Optimization Authors: Baiyang Liu, Lin Yang, Junzhou Huang, Peter Meer, Leiguang Gong and.
Lecture 8:Eigenfaces and Shared Features
Lecture: Face Recognition and Feature Reduction
Recognition: Face Recognition
LSI, SVD and Data Management
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Singular Value Decomposition
Object Modeling with Layers
Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction
Feature space tansformation methods
Maths for Signals and Systems Linear Algebra in Engineering Lecture 6, Friday 21st October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL.
Lecture 15: Least Square Regression Metric Embeddings
NonLinear Dimensionality Reduction or Unfolding Manifolds
Ronen Basri Tal Hassner Lihi Zelnik-Manor Weizmann Institute Caltech
Presentation transcript:

Approximate Nearest Subspace Search with Applications to Pattern Recognition Ronen Basri, Tal Hassner, Lihi Zelnik-Manor presented by Andrew Guillory and Ian Simon

The Problem Given n linear subspaces S i :

The Problem Given n linear subspaces S i : And a query point q:

The Problem Given n linear subspaces S i : And a query point q: Find the subspace S i that minimizes dist(S i,q).

Why? object appearance variation = subspace – fast queries on object database

Why? object appearance variation = subspace – fast queries on object database Other reasons?

Approach Solve by reduction to nearest neighbor. – point-to-point distances

Approach Solve by reduction to nearest neighbor. – point-to-point distances not actual reduction

Approach Solve by reduction to nearest neighbor. – point-to-point distances In higher-dimensional space. not actual reduction

Point-Subspace Distance Use squared distance.

Point-Subspace Distance Use squared distance.

Point-Subspace Distance Use squared distance. Squared point-subspace distance can be represented as a dot product.

The Reduction Let: Remember:

The Reduction Let: Then: Remember:

The Reduction

constant over query

The Reduction ?constant over query

The Reduction Z T Z = I ?constant over query

The Reduction Z T Z = I Z is d-by-(d-k), columns orthonormal. ?constant over query

The Reduction Z T Z = I Z is d-by-(d-k), columns orthonormal. ?constant over query

The Reduction For query point q:

The Reduction For query point q: Can we decrease the additive constant?

Observation 1 All data points lie on a hyperplane.

Observation 1 All data points lie on a hyperplane. Let: Now the hyperplane contains the origin.

Observation 2 After hyperplane projection: All data points lie on a hypersphere.

Observation 2 After hyperplane projection: All data points lie on a hypersphere. Let: Now the query point lies on the hypersphere.

Observation 2 After hyperplane projection: All data points lie on a hypersphere. Let: Now the query point lies on the hypersphere.

Reduction Geometry What is happening?

Reduction Geometry What is happening?

Finally Additive constant depends only on dimension of points and subspaces. This applies to linear subspaces, all of the same dimension.

Extensions subspaces of different dimension – lines and planes, e.g. – Not all data points have the same norm. Add extra dimension to fix this.

Extensions subspaces of different dimension – lines and planes, e.g. – Not all data points have the same norm. Add extra dimension to fix this. affine subspaces – Again, not all data points have the same norm.

Approximate Nearest Neighbor Search Find point x with distance d(x, q) <= (1 + ε) min i d(x i,q) Tree based approaches: KD-trees, metric / ball trees, cover trees Locality sensitive hashing This paper uses multiple KD-Trees with (different) random projections

KD-Trees Decompose space into axis aligned rectangles Image from Dan Pelleg

Random Projections Multiply data with a random matrix X with X(i,j) drawn from N(0,1) Several different justifications – Johnson-Lindenstrauss (data set that is small compared to dimensionality) – Compressed Sensing (data set that is sparse in some linear basis) – RP-Trees (data set that has small doubling dimension)

Results Two goals – show their method is fast – show nearest subspace is useful Four experiments – Synthetic Experiments – Image Approximation – Yale Faces – Yale Patches

Image Reconstruction

Yale Faces

Questions / Issues Should random projections be applied before or after the reduction? Why does the effective distance error go down with the ambient dimensionality? The reduction tends to make query points far away from the points in the database. Are there better approximate nearest neighbor algorithms in this case?