Fast Similarity Search for Learned Metrics Prateek Jain, Brian Kulis, and Kristen Grauman Department of Computer Sciences University of Texas at Austin.

Slides:



Advertisements
Similar presentations
Registration for Robotics Kurt Konolige Willow Garage Stanford University Patrick Mihelich JD Chen James Bowman Helen Oleynikova Freiburg TORO group: Giorgio.
Advertisements

Algorithmic High-Dimensional Geometry 1 Alex Andoni (Microsoft Research SVC)
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Searching on Multi-Dimensional Data
MIT CSAIL Vision interfaces Towards efficient matching with random hashing methods… Kristen Grauman Gregory Shakhnarovich Trevor Darrell.
Efficiently searching for similar images (Kristen Grauman)
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
MIT CSAIL Vision interfaces Approximate Correspondences in High Dimensions Kristen Grauman* Trevor Darrell MIT CSAIL (*) UT Austin…
The Pyramid Match Kernel: Discriminative Classification with Sets of Image Features Kristen Grauman Trevor Darrell MIT.
Similarity Search in High Dimensions via Hashing
Low Complexity Keypoint Recognition and Pose Estimation Vincent Lepetit.
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
VLSH: Voronoi-based Locality Sensitive Hashing Sung-eui Yoon Authors: Lin Loi, Jae-Pil Heo, Junghwan Lee, and Sung-Eui Yoon KAIST
Computer Vision Group, University of BonnVision Laboratory, Stanford University Abstract This paper empirically compares nine image dissimilarity measures.
Introduction to Data-driven Animation Jinxiang Chai Computer Science and Engineering Texas A&M University.
Coherency Sensitive Hashing (CSH) Simon Korman and Shai Avidan Dept. of Electrical Engineering Tel Aviv University ICCV2011 | 13th International Conference.
1 Large Scale Similarity Learning and Indexing Part II: Learning to Hash for Large Scale Search Fei Wang and Jun Wang IBM TJ Watson Research Center.
Fast and Compact Retrieval Methods in Computer Vision Part II A. Torralba, R. Fergus and Y. Weiss. Small Codes and Large Image Databases for Recognition.
1 Jun Wang, 2 Sanjiv Kumar, and 1 Shih-Fu Chang 1 Columbia University, New York, USA 2 Google Research, New York, USA Sequential Projection Learning for.
1998/5/21by Chang I-Ning1 ImageRover: A Content-Based Image Browser for the World Wide Web Introduction Approach Image Collection Subsystem Image Query.
Computing Sketches of Matrices Efficiently & (Privacy Preserving) Data Mining Petros Drineas Rensselaer Polytechnic Institute (joint.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Nearest Neighbor Retrieval Using Distance-Based Hashing Michalis Potamias and Panagiotis Papapetrou supervised by Prof George Kollios A method is proposed.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Multiple Object Class Detection with a Generative Model K. Mikolajczyk, B. Leibe and B. Schiele Carolina Galleguillos.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Object Recognition by Discriminative Methods Sinisa Todorovic 1st Sino-USA Summer School in VLPR July, 2009.
Agenda Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based.
Large Scale Recognition and Retrieval. What does the world look like? High level image statistics Object Recognition for large-scale search Focus on scaling.
Efficient Image Search and Retrieval using Compact Binary Codes
Indexing Techniques Mei-Chen Yeh.
Approximation algorithms for large-scale kernel methods Taher Dameh School of Computing Science Simon Fraser University March 29 th, 2010.
Image Based Positioning System Ankit Gupta Rahul Garg Ryan Kaminsky.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Beyond Locality Sensitive Hashing Alex Andoni (Microsoft Research) Joint with: Piotr Indyk (MIT), Huy L. Nguyen (Princeton), Ilya Razenshteyn (MIT)
Nearest Neighbor Paul Hsiung March 16, Quick Review of NN Set of points P Query point q Distance metric d Find p in P such that d(p,q) < d(p’,q)
Example: line fitting. n=2 Model fitting Measure distances.
Minimal Loss Hashing for Compact Binary Codes
Mingyang Zhu, Huaijiang Sun, Zhigang Deng Quaternion Space Sparse Decomposition for Motion Compression and Retrieval SCA 2012.
Geodesic Flow Kernel for Unsupervised Domain Adaptation Boqing Gong University of Southern California Joint work with Yuan Shi, Fei Sha, and Kristen Grauman.
Similarity Searching in High Dimensions via Hashing Paper by: Aristides Gionis, Poitr Indyk, Rajeev Motwani.
An Approximate Nearest Neighbor Retrieval Scheme for Computationally Intensive Distance Measures Pratyush Bhatt MS by Research(CVIT)
Approximate Nearest Neighbors: Towards Removing the Curse of Dimensionality Piotr Indyk, Rajeev Motwani The 30 th annual ACM symposium on theory of computing.
KNN & Naïve Bayes Hongning Wang Today’s lecture Instance-based classifiers – k nearest neighbors – Non-parametric learning algorithm Model-based.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
3D Motion Data Mining Multimedia Project Multimedia and Network Lab, Department of Computer Science.
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
776 Computer Vision Jan-Michael Frahm Spring 2012.
Optimal Data-Dependent Hashing for Nearest Neighbor Search Alex Andoni (Columbia University) Joint work with: Ilya Razenshteyn.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Similarity Measurement and Detection of Video Sequences Chu-Hong HOI Supervisor: Prof. Michael R. LYU Marker: Prof. Yiu Sang MOON 25 April, 2003 Dept.
Summer School on Hashing’14 Dimension Reduction Alex Andoni (Microsoft Research)
IIIT Hyderabad Diversity in Image Retrieval: Randomization and Learned Metrics P Vidyadhar Rao MS by Research CVIT, IIIT Hyderabad
Cross-modal Hashing Through Ranking Subspace Learning
KNN & Naïve Bayes Hongning Wang
Invariant Local Features Image content is transformed into local feature coordinates that are invariant to translation, rotation, scale, and other imaging.
Semi-Supervised Clustering
Information Retrieval
Metric Learning for Clustering
Sublinear Algorithmic Tools 3
Lecture 11: Nearest Neighbor Search
Approximate Correspondences in High Dimensions
K Nearest Neighbor Classification
Data-Dependent Hashing for Nearest Neighbor Search
Locality Sensitive Hashing
Scale-Space Representation for Matching of 3D Models
Minwise Hashing and Efficient Search
President’s Day Lecture: Advanced Nearest Neighbor Search
Topological Signatures For Fast Mobility Analysis
Ronen Basri Tal Hassner Lihi Zelnik-Manor Weizmann Institute Caltech
Presentation transcript:

Fast Similarity Search for Learned Metrics Prateek Jain, Brian Kulis, and Kristen Grauman Department of Computer Sciences University of Texas at Austin

Motivation Fast image search is a useful component for a number of vision problems. ? Object categorization

Motivation Example-based pose estimation ? Fast image search is a useful component for a number of vision problems.

Motivation Structure from Motion ? Fast image search is a useful component for a number of vision problems.

Problem Search must be both fast and accurate –“Generic” distances or low-dimensional representations amenable to fast search, but may be inaccurate for a given problem. –Learned task-specific distance functions more accurate, but current methods cannot guarantee fast search for them. Our approach: –Develop approximate similarity search method for learned metrics –Encode side-information into randomized locality- sensitive hash functions –Applicable for a variety of image search tasks

Metric learning for image distances –Weinberger et al. 2004, Hertz et al. 2004, Frome et al. 2007, Varma & Ray 2007 Embedding functions to reduce cost of expensive distances –Athitsos et al. 2004, Grauman & Darrell 2005, Torralba et al Search structures based on spatial partitioning and recursive decompositions –Beis & Lowe 1997, Obdrzalek & Matas 2005, Nister & Stewenius 2006, Uhlmann 1991 Related work Locality-sensitive hashing (LSH) for vision applications –Shakhnarovich et al. 2003, Frome et al. 2004, Grauman & Darrell 2004 Data-dependent variants of LSH –Shakhnarovich et al. 2003, Georgescu et al. 2003

Metric learning There are various ways to judge appearance/shape similarity… but often we know more about (some) data than just their appearance.

Metric learning Exploit partially labeled data and/or (dis)similarity constraints to construct more useful distance function Various existing techniques

Problem-specific knowledge Example sources of similarity constraints Detected video shots, tracked objects User feedback Partially labeled image databases Fully labeled image databases

Problem: How to guarantee fast search for a learned metric? Exact search methods break down in high-d spaces, rely on good partitioning heuristics, and can degenerate to linear scan in worst case. Approximate search techniques are defined only for particular “generic” metrics, e.g. Hamming distance, L p norms, inner product.

Mahalanobis distances Distance parameterized by p.d. d × d matrix A: Similarity measure is associated generalized inner product (kernel)

Information-theoretic (LogDet) metric learning [Davis, Kulis, Jain, Sra, and Dhillon, ICML 2007] Formulation: Advantages: -Simple, efficient algorithm -Can be applied in kernel space

Locality Sensitive Hashing (LSH) Q h r 1 …r k XiXi N h << N Q [Indyk and Motwani 1998, Charikar 2002] Guarantee “approximate”- nearest neighbors ((1+ε)- accurate) in sub-linear time, given appropriate hash functions.

LSH functions for dot products The probability that a random hyperplane separates two unit vectors depends on the angle between them: [Goemans and Williamson 1995, Charikar 2004] High dot product: unlikely to split Lower dot product: likely to split Corresponding hash function:

LSH functions for learned metrics It should be unlikely that a hash function will split examples like those having similarity constraints… …but likely that it splits those having dissimilarity constraints.

LSH functions for learned metrics Given learned metric with We generate parameterized hash functions for : This satisfies the locality-sensitivity condition:

Implicit hashing formulation Image data often high-dimensional—must work in kernel space High-d inputs are sparse, but may be dense can’t work with. We derive an implicit update rule that simultaneously updates metric and hash function parameters. Integrates metric learning and hashing

Implicit hashing formulation S is c x c matrix of coefficients that determine how much weight each pair of the c constrained inputs contributes to learned parameters. We show that the same hash function can be computed indirectly via: Possible due to property of information-theoretic metric learning

Recap: data flow 1.Receive constraints and base metric. 2.Learning stage: simultaneously update metric and hash functions. 3.Hash database examples into table. 4.When a query arrives, hash into existing table for approximate neighbors under learned metric.

Results Object Categorization Pose Estimation Patch Indexing Caltech 101, O(10 6 ) dimensions, 4k points Poser data, 24k dimensions,.5 million points Photo Tourism data, 4096 dimensions, 300k points

Results: object categorization Best accuracy to date with a single metric / kernel. Caltech-101 database [CORR] [PMK] ML = metric learning

Results: object categorization Query time controlled by required accuracy e.g., search less than 2% of database examples for accuracy close to linear scan k-NN error rate (101 classes) Epsilon (ε) slower search faster search

Results: object categorization Query time controlled by required accuracy e.g., search less than 2% of database examples for accuracy close to linear scan Epsilon (ε) slower search faster search k-NN error rate (101 classes)

Results: pose estimation 500,000 synthetic images Measure mean error per joint between query and NN –Random 2 database images: 34.5 cm between each joint Average query time: –ML linear scan: sec –ML hashing: 1.39 sec Error (cm)

Results: patch indexing O(10 5 ) patches Photo Tourism data: goal is to match patches that correspond to same point on 3d object More accurate matches → better reconstruction Huge search pool [Photo Tourism data provided by Snavely, Seitz, Szeliski, Winder & Brown]

Recall Number of patches retrieved Learned metric improves recall Search 100% of data Search 0.8% of data Our technique maintains accuracy while searching less than 1% of the database. Results: patch indexing Photo Tourism data

Summary Content-based queries demand fast search algorithms for useful image metrics. Contributions: –Semi-supervised hash functions for class of learned metrics and kernels –Theoretical guarantees of accuracy on nearest neighbor searches –Validation with pose estimation, object categorization, and patch indexing tasks.