Large Scale Discovery of Spatially Related Images Ondřej Chum and Jiří Matas Center for Machine Perception Czech Technical University Prague.

Slides:



Advertisements
Similar presentations
Recognising Panoramas M. Brown and D. Lowe, University of British Columbia.
Advertisements

Image Retrieval with Geometry-Preserving Visual Phrases
Context-based object-class recognition and retrieval by generalized correlograms by J. Amores, N. Sebe and P. Radeva Discussion led by Qi An Duke University.
Location Recognition Given: A query image A database of images with known locations Two types of approaches: Direct matching: directly match image features.
Aggregating local image descriptors into compact codes
Three things everyone should know to improve object retrieval
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Content-Based Image Retrieval
Instance-level recognition II.
VisualRank: Applying PageRank to Large-Scale Image Search Yushi Jing, Member, IEEE, and Shumeet Baluja, Member, IEEE.
Mixture of trees model: Face Detection, Pose Estimation and Landmark Localization Presenter: Zhang Li.
CS4670 / 5670: Computer Vision Bag-of-words models Noah Snavely Object
Discrete-Continuous Optimization for Large-scale Structure from Motion David Crandall, Andrew Owens, Noah Snavely, Dan Huttenlocher Presented by: Rahul.
Special Topic on Image Retrieval Local Feature Matching Verification.
Image alignment Image from
Bag-of-features models Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Multimedia Indexing and Retrieval Kowshik Shashank Project Advisor: Dr. C.V. Jawahar.
CVPR 2008 James Philbin Ondˇrej Chum Michael Isard Josef Sivic
Packing bag-of-features ICCV 2009 Herv´e J´egou Matthijs Douze Cordelia Schmid INRIA.
Bundling Features for Large Scale Partial-Duplicate Web Image Search Zhong Wu ∗, Qifa Ke, Michael Isard, and Jian Sun CVPR 2009.
Bag of Features Approach: recent work, using geometric information.
Effective Image Database Search via Dimensionality Reduction Anders Bjorholm Dahl and Henrik Aanæs IEEE Computer Society Conference on Computer Vision.
Robust and large-scale alignment Image from
WISE: Large Scale Content-Based Web Image Search Michael Isard Joint with: Qifa Ke, Jian Sun, Zhong Wu Microsoft Research Silicon Valley 1.
Small Codes and Large Image Databases for Recognition CVPR 2008 Antonio Torralba, MIT Rob Fergus, NYU Yair Weiss, Hebrew University.
Object retrieval with large vocabularies and fast spatial matching
Lecture 28: Bag-of-words models
Recognising Panoramas
Automatic Panoramic Image Stitching using Local Features Matthew Brown and David Lowe, University of British Columbia.
Bag-of-features models
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Multiple Object Class Detection with a Generative Model K. Mikolajczyk, B. Leibe and B. Schiele Carolina Galleguillos.
Keypoint-based Recognition and Object Search
10/31/13 Object Recognition and Augmented Reality Computational Photography Derek Hoiem, University of Illinois Dali, Swans Reflecting Elephants.
Object Recognition and Augmented Reality
Indexing Techniques Mei-Chen Yeh.
Keypoint-based Recognition Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/04/10.
Near Duplicate Image Detection: min-Hash and tf-idf weighting
A General Framework for Tracking Multiple People from a Moving Camera
Learning a Fast Emulator of a Binary Decision Process Center for Machine Perception Czech Technical University, Prague ACCV 2007, Tokyo, Japan Jan Šochman.
A Statistical Approach to Speed Up Ranking/Re-Ranking Hong-Ming Chen Advisor: Professor Shih-Fu Chang.
1 Total Recall: Automatic Query Expansion with a Generative Feature Model for Object Retrieval Ondrej Chum, James Philbin, Josef Sivic, Michael Isard and.
Example: line fitting. n=2 Model fitting Measure distances.
Fast Similarity Search for Learned Metrics Prateek Jain, Brian Kulis, and Kristen Grauman Department of Computer Sciences University of Texas at Austin.
Collective Vision: Using Extremely Large Photograph Collections Mark Lenz CameraNet Seminar University of Wisconsin – Madison February 2, 2010 Acknowledgments:
10/31/13 Object Recognition and Augmented Reality Computational Photography Derek Hoiem, University of Illinois Dali, Swans Reflecting Elephants.
18 th August 2006 International Conference on Pattern Recognition 2006 Epipolar Geometry from Two Correspondences Michal Perďoch, Jiří Matas, Ondřej Chum.
CVPR 2006 New York City Spatial Random Partition for Common Visual Pattern Discovery Junsong Yuan and Ying Wu EECS Dept. Northwestern Univ.
Peter Henry1, Michael Krainin1, Evan Herbst1,
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Geometric Transformations
Total Recall: Automatic Query Expansion with a Generative Feature Model for Object Retrieval O. Chum, et al. Presented by Brandon Smith Computer Vision.
Discovering Objects and their Location in Images Josef Sivic 1, Bryan C. Russell 2, Alexei A. Efros 3, Andrew Zisserman 1 and William T. Freeman 2 Goal:
Bundling Features for Large Scale Partial-Duplicate Web Image Search Zhong Wu ∗, Qifa Ke, Michael Isard, and Jian Sun Microsoft Research.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
Invariant Local Features Image content is transformed into local feature coordinates that are invariant to translation, rotation, scale, and other imaging.
IIIT HYDERABAD Techniques for Organization and Visualization of Community Photo Collections Kumar Srijan Faculty Advisor : Dr. C.V. Jawahar.
776 Computer Vision Jan-Michael Frahm Spring 2012.
25 Years of RANSAC 25 Years of RANSAC Ondřej Chum
The topic discovery models
Capturing, Processing and Experiencing Indian Monuments
Nonparametric Semantic Segmentation
Approximate Models for Fast and Accurate Epipolar Geometry Estimation
Jiří Matas and Onřej Chum Centre for Machine Perception
The topic discovery models
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
The topic discovery models
Minwise Hashing and Efficient Search
Presentation transcript:

Large Scale Discovery of Spatially Related Images Ondřej Chum and Jiří Matas Center for Machine Perception Czech Technical University Prague

2 /26 Related Vision Problems Organize my holiday snapshots –Schaffalitzky and Zisserman ECCV’02 Find images containing a given “object” (“window”) –Sivic ICCV‘03, Nister CVPR‘06, Jegou CVPR’07, Philbin CVPR‘07, Chum ICCV’07 Find small “object” in a film –Sivic and Zisserman CVPR’04 Match and reconstruct Saint Marco –Snavely, Seitz and Szeliski SIGGRAPH’06 Find and match ALL spatially related images in a large database, using only visual information, i.e. not using (flicker) tags, EXIF info, GPS, …. This Work O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

3 /26 Visual Only Approach Large database ( images in our experiments) Find spatially related clusters Fast method, even for sizes up to 2 50 images Probability of successful discovery of spatial relation of images independent of database size O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

4 /26 Image Clustering and its Time Complexity Standard Approach (using image retrieval): Quadratic method in the size of database D -- O(D 2 ) the multiplicative constant at the quadratic term ~ 1 – quadratic even for small D 1.Take each image in turn 2.Use a image retrieval system to retrieve related images 3.Compute connected components of the graph Proposed method 1.Seed Generation – hashing characterize images by pseudo-random numbers stored in a hash table time complexity equal to the sum of variances of Poisson distributions linear for database size D ¼ Seed Growing – retrieval complete the clusters only for cluster members c << D, complexity O(cD) O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

5 /26 Building on Two Methods Fast (low recall) seed generation based on hashing Thorough (high recall) seed growing based on image retrieval Chum, Philbin, Isard, and Zisserman: Scalable Near Identical Image and Shot Detection CIVR 2007 Chum, Philbin, Sivic, Isard, and Zisserman: Total Recall: Automatic Query Expansion with a Generative Feature Model for Object Retrieval ICCV 2007 O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

6 /26 Image Representation Feature detector SIFT descriptor [Lowe’04] Visual vocabulary Vector quantization … Bag of words Set of words O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

7 /26 Hypothesizing Seeds with min-Hash A 1 ∩ A 2 A 1 U A 2 A1A1 A2A2 Image similarity measured as a set overlap (using min-Hash algorithm) Spatially related images share visual words Problem: Robustly estimate set overlap of high dimensional sparse binary vectors in constant time independent of the dimensionality (d ¼ 10 5 ) Set overlap probabilistically estimated via min-Hash Similar approach as LSH (locally sensitive hashing) O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

8 /26 min-Hash According to some (replicable) key select a small number of non-zero elements Similar vectors should have similar selected elements Key = generate a random number (a hash) for each dimension, choose nonzero element with minimal value of the key

9 /26 Seed Generation: Probability of Success An image pair forms a seed if at least one of k s-tuples of min-Hashes agrees. Probability that an image pair is retrieved is a function of the similarity: where s,k are user-controllable parameters of the method: s governs the size of the hashing table k is number of hashing tables Successfully retrieved pair of images = at least one collision in one of the tables (equivalent to AND-OR) O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

10 /26 Probability of Retrieving an Image Pair similarity (set overlap) Near duplicate Images Images of the same object and unrelated images 8.9 % (sim = 0.057) 5.1% (sim = 0.047) 13.9 % (sim = 0,066) 100% (sim = 0.746) 100% (sim = 0.322) 99.5% (sim = 0,217) probability of retrieval O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

11 /26 Spatially Related Images 18.9 % (sim = 0,074)5.1 % (sim = 0,047) similarity (set overlap) probability of retrieval (log scale) 13.9 % 8.9 % 5.1 % 9.8 % 7.2 % 8.9 % 13.9 % 16.3 % 10.7 %

12 /26 10% 7% 4% 5% 4% Seed Generation P (no seed) = 6% % %68.88 % O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

13 / % Seed Generation P (no seed) = %31.84 %1.94 % Resemblance to RANSAC Related image pair ~ an all inlier sample (there is no need to enumerate them all, one hit is sufficient) Probability of retrieving an image pair ~ fraction of inliers The number of related image pairs ~ how many times we can try O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

14 /26 At Least One Seed in Cluster cluster size P(no seed ) similarity = probability of retrieval 6.2% 10.4% 16.1% Estimate of the probability of failure plot against the size of the cluster assumption used in this plot: all images in the cluster are related O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

15 /26 backproject features Growing the Seed Application of Total Recall –Combining average query expansion and transitive closure –3D geometric constraint (not only affine transformation) –Tighter geometric constraints (10 pixel threshold) query enhanced query Average query expansion (from possibly multiple coplanar structures) Transitive closure crawl O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

16 /26 Summary of the Method Unknown structure min-Hash seeds x Spatial verification Query Expansion Rejected seed Missed cluster Seed Cluster skeleton Failed retrieval Images O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

17 /26 Experiment 1 Univ. of Kentucky Dataset [Nister & Stewenius] 2550 clusters of size 4 – very small clusters “partial” ground truth: “different” cluster share the same background How many clusters have at least one seed? CONTRAST – DIFFERENT TASK If we were looking for ALL results not ANY (seed) the standard retrieval measure on this dataset would be only 1.63 out of % O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

18 /26 Experimental Validation UKY dataset cluster size P(no seed) similarity = probability of retrieval 6.2% 10.4% 16.1% + In University of Kentucky dataset “average” similarity slightly above 0.06 O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

19 /26 Experimental Results on 100k Images Hertford Keble Magdalen Pitt Rivers Radcliffe Camera All Soul's Ashmolean Balliol Bodleian Christ Church Cornmarket Images downloaded from FLICKR Includes 11 Oxford Landmarks with manually labelled ground truth O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

20 /26 Experimental Results on 100k Images Settings scalable to millions images, also finding small clusters Settings scalable to billions images, only finding larger clusters Timing: 17 min 13 sec + 16 min 20 sec = sec / image O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

21 /26 Application – Object Labelling Factorizing the clusters using multiple constrains Matches between images Weak geometric constraints (coplanarity, disparity) Photographer’s psychology – tends to take pictures of single objects

22 /26

23 /26

24 /26 Automatic 3D Reconstruction O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

25 /26 O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

26 /26 Conclusions Novel method for fast clustering in large collections Combines fast low recall method (seed generation) and thorough (total recall) method for seed growing Probability of finding a cluster rapidly increases with its size and is independent of the size of the database Can be incrementally updated as the database grows Efficient: sec / image on a single PC Fully parallelizable A state of the art near duplicate detection comes as a bonus (as a part of seed generation) O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images

27 /26 Thank you! Thanks to Daniel Martinec, Michal Perďoch, James Philbin, Jakub Pokluda Technical Report available O.Chum, J. Matas: Large Scale Discovery of Spatially Related Images