Presentation is loading. Please wait.

Presentation is loading. Please wait.

Approximate Correspondences in High Dimensions

Similar presentations


Presentation on theme: "Approximate Correspondences in High Dimensions"— Presentation transcript:

1 Approximate Correspondences in High Dimensions
Kristen Grauman1,2 and Trevor Darrell1 1CSAIL, Massachusetts Institute of Technology 2Department of Computer Sciences, University of Texas-Austin Problem Results The correspondence between sets of local feature vectors is often a good measure of similarity, but it is computationally expensive. VG pyramids’ matching scores consistently highly correlated with the optimal matching, even for high dimensional features. (ETH-80 image data, SIFT features, k=10, L=5, results from 10 runs) flakes snow cool ice ski cold No explicit search for matches! Accuracy of existing matching approximations declines linearly with the feature dimension. The Vocabulary-Guided Pyramid Match Our approach Uniform bins Vocabulary-guided bins Data-dependent pyramid structure allows more gradual distance ranges. Form multi-resolution decomposition of the feature space to efficiently count “implicit” matches without directly comparing features Exploit structure in feature space when placing partitions in order to fully leverage their grouping power Approximate partial matching Linear-time match Mercer kernel Accurate for feature dimensions > 100 Uniformly shaped bins result in decreased matching accuracy for high-dimensional features… Tune pyramid partitions to the feature distribution Explicit correspondence fields are more accurate and faster to compute. Hierarchical k-means over corpus of features Record diameters of the irregularly shaped cells Optimal partial match Vocabulary-guided (VG) pyramid match cost: time Number of matches in bin i,j Number of new matches for jth bin at ith level Number of matches in bin i,j’s children The Pyramid Match set of features → histogram pyramid [Grauman and Darrell, ICCV 2005] In time, approximate the optimal partial matching cost: use multi-resolution histograms to count matches that are possible within a discrete set of distances. Improved object recognition when used as a kernel in an SVM. Weighting options: diameter of cell i,j Pyramid matching method Mean recognition rate/class (d=128/d=10) Time/match (s) (d=128/d=10) Vocabulary-guided bins Uniform bins 99.0 / 97.7 64.9 / 96.5 6.1e-4/6.2e-4 1.5e-3 / 5.7e-4 (Caltech-4 data set, Harris and MSER-detected SIFT features) admits a Mercer kernel input-specific upper bound Pyramid match cost: VG pyramid structure stored once in Histograms stored sparsely in entries Inserting point sets into histograms adds time Match time still only Future work Sub-linear time PM hashing (ongoing) Distortion bounds for the VG-PM? Learning weights on pyramid bins Beyond geometric vocabularies Weight according to bin size Number of new matches at level i counted by difference in histogram intersections across levels


Download ppt "Approximate Correspondences in High Dimensions"

Similar presentations


Ads by Google