© 2009 IBM Corporation IBM Research Xianglong Liu 1, Yadong Mu 2, Bo Lang 1 and Shih-Fu Chang 2 1 Beihang University, Beijing, China 2 Columbia University,

Slides:



Advertisements
Similar presentations
© 2009 IBM Corporation IBM Research Xianglong Liu 1, Junfeng He 2,3, and Bo Lang 1 1 Beihang University, Beijing, China 2 Columbia University, New York,
Advertisements

Weakly supervised learning of MRF models for image region labeling Jakob Verbeek LEAR team, INRIA Rhône-Alpes.
Location Recognition Given: A query image A database of images with known locations Two types of approaches: Direct matching: directly match image features.
Nearest Neighbor Search in High Dimensions Seminar in Algorithms and Geometry Mica Arie-Nachimson and Daniel Glasner April 2009.
Data Mining Feature Selection. Data reduction: Obtain a reduced representation of the data set that is much smaller in volume but yet produces the same.
Presented by Xinyu Chang
Active Learning for Streaming Networked Data Zhilin Yang, Jie Tang, Yutao Zhang Computer Science Department, Tsinghua University.
VisualRank: Applying PageRank to Large-Scale Image Search Yushi Jing, Member, IEEE, and Shumeet Baluja, Member, IEEE.
MIT CSAIL Vision interfaces Towards efficient matching with random hashing methods… Kristen Grauman Gregory Shakhnarovich Trevor Darrell.
Query Specific Fusion for Image Retrieval
Low Complexity Keypoint Recognition and Pose Estimation Vincent Lepetit.
Exploring Reduction for Long Web Queries Niranjan Balasubramanian, Giridhar Kuamaran, Vitor R. Carvalho Speaker: Razvan Belet 1.
Presented by Arshad Jamal, Rajesh Dhania, Vinkal Vishnoi Active hashing and its application to image and text retrieval Yi Zhen, Dit-Yan Yeung, Published.
Stephan Gammeter, Lukas Bossard, Till Quack, Luc Van Gool.
Multimedia Indexing and Retrieval Kowshik Shashank Project Advisor: Dr. C.V. Jawahar.
Presented by Relja Arandjelović Iterative Quantization: A Procrustean Approach to Learning Binary Codes University of Oxford 21 st September 2011 Yunchao.
Small Codes and Large Image Databases for Recognition CVPR 2008 Antonio Torralba, MIT Rob Fergus, NYU Yair Weiss, Hebrew University.
1 Large Scale Similarity Learning and Indexing Part II: Learning to Hash for Large Scale Search Fei Wang and Jun Wang IBM TJ Watson Research Center.
Fast and Compact Retrieval Methods in Computer Vision Part II A. Torralba, R. Fergus and Y. Weiss. Small Codes and Large Image Databases for Recognition.
ACM Multimedia th Annual Conference, October , 2004
1 Jun Wang, 2 Sanjiv Kumar, and 1 Shih-Fu Chang 1 Columbia University, New York, USA 2 Google Research, New York, USA Sequential Projection Learning for.
Sparse Solutions for Large Scale Kernel Machines Taher Dameh CMPT820-Multimedia Systems Dec 2 nd, 2010.
Ensemble Learning: An Introduction
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
Similarity Search in High Dimensions via Hashing Aristides Gionis, Protr Indyk and Rajeev Motwani Department of Computer Science Stanford University presented.
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
Nearest Neighbor Retrieval Using Distance-Based Hashing Michalis Potamias and Panagiotis Papapetrou supervised by Prof George Kollios A method is proposed.
Online Learning Algorithms
Joint Image Clustering and Labeling by Matrix Factorization
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
Search Result Diversification by M. Drosou and E. Pitoura Presenter: Bilge Koroglu June 14, 2011.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
Data mining and machine learning A brief introduction.
Mehdi Kargar Aijun An York University, Toronto, Canada Keyword Search in Graphs: Finding r-cliques.
Recent Advances of Compact Hashing for Large-Scale Visual Search Shih-Fu Chang Columbia University October 2012 Joint work with Junfeng He (Facebook),
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
Skewing: An Efficient Alternative to Lookahead for Decision Tree Induction David PageSoumya Ray Department of Biostatistics and Medical Informatics Department.
Fast Similarity Search for Learned Metrics Prateek Jain, Brian Kulis, and Kristen Grauman Department of Computer Sciences University of Texas at Austin.
Machine Learning in Ad-hoc IR. Machine Learning for ad hoc IR We’ve looked at methods for ranking documents in IR using factors like –Cosine similarity,
Minimal Loss Hashing for Compact Binary Codes
IEEE Int'l Symposium on Signal Processing and its Applications 1 An Unsupervised Learning Approach to Content-Based Image Retrieval Yixin Chen & James.
INTRODUCTION Heesoo Myeong and Kyoung Mu Lee Department of ECE, ASRI, Seoul National University, Seoul, Korea Tensor-based High-order.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
Zibin Zheng DR 2 : Dynamic Request Routing for Tolerating Latency Variability in Cloud Applications CLOUD 2013 Jieming Zhu, Zibin.
INTERACTIVELY BROWSING LARGE IMAGE DATABASES Ronald Richter, Mathias Eitz and Marc Alexa.
1 A Compact Feature Representation and Image Indexing in Content- Based Image Retrieval A presentation by Gita Das PhD Candidate 29 Nov 2005 Supervisor:
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Learning Spectral Clustering, With Application to Speech Separation F. R. Bach and M. I. Jordan, JMLR 2006.
Project by: Cirill Aizenberg, Dima Altshuler Supervisor: Erez Berkovich.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Automatic Video Tagging using Content Redundancy Stefan Siersdorfer 1, Jose San Pedro 2, Mark Sanderson 2 1 L3S Research Center, Germany 2 University of.
Answering Top-k Queries Using Views Gautam Das (Univ. of Texas), Dimitrios Gunopulos (Univ. of California Riverside), Nick Koudas (Univ. of Toronto), Dimitris.
Approximate Nearest Neighbors: Towards Removing the Curse of Dimensionality Piotr Indyk, Rajeev Motwani The 30 th annual ACM symposium on theory of computing.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
KNN & Naïve Bayes Hongning Wang Today’s lecture Instance-based classifiers – k nearest neighbors – Non-parametric learning algorithm Model-based.
Unsupervised Auxiliary Visual Words Discovery for Large-Scale Image Object Retrieval Yin-Hsi Kuo1,2, Hsuan-Tien Lin 1, Wen-Huang Cheng 2, Yi-Hsuan Yang.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Cross-modal Hashing Through Ranking Subspace Learning
KNN & Naïve Bayes Hongning Wang
Presenter: Chu-Song Chen
Probabilistic Data Management
Locality Sensitive Hashing
Overfitting and Underfitting
Minwise Hashing and Efficient Search
Presentation transcript:

© 2009 IBM Corporation IBM Research Xianglong Liu 1, Yadong Mu 2, Bo Lang 1 and Shih-Fu Chang 2 1 Beihang University, Beijing, China 2 Columbia University, New York, USA Compact Hashing for Mixed Image-Keyword Query over Multi-Label Images

 Introduction – Motivation – Our Solution  Boosted Shared Hashing – Formulation – Optimization – The Retrieval Stage  Experiments  Conclusion Outline

 Yet another image search paradigm – Query image provides content descriptor “Image + Keyword” based Visual Search (1/4) 3 Query Image Search-By-Example Results on Page #1 – Textual keywords greatly narrow the semantic gap!

 Challenge-1: Noisy or unknown label information – Database Images: labels are unknown and expensive to annotate – Training Images: a small set, and manually annotated – Query: Image + Keyword (or label) “Image + Keyword” based Visual Search (2/4) 4 Visual FeaturesLabels Database Training Set Query √ √√ √√ Problem Settings

 Challenge-2: Scalability to Web-Scale Data – Linear scan is infeasible – Approximate nearest neighbor (ANN) balance the performance and computational complexity Tree-based methods (KD tree, metric tree, ball tree, etc. ) Hashing-based methods: efficient index and search “Image + Keyword” based Visual Search (3/4) 5 … … wkwk 1 0/-1 HashingHashing Tables Bucket Indexed Image

6  Challenge-3: Diverse Semantics User intention is ambiguous / diverse “Image + Keyword” based Visual Search (4/4) Keywords Cat Dog+Cat Query Image Top Results Query-adaptive hashing over multi-label data

 Supervised Hashing Related Works 7 Naïve solutions for Hashing with Multi-Label Data Universal hash function: Semi-Supervised Hash [Wang10a] Sequential Projection Hash [Wang10b] Universal hash function: Semi-Supervised Hash [Wang10a] Sequential Projection Hash [Wang10b] Per-Label hash function : Bit-Selection Hash [Mu 11] Per-Label hash function : Bit-Selection Hash [Mu 11] Universal ones: hard to learn Complicate bit selection Universal ones: hard to learn Complicate bit selection Redundancy

 Key idea: to encourage sparse association between hashing functions and labels by exploiting shared subspaces among the labels Overview of Our Solution (1/2) √ √ 2 2 √ 1 1 √ 4 4 √ √ Hashing functionsBit-label association Query labels Image Select optimal hash bits per query denotes samples associated with both and

Overview of Our Solution (2/2) 9 Labels Visual Features Training Data Labels Visual Features Query Boosted Shared Hashing Indexed Database Visual Features Database Data Hash Functions Bit-Label Association Matrix Selected Bits Output Ranked Images

Data Structure Multi-label data : x i is associated with the k-th label x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 … = 1= -1= 0 x l1 x l2 x l3 … … Neighbor graph homogeneous neighbors: with the same label heterogeneous neighbors: with different labels x x Homogeneous Neighbors Heterogeneous Neighbors Label x 10

Objective Function  Encourage prediction f of hashing function h on neighbor pairs to have the same sign with neighbor matrix Z : 11 Exponential Loss Function : Homogeneous: z ij =1, expect h b (x i ) h b (x j )=1 Heterogeneous: z ij =-1, expect h b (x i ) h b (x j )=-1 Hashing function z ij f(x i,x j,k)=1 Hashing prediction Active Label Set S(b): labels associated with the b -th bit

 Boosting style: to learn a hashing function that tries to correct the previous mistakes by updating weights on neighbor pairs Sequential Learning: Boosting x l1 x l2 x l3 … x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 … … x l1 x l2 x l3 … x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 … … x l1 x l2 x l3 … x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 … … > 0< 0= 0 weightsprediction error 12

= 1= -1= 0 √ √ √ √ ╳ √ x l1 x l2 x l3 x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 … ╳ √ ╳ ╳ … - Optimize: hashing function 13  Taylor expansion  Relaxation of sign function efficiently solved by eigen-decomposition

Optimize: active label set  Find a label subset S(b) that gives minimum loss – Intuitive way: exhaustively compare all possible 2 L subsets – A greedy selection O(L 2 ) : Initialize S(b) : the label giving minimum loss; Expand S(b) : add label giving the most loss decrease among all rest labels Terminated when the gain is incremental (<5%) Association matrix on CIFAR-10 14

 Bit selection based on matching-score – Select bits that are most confident across all query labels – Measured by Jaccard index: computed between a (any column of matrix A ) and query labels : Query-Adaptive Search is the learned bit-label association matrix cat dog hihi hjhj hkhk... Association matrix on CIFAR-10 15

Experiments  Datasets – Multi-category: CIFAR-10 (60K) – Multi-label: NUS-WIDE (270K)  Baselines: – SPLH [Wang 10a], SH [Weiss 08], and LSH [Indyk 98]  Setting: – 15 homogeneous and 30 heterogeneous neighbors without tuning. – same # bits per query for all methods – Average performance of 10 independent runs 16

17 CIFAR-10  32x32 color images, 10 semantic categories (e.g., airplane, frog, truck etc.)  3,000 images as training data  1,000 random samples as the queries  384-D GIST features

Impact of Sharing  Greedy sharing: S(b)  All sharing: each hashing function is universal for all labels  Random sharing: uniformly sample specific number ( the averaged size of S(b) ) of labels to be active 18

19 NUS-WIDE  Select 25 most-frequent tags (“sky”, “clouds”, “person”, etc.) from 81 tags  5,000 images as training set  1,000 images with two randomly selected labels as the query set  Groundtruth for each query: images with both (1) the same labels; and (2) the closest distances of their visual features  Concatenate 500-D Bow (SIFT) and 225-D block-wise color moment

Examples 20

 Summary and contributions – the first compact hashing technique for mixed image-keyword search over multi-label images – an efficient Boosting-style algorithm to sequentially learn the hashing functions and active label set for multi-label images – A simple hashing function selection adaptive to query labels  Future work – Theoretical analysis of performance guarantee – Extension to non-linear and reweighted hashing Summary and Conclusion 21

Thank you!

Convergency 23

Sharing Rate 24