Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Aggregating local image descriptors into compact codes
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Three things everyone should know to improve object retrieval
Presented by Xinyu Chang
An Introduction of Support Vector Machine
Support Vector Machines
SVM—Support Vector Machines
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Machine learning continued Image source:
Multi-layer Orthogonal Codebook for Image Classification Presented by Xia Li.
1 Part 1: Classical Image Classification Methods Kai Yu Dept. of Media Analytics NEC Laboratories America Andrew Ng Computer Science Dept. Stanford University.
Computer Vision for Human-Computer InteractionResearch Group, Universität Karlsruhe (TH) cv:hci Dr. Edgar Seemann 1 Computer Vision: Histograms of Oriented.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
CS4670 / 5670: Computer Vision Bag-of-words models Noah Snavely Object
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Addressing the Medical Image Annotation Task using visual words representation Uri Avni, Tel Aviv University, Israel Hayit GreenspanTel Aviv University,
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
Discriminative and generative methods for bags of features
Bag-of-features models Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Beyond bags of features: Part-based models Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Bayes Rule How is this rule derived? Using Bayes rule for probabilistic inference: –P(Cause | Evidence): diagnostic probability –P(Evidence | Cause): causal.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Fuzzy Support Vector Machines (FSVMs) Weijia Wang, Huanren Zhang, Vijendra Purohit, Aditi Gupta.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
1 Image Recognition - I. Global appearance patterns Slides by K. Grauman, B. Leibe.
Lecture 28: Bag-of-words models
A Study of Approaches for Object Recognition
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman ICCV 2003 Presented by: Indriyati Atmosukarto.
Beyond bags of features: Adding spatial information Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Multi-Class Object Recognition Using Shared SIFT Features
Bag-of-features models
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
Visual Categorization with Bag of Keypoints
Object Class Recognition Using Discriminative Local Features Gyuri Dorko and Cordelia Schmid.
5/30/2006EE 148, Spring Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Discriminative and generative methods for bags of features
Machine learning & category recognition Cordelia Schmid Jakob Verbeek.
Review: Intro to recognition Recognition tasks Machine learning approach: training, testing, generalization Example classifiers Nearest neighbor Linear.
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Exercise Session 10 – Image Categorization
Multiclass object recognition
Step 3: Classification Learn a decision rule (classifier) assigning bag-of-features representations of images to different classes Decision boundary Zebra.
Building local part models for category-level recognition C. Schmid, INRIA Grenoble Joint work with G. Dorko, S. Lazebnik, J. Ponce.
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Svetlana Lazebnik, Cordelia Schmid, Jean Ponce
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Beyond Sliding Windows: Object Localization by Efficient Subwindow Search The best paper prize at CVPR 2008.
Jakob Verbeek December 11, 2009
An Introduction to Support Vector Machine (SVM)
CS 1699: Intro to Computer Vision Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh October 29, 2015.
Lecture 08 27/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Poselets: Body Part Detectors Trained Using 3D Human Pose Annotations ZUO ZHEN 27 SEP 2011.
CS654: Digital Image Analysis
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Does one size really fit all? Evaluating classifiers in a Bag-of-Visual-Words classification Christian Hentschel, Harald Sack Hasso Plattner Institute.
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Machine Learning and Category Representation Jakob Verbeek November 25, 2011 Course website:
Lecture IX: Object Recognition (2)
Video Google: Text Retrieval Approach to Object Matching in Videos
Paper Presentation: Shape and Matching
By Suren Manvelyan, Crocodile (nile crocodile?) By Suren Manvelyan,
Video Google: Text Retrieval Approach to Object Matching in Videos
Feature descriptors and matching
Presentation transcript:

Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning in Computer – 2004 Presented By: Xinwu Mo Prasad Samarakoon

Outline Introduction Method Experiments Conclusion

Outline Introduction – Visual Categorization Is NOT – Expected Goals – Bag of Words Analogy Method Experiments Conclusion

Introduction A method for generic visual categorization Face

PrasadXinwu Visual Categorization Is NOT Recognition Concerns the identification of particular object instances

Visual Categorization Is NOT Content based image retrieval Retrieving images on the basis of low-level image features

Visual Categorization Is NOT Detection Deciding whether or not a member of one visual category is present in a given image FaceYes CatNo

Visual Categorization Is NOT Detection Deciding whether or not a member of one visual category is present in a given image « One Visual Category » - sounds similar Yet most of the existing detection techniques require – Precise manual alignment of the training images – Segregation of these images into different views Bags of Keypoints don’t need any of these

Expected Goals Should be readily extendable Should handle the variations in view, imaging, lighting condition, occlusion Should handle intra class variations

Bag of Words Analogy Image Credits: Cordelia Schmid

Bag of Words Analogy Image Credits: Li Fei Fei

Bag of Words Analogy Image Credits: Li Fei Fei

Bag of Words Analogy Zhu et al – 2002 have used this method for categorization using small square image windows – called keyblocks But keyblocks don’t posses any invariance properties that Bags of Keypoints posses

Outline Introduction Method – Detection And Description of Image Patches – Assignment of Patch Descriptors – Contruction of Bag of Keypoints – Application of Multi-Class Classifier Experiments Conclusion

Method 4 main steps – Detection And Description of Image Patches – Assignment of Patch Descriptors – Contruction of Bag of Keypoints – Application of Multi-Class Classifier Categorization by Naive Bayes Categorization by SVM Designed to maximize classification accuracy while minimizing computational effort

Detection And Description of Image Patches Descriptors should be invariant to variation but have enough information to discriminate different categories Image Credits: Li Fei Fei

Detection And Description of Image Patches Detection – Harris affine detector – Last presentation by Guru and Shreyas Description – SIFT descriptor – 128 dimensional vector – 8 * (4*4)

Assignment of Patch Descriptors When a new query image is given, the derived descriptors should be assigned to ones that are already in our training dataset Check them with All the descriptors available in the training dataset – too expensive Only a few of them – but not too few The number of descriptors should be carefully selected

Assignment of Patch Descriptors Each patch has a descriptor, which is a point in some high-dimensional space (128) Image Credits: K. Grauman, B. Leibe

Assignment of Patch Descriptors Close points in feature space, means similar descriptors, which indicates similar local content Image Credits: K. Grauman, B. Leibe

Assignment of Patch Descriptors To reduce the huge number of descriptors involved ( ), they are clustered Using K-means K-means is run several times using different K values and initial positions One with the lowest empirical risk is used Image Credits: K. Grauman, B. Leibe

Assignment of Patch Descriptors Now the descriptor space looks like Feature space is quantized These cluster centers are the prototype words They make the vocabulary Image Credits: K. Grauman, B. Leibe

Assignment of Patch Descriptors When a query image comes Its descriptors are attached to the nearest cluster center That particular word is present in the query image Image Credits: K. Grauman, B. Leibe

Assignment of Patch Descriptors Vocabulary should be – Large enough to distinguish relevant changes in the image parts – Not so large that noise starts affecting the categorization

Contruction of Bags of Keypoints Summarize entire image based on its distribution (histogram) of word occurrences Image Credits: Li Fei Fei

Application of Multi-Class Classifier Apply a multi-class classifier, treat the bag of keypoints as the feature vector, thus determine which category or categories to assign to the image – Categorization by Naive Bayes – Categorization by SVM

Categorization by Naive Bayes Can be viewed as the maximum a posteriori probability classifier for a generative model To avoid zero probabilities of, Laplace smoothing is used

Categorization by SVM Find a hyperplane which separates two-class data with maximal margin

Categorization by SVM Classification function: f(x) = sign(w T x+b) where w, b parameters of the hyperplane

Categorization by SVM Data sets not always linearly separable – error weighting constant to penalizes misclassification of samples in proportion to their distance from the classification boundary – A mapping φ is made from the original data space of X to another feature space This is used in Bag of Keypoints

Categorization by SVM What do you mean by mapping function?

Categorization by SVM Can be formulated in terms of scalar products in the second feature space, by introducing the kernel Then the decision function becomes

Outline Introduction Method Experiments Conclusion

Experiments Some samples from the inhouse dataset

Experiments Impact of the number of clusters on classifier accuracy and evaluate the performance of – Naive Bayes classifier – SVM Three performance measures are used – Confusion matrix – Overall error rate – Mean ranks

Results

For K = 1000 – Naive Bayes 28%SVM 15%

Experiments Performance of SVM in another dataset

Results

Multiple objects of the same category/ partial view Misclassifications

Outline Introduction Method Experiments Conclusion – Future Work

Conclusion Advantages – Bag of Keypoints is simple – Computationally efficient – Invariant to affine transformations, occlusions, lighting, intra-class variations

Future Work Extend to more visual categories Extend the categorizer to incorporate geometric information Make the method robust when the object of interest is occupying only a small fraction of the image Investigate many alternatives for each of the four steps of the basic method

Q&A Thank You