5/30/2006EE 148, Spring 20061 Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Tamara Berg Object Recognition – BoF models Recognizing People, Objects, & Actions 1.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Presented by Xinyu Chang
Machine learning continued Image source:
Indian Statistical Institute Kolkata
MIT CSAIL Vision interfaces Approximate Correspondences in High Dimensions Kristen Grauman* Trevor Darrell MIT CSAIL (*) UT Austin…
CS395: Visual Recognition Spatial Pyramid Matching Heath Vinicombe The University of Texas at Austin 21 st September 2012.
Learning Visual Similarity Measures for Comparing Never Seen Objects Eric Nowak, Frédéric Jurie CVPR 2007.
Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem
Discriminative and generative methods for bags of features
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Generic Object Recognition -- by Yatharth Saraf A Project on.
Multi-Class Object Recognition Using Shared SIFT Features
Image Categorization by Learning and Reasoning with Regions Yixin Chen, University of New Orleans James Z. Wang, The Pennsylvania State University Published.
Visual Categorization with Bag of Keypoints
Object Class Recognition Using Discriminative Local Features Gyuri Dorko and Cordelia Schmid.
Redaction: redaction: PANAKOS ANDREAS. An Interactive Tool for Color Segmentation. An Interactive Tool for Color Segmentation. What is color segmentation?
Discriminative and generative methods for bags of features
© 2013 IBM Corporation Efficient Multi-stage Image Classification for Mobile Sensing in Urban Environments Presented by Shashank Mujumdar IBM Research,
Exercise Session 10 – Image Categorization
Bag-of-Words based Image Classification Joost van de Weijer.
Step 3: Classification Learn a decision rule (classifier) assigning bag-of-features representations of images to different classes Decision boundary Zebra.
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
Data mining and machine learning A brief introduction.
Watch, Listen and Learn Sonal Gupta, Joohyun Kim, Kristen Grauman and Raymond Mooney -Pratiksha Shah.
Marcin Marszałek, Ivan Laptev, Cordelia Schmid Computer Vision and Pattern Recognition, CVPR Actions in Context.
Why Categorize in Computer Vision ?. Why Use Categories? People love categories!
Svetlana Lazebnik, Cordelia Schmid, Jean Ponce
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
CSSE463: Image Recognition Day 31 Today: Bayesian classifiers Today: Bayesian classifiers Tomorrow: project day. Tomorrow: project day. Questions? Questions?
Representations for object class recognition David Lowe Department of Computer Science University of British Columbia Vancouver, Canada Sept. 21, 2006.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Beyond Sliding Windows: Object Localization by Efficient Subwindow Search The best paper prize at CVPR 2008.
Today Ensemble Methods. Recap of the course. Classifier Fusion
School of Engineering and Computer Science Victoria University of Wellington Copyright: Peter Andreae, VUW Image Recognition COMP # 18.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Sparse Bayesian Learning for Efficient Visual Tracking O. Williams, A. Blake & R. Cipolloa PAMI, Aug Presented by Yuting Qi Machine Learning Reading.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
1 CSC 4510, Spring © Paula Matuszek CSC 4510 Support Vector Machines (SVMs)
Kylie Gorman WEEK 1-2 REVIEW. CONVERTING AN IMAGE FROM RGB TO HSV AND DISPLAY CHANNELS.
1 Classification and Feature Selection Algorithms for Multi-class CGH data Jun Liu, Sanjay Ranka, Tamer Kahveci
Lecture 08 27/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Poselets: Body Part Detectors Trained Using 3D Human Pose Annotations ZUO ZHEN 27 SEP 2011.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #23.
Irfan Ullah Department of Information and Communication Engineering Myongji university, Yongin, South Korea Copyright © solarlits.com.
Goggle Gist on the Google Phone A Content-based image retrieval system for the Google phone Manu Viswanathan Chin-Kai Chang Ji Hyun Moon.
On Using SIFT Descriptors for Image Parameter Evaluation Authors: Patrick M. McInerney 1, Juan M. Banda 1, and Rafal A. Angryk 2 1 Montana State University,
9.913 Pattern Recognition for Vision Class9 - Object Detection and Recognition Bernd Heisele.
Gist 2.3 John H. Phan MIBLab Summer Workshop June 28th, 2006.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Does one size really fit all? Evaluating classifiers in a Bag-of-Visual-Words classification Christian Hentschel, Harald Sack Hasso Plattner Institute.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Machine Learning and Category Representation Jakob Verbeek November 25, 2011 Course website:
Another Example: Circle Detection
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Bag-of-Visual-Words Based Feature Extraction
Data Driven Attributes for Action Detection
Perceptrons Lirong Xia.
Introductory Seminar on Research: Fall 2017
Features & Decision regions
Machine Learning Week 1.
Categorization by Learning and Combing Object Parts
Brief Review of Recognition + Context
CSCI N317 Computation for Scientific Applications Unit Weka
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Perceptrons Lirong Xia.
Presentation transcript:

5/30/2006EE 148, Spring Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray Presented by Yun-hsueh Liu

5/30/2006EE 148, Spring What is Generic Visual Categorization? Categorization: distinguish different classes Generic Visual Categorization: Generic  to cope with many object types simultaneously readily extended to new object types. Handle the variation in view, imaging, lighting, occlusion, and typical object and scene variations

5/30/2006EE 148, Spring Previous Work in Computational Vision Single Category Detection Decide if a member of one visual category is present in a given image. (faces, cars, targets) Content Based Image Retrieval Retrieve images on the basis of low-level image features, such as colors or textures. Recognition Distinguish between images of structurally distinct objects within one class. (say, different cell phones)

5/30/2006EE 148, Spring Bag-of-Keypoints Approach Interesting Point Detection Key Patch Extraction Feature Descriptors Bag of Keypoints Multi-class Classifier

5/30/2006EE 148, Spring SIFT Descriptors Interesting Point Detection Key Patch Extraction Feature Descriptors Bag of Keypoints Multi-class Classifier

5/30/2006EE 148, Spring Bag of Keypoints (1) Construction of a vocabulary Kmeans clustering  find “centroids” (on all the descriptors we find from all the training images) Define a “vocabulary” as a set of “centroids”, where every centroid represents a “word”. Interesting Point Detection Key Patch Extraction Feature Descriptors Bag of Keypoints Multi-class Classifier

5/30/2006EE 148, Spring Bag of Keypoints (2) Histogram Counts the number of occurrences of different visual words in each image Interesting Point Detection Key Patch Extraction Feature Descriptors Bag of Keypoints Multi-class Classifier

5/30/2006EE 148, Spring Multi-class Classifier In this paper, classification is based on conventional machine learning approaches Naïve Bayes Support Vector Machine (SVM) Interesting Point Detection Key Patch Extraction Feature Descriptors Bag of Keypoints Multi-class Classifier

5/30/2006EE 148, Spring Multi-class classifier – Naïve Bayes (1) Let V = {v i }, i = 1,…,N, be a visual vocabulary, in which each v i represents a visual word (cluster centers) from the feature space. A set of labeled images I = {I i }. Denote C j to represent our Classes, where j = 1,..,M N(t,i) = number of times v i occurs in image I i (keypoint histogram) Score approach: want to determine P(C j |I i ), where (*)

5/30/2006EE 148, Spring Multi-class Classifier – Naïve Bayes (2) Goal: Find one specific class C j so that has maximum value In order to avoid zero probability, use Laplace smoothing:

5/30/2006EE 148, Spring Multi-class classifier – Support Vector Machine (SVM) Input: the keypoints histogram for each image Multi-class  one-against-all approach Linear SVM gives better performances than quadratic or cubic SVM Goal: find hyperplanes which separate multi-class data with maximun margin

5/30/2006EE 148, Spring Multi-class classifier – SVM (2)

5/30/2006EE 148, Spring Evaluation of Multi-class Classifiers Three performance measures: The confusion matrix Each column of the matrix represents the instances in a predicted class Each row represents the instances in an actual class The overall error rate  = Pr(output class = true class) The mean ranks The mean position of the correct labels when labels output by the multi- class classifier are sorted by the classifier score.

5/30/2006EE 148, Spring n-Fold Cross Validation What is “fold”? Randomly break the dataset into n partitions Example: suppose n = 10 Training on 2, 3,…,10; testing on 1 = result 1 Training on 1, 3,…,10; testing on 2 = result 2 … Answer = Average of result 1, result 2, ….

5/30/2006EE 148, Spring Experiment on Naïve Bayes – k’s effect Present the overal error rate as a function of # of clusters k Result Error rate decreases as k increases Selecting point: k = 1000 After passing the selecting point, the error rate decreases slowly

5/30/2006EE 148, Spring Experiment on Naïve Bayes – Confusion Matrix facesbuildingstreescarsphonesbikesbooks faces buildings trees cars phones bikes books error rate mean rank

5/30/2006EE 148, Spring Experiment on SVM – Confusion Matrix facesbuildingstreescarsphonesbikesbooks faces buildings trees cars phones bikes books error rate mean rank

5/30/2006EE 148, Spring Interpretation of Results The confusion matrix In general, SVM has more correct predictions than Naïve Bayes does The overall error rate In general, Naïve Bayes > SVM The Mean Rank In general, SVM < Naïve Bayes

5/30/2006EE 148, Spring Why do we have errors? There are objects from more than 2 classes in one image The data set is not totally clean (noise) Each image is given only one training label

5/30/2006EE 148, Spring Conclusion Bag-of-Keypoints is a new and efficient generic visual categorizer. Evaluated on a seven-category database, this method is proved that it is robust to Choice of clusters, background clutter, multiple objects Any Questions? Thank you for listening to my presentation!! :)