UNBIASED LOOK AT DATASET BIAS Antonio Torralba Massachusetts Institute of Technology Alexei A. Efros Carnegie Mellon University CVPR 2011.

Slides:



Advertisements
Similar presentations
James Hays and Alexei A. Efros Carnegie Mellon University CVPR IM2GPS: estimating geographic information from a single image Wen-Tsai Huang.
Advertisements

Rapid Object Detection using a Boosted Cascade of Simple Features Paul Viola, Michael Jones Conference on Computer Vision and Pattern Recognition 2001.
Rapid Object Detection using a Boosted Cascade of Simple Features Paul Viola, Michael Jones Conference on Computer Vision and Pattern Recognition 2001.
What makes an image memorable?
3 Small Comments Alex Berg Stony Brook University I work on recognition: features – action recognition – alignment – detection – attributes – hierarchical.
Proactive Learning: Cost- Sensitive Active Learning with Multiple Imperfect Oracles Pinar Donmez and Jaime Carbonell Pinar Donmez and Jaime Carbonell Language.
Data-driven Visual Similarity for Cross-domain Image Matching
1 Building a Dictionary of Image Fragments Zicheng Liao Ali Farhadi Yang Wang Ian Endres David Forsyth Department of Computer Science, University of Illinois.
Object-centric spatial pooling for image classification Olga Russakovsky, Yuanqing Lin, Kai Yu, Li Fei-Fei ECCV 2012.
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
Sketch Tokens: A Learned Mid-level Representation for Contour and Object Detection CVPR2013 POSTER.
Large dataset for object and scene recognition A. Torralba, R. Fergus, W. T. Freeman 80 million tiny images Ron Yanovich Guy Peled.
Proceedings of the IEEE 2010 Antonio Torralba, MIT Jenny Yuen, MIT Bryan C. Russell, MIT.
Recognition: A machine learning approach
Tour the World: building a web-scale landmark recognition engine ICCV 2009 Yan-Tao Zheng1, Ming Zhao2, Yang Song2, Hartwig Adam2 Ulrich Buddemeier2, Alessandro.
Small Codes and Large Image Databases for Recognition CVPR 2008 Antonio Torralba, MIT Rob Fergus, NYU Yair Weiss, Hebrew University.
Fast and Compact Retrieval Methods in Computer Vision Part II A. Torralba, R. Fergus and Y. Weiss. Small Codes and Large Image Databases for Recognition.
Statistical Recognition Slides adapted from Fei-Fei Li, Rob Fergus, Antonio Torralba, and Kristen Grauman.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
The Human Body For Miss. McGuire’s Fourth Grade Class.
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
Let Computer Draw Qingyuan Kong. Goal Give me a picture “Obama stands in front of a pyramid”
Object Recognition: Conceptual Issues Slides adapted from Fei-Fei Li, Rob Fergus, Antonio Torralba, and K. Grauman.
Where computer vision needs help from computer science (and machine learning) Bill Freeman Electrical Engineering and Computer Science Dept. Massachusetts.
Object Recognition: Conceptual Issues Slides adapted from Fei-Fei Li, Rob Fergus, Antonio Torralba, and K. Grauman.
Agenda Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based.
Visual Object Recognition Rob Fergus Courant Institute, New York University
Extreme, Non-parametric Object Recognition 80 million tiny images (Torralba et al)
Spatial Pyramid Pooling in Deep Convolutional
Predicting Matchability - CVPR 2014 Paper -
Opportunities of Scale, Part 2 Computer Vision James Hays, Brown Many slides from James Hays, Alyosha Efros, and Derek Hoiem Graphic from Antonio Torralba.
Creating and Exploring a Large Photorealistic Virtual Space INRIA / CSAIL / Adobe First IEEE Workshop on Internet Vision, associated with CVPR 2008.
Internet-scale Imagery for Graphics and Vision James Hays cs195g Computational Photography Brown University, Spring 2010.
Salient Object Detection by Composition
Exercise Session 10 – Image Categorization
Overcoming Dataset Bias: An Unsupervised Domain Adaptation Approach Boqing Gong University of Southern California Joint work with Fei Sha and Kristen Grauman.
Studying Relationships Between Human Gaze, Description, and Computer Vision Kiwon Yun 1, Yifan Peng 1 Dimitris Samaras 1, Gregory J. Zelinsky 1,2, Tamara.
Computer Vision CS 776 Spring 2014 Recognition Machine Learning Prof. Alex Berg.
Last part: datasets and object collections. CMU/MIT frontal facesvasc.ri.cmu.edu/idb/html/face/frontal_images cbcl.mit.edu/software-datasets/FaceData2.html.
Model comparison and challenges II Compositional bias of salient object detection benchmarking Xiaodi Hou K-Lab, Computation and Neural Systems California.
Incremental Learning Chris Mesterharm Fordham University.
Reading Between The Lines: Object Localization Using Implicit Cues from Image Tags Sung Ju Hwang and Kristen Grauman University of Texas at Austin Jingnan.
80 million tiny images: a large dataset for non-parametric object and scene recognition CS 4763 Multimedia Systems Spring 2008.
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
School of Engineering and Computer Science Victoria University of Wellington Copyright: Peter Andreae, VUW Image Recognition COMP # 18.
Geodesic Flow Kernel for Unsupervised Domain Adaptation Boqing Gong University of Southern California Joint work with Yuan Shi, Fei Sha, and Kristen Grauman.
Object detection, deep learning, and R-CNNs
The Viola/Jones Face Detector A “paradigmatic” method for real-time object detection Training is slow, but detection is very fast Key ideas Integral images.
1 CHUKWUEMEKA DURUAMAKU.  Machine learning, a branch of artificial intelligence, concerns the construction and study of systems that can learn from data.
Image Classification over Visual Tree Jianping Fan Dept of Computer Science UNC-Charlotte, NC
Recognition Using Visual Phrases
Context Neelima Chavali ECE /21/2013. Roadmap Introduction Paper1 – Motivation – Problem statement – Approach – Experiments & Results Paper 2 Experiments.
SUN Database: Large-scale Scene Recognition from Abbey to Zoo Jianxiong Xiao *James Haysy Krista A. Ehinger Aude Oliva Antonio Torralba Massachusetts Institute.
Carl Vondrick, Aditya Khosla, Tomasz Malisiewicz, Antonio Torralba Massachusetts Institute of Technology
Internet-scale Imagery for Graphics and Vision James Hays cs129 Computational Photography Brown University, Spring 2011.
ICCV 2009 Tilke Judd, Krista Ehinger, Fr´edo Durand, Antonio Torralba.
University of Ioannina
Machine Learning Dr. Mohamed Farouk.
Li Fei-Fei, UIUC Rob Fergus, MIT Antonio Torralba, MIT
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
MACHINE LEARNING.
Finding Clusters within a Class to Improve Classification Accuracy
CS 1674: Intro to Computer Vision Scene Recognition
CVPR 2014 Orientational Pyramid Matching for Recognizing Indoor Scenes
Recognizing and Learning Object categories
Rob Fergus Computer Vision
Brief Review of Recognition + Context
a chicken and egg problem…
Similarity based on Shape and Appearance
Objects as Attributes for Scene Classification
Presentation transcript:

UNBIASED LOOK AT DATASET BIAS Antonio Torralba Massachusetts Institute of Technology Alexei A. Efros Carnegie Mellon University CVPR 2011

Outline  1. Introduction  2. Measuring Dataset Bias  3. Measuring Dataset’s Value  4. Discussion

Name That Dataset!  Let’s play a game!

Answer 1Caltech-101 2UIUC 3MSRC 4Tiny Images 5ImageNet 6PASCAL VOC 7LabelMe 8SUNS Scenes 10Corel 11Caltech COIL-100

UIUC test set is not the same as its training set COIL is a lab-based dataset Caltech101 and Caltech256 are predictably confused with each other

Caltech 101 Caltech256 Caltech256  Pictures of objects belonging to 101 categories. About 40 to 800 images per category  Most categories have about 50 images  Collected in September 2003  The size of each image is roughly 300 x 200 pixels

LabelMe  A project created by the MIT Computer Science and Artificial Intelligence Laboratory (CSAIL)  A dataset of digital images with annotations  The most applicable use of LabelMe is in computer vision research  As of October 31, 2010, LabelMe has 187,240 images, 62,197 annotated images, and 658,992 labeled objects

Bias  Urban scenes Rural landscapes  Professional photographs Amateur snapshots  Entire scenes Single objects

The Rise of the Modern Dataset  COIL-100 dataset (a hundred household objects on a black background)  Corel and 15 Scenes were Professional collections visual complexity  Caltech-101 (101 objects using Google and cleaned by hand) wilderness of the Internet  MSRC and LabelMe (both researcher-collected sets), complex scenes with many objects

The Rise of the Modern Dataset  PASCAL Visual Object Classes (VOC) was a reaction against the lax training and testing standards of previous datasets  The batch of very-large-scale, Internet-mined datasets – Tiny Images, ImageNet, and SUN09 – can be considered a reaction against the inadequacies of training and testing on datasets that are just too small for the complexity of the real world

Outline  2. Measuring Dataset Bias Cross-dataset generalization Negative Set Bias

Cross-dataset generalization

Negative Set Bias  Evaluate the relative bias in the negative sets of different datasets (e.g. is a “not car” in PASCAL different from “not car” in MSRC?).  For each dataset, we train a classifier on its own set of positive and negative instances. Then, during testing, the positives come from that dataset, but the negatives come from all datasets combined

Outline  3. Measuring Dataset’s Value

Measuring Dataset’s Value  Given a particular detection task and benchmark, there are two basic ways of improving the performance  The first solution is to improve the features, the object representation and the learning algorithm for the detector  The second solution is to simply enlarge the amount of data available for training

Market Value for a car sample across datasets

Outline  4. Discussion

Discussion  Caltech-101 is extremely biased with virtually no observed generalization, and should have been retired long ago (as arguedby [14] back in 2006)  MSRC has also fared very poorly.  PASCAL VOC, ImageNet and SUN09, have fared comparatively well