Beyond Nouns Exploiting Preposition and Comparative adjectives for learning visual classifiers.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Presented by, Biswaranjan Panda and Moutupsi Paul Beyond Nouns -Exploiting Prepositions and Comparative Adjectives for Learning Visual Classifiers Ref.
Weakly supervised learning of MRF models for image region labeling Jakob Verbeek LEAR team, INRIA Rhône-Alpes.
Context-based object-class recognition and retrieval by generalized correlograms by J. Amores, N. Sebe and P. Radeva Discussion led by Qi An Duke University.
CVPR2013 Poster Representing Videos using Mid-level Discriminative Patches.
Patch to the Future: Unsupervised Visual Prediction
Human Action Recognition by Learning Bases of Action Attributes and Parts.
Automatic Image Annotation and Retrieval using Cross-Media Relevance Models J. Jeon, V. Lavrenko and R. Manmathat Computer Science Department University.
Generic Object Detection using Feature Maps Oscar Danielsson Stefan Carlsson
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
CVPR SLAM 2007 Using Group Prior to Identify People In Consumer Images Andrew C. Gallagher Tsuhan Chen Carnegie Mellon University Eastman Kodak Company.
Presented by Zeehasham Rasheed
Boosting Main idea: train classifiers (e.g. decision trees) in a sequence. a new classifier should focus on those cases which were incorrectly classified.
Beyond Nouns Abhinav Gupta and Larry S. Davis University of Maryland, College Park Exploiting Prepositions and Comparative Adjectives for Learning Visual.
Semi-Supervised Learning
Exercise Session 10 – Image Categorization
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
Disambiguation of References to Individuals Levon Lloyd (State University of New York) Varun Bhagwan, Daniel Gruhl (IBM Research Center) Varun Bhagwan,
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
Automatic Extraction of Opinion Propositions and their Holders Steven Bethard, Hong Yu, Ashley Thornton, Vasileios Hatzivassiloglou and Dan Jurafsky Department.
1 Wikification CSE 6339 (Section 002) Abhijit Tendulkar.
SVCL Automatic detection of object based Region-of-Interest for image compression Sunhyoung Han.
Exploiting Ontologies for Automatic Image Annotation M. Srikanth, J. Varner, M. Bowden, D. Moldovan Language Computer Corporation
Distributional Part-of-Speech Tagging Hinrich Schütze CSLI, Ventura Hall Stanford, CA , USA NLP Applications.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Languages and Images Virginia Tech ECE /04/25 Stanislaw Antol.
Beyond Co-occurrence: Discovering and Visualizing Tag Relationships from Geo-spatial and Temporal Similarities Date : 2012/8/6 Resource : WSDM’12 Advisor.
Why Categorize in Computer Vision ?. Why Use Categories? People love categories!
Ontology-Driven Automatic Entity Disambiguation in Unstructured Text Jed Hassell.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
Fine-Grained Location Extraction from Tweets with Temporal Awareness Date:2015/03/19 Author:Chenliang Li, Aixin Sun Source:SIGIR '14 Advisor:Jia-ling Koh.
Experimental Evaluation of Learning Algorithms Part 1.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Chapter 4: Pattern Recognition. Classification is a process that assigns a label to an object according to some representation of the object’s properties.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
Object Recognition a Machine Translation Learning a Lexicon for a Fixed Image Vocabulary Miriam Miklofsky.
MedIX – Summer 07 Lucia Dettori (room 745)
Using Webcast Text for Semantic Event Detection in Broadcast Sports Video IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 10, NO. 7, NOVEMBER 2008.
Tell Me What You See and I will Show You Where It Is Jia Xu 1 Alexander G. Schwing 2 Raquel Urtasun 2,3 1 University of Wisconsin-Madison 2 University.
Some questions -What is metadata? -Data about data.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
Presented By- Shahina Ferdous, Student ID – , Spring 2010.
Total Recall: Automatic Query Expansion with a Generative Feature Model for Object Retrieval O. Chum, et al. Presented by Brandon Smith Computer Vision.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
Class Imbalance in Text Classification
Annotation Framework & ImageCLEF 2014 JAN BOTOREK, PETRA BUDÍKOVÁ
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
1.Learn appearance based models for concepts 2.Compute posterior probabilities or Semantic Multinomial (SMN) under appearance models. -But, suffers from.
Feature Assignment LBSC 878 February 22, 1999 Douglas W. Oard and Dagobert Soergel.
Semantic search-based image annotation Petra Budíková, FI MU CEMI meeting, Plzeň,
Opinion spam and Analysis 소프트웨어공학 연구실 G 최효린 1 / 35.
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
CNN-RNN: A Unified Framework for Multi-label Image Classification
Erasmus University Rotterdam
Lecture 15: Text Classification & Naive Bayes
Machine Learning Week 1.
Quanzeng You, Jiebo Luo, Hailin Jin and Jianchao Yang
Statistical NLP: Lecture 9
RGB-D Image for Scene Recognition by Jiaqi Guo
Tagging Review Comments Rationale #10 Week 13
CS 620 Class Presentation Using WordNet to Improve User Modelling in a Web Document Recommender System Using WordNet to Improve User Modelling in a Web.
LECTURE 23: INFORMATION THEORY REVIEW
Presented by Wanxue Dong
Information Retrieval
Evaluating Classifiers
Statistical NLP : Lecture 9 Word Sense Disambiguation
Presentation transcript:

Beyond Nouns Exploiting Preposition and Comparative adjectives for learning visual classifiers

Nouns Nouns Images Widely used Object detection Tag recommendation etc Co-occurrence over large data -> relations

Co-occurrence Examples Jack and Jill Laurel and Hardy

Co-occurrence

Co-occurrence Co-occurrence is good To detect whether the objects are there Eg. Suggesting tags Co-occurrence is bad When we want to disambiguate We might never be able to tell Jack and Jill apart

Relationships How do they help? Additional constraints limit possibilities Other intuitive examples Aptitude questions Like ILP? “Since cars are typically found on streets, it is difficult to resolve the correspondence using co- occurrence alone”

Relationships

Relationships How do we learn them? Manually annotate Learn model Use relationship to predict in testing …. Or …

The Problem We have a weakly labeled dataset (tags only) Relationship model helps us label it strongly Strong labeling helps us derive relationship model Therefore, EM. The labeling (assignment) is treated as the missing data.

Feature Representation

Model

EM

EM M For the noun assignment done earlier, we learn relationship and object classifiers The relationship classifier is modeled on a single feature … GOTO E Boot strapping can be done using any image annotation approach

Model

Inference

Inference

Testing Likelihood models Nouns: Nearest Neighbor Relationships: Decision Stump Evaluation on: subset of Corel5k Training on: 850 images with (173) nouns and (19) hand labeled relationships

Evaluation – resolution of correspondence ambiguities Metrics range of semantics number of unique nouns correct (?) frequency correct number of total nouns correct (?) Compared with Image annotation algorithms Human assisted annotation (are machines better at relationships than us?)

Results

Examples

Evaluation – labeling new images Tested on random subset of Corel5k, based on learnt vocabulary Labels verified by “us” (Corel is misleading) Precision/Recall While recall rates are reported with respect to corel annotations, precision rates are reported with respect to correctness defined by human observers.

Results

Results

Examples