1 Data Collection and Normalization for the Scenario- Based Lexical Knowledge Resource of a Text-to-Scene Conversion System Margit Bowler.

Slides:



Advertisements
Similar presentations
Relevance Feedback & Query Expansion
Advertisements

MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
1 Extended Gloss Overlaps as a Measure of Semantic Relatedness Satanjeev Banerjee Ted Pedersen Carnegie Mellon University University of Minnesota Duluth.
Applying Crowd Sourcing and Workflow in Social Conflict Detection By: Reshmi De, Bhargabi Chakrabarti 28/03/13.
Precision and Recall.
Evaluating Search Engine
Sentiment Lexicon Creation from Lexical Resources BIS 2011 Bas Heerschop Erasmus School of Economics Erasmus University Rotterdam
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
Semantic distance in WordNet: An experimental, application-oriented evaluation of five measures Presenter: Cosmin Adrian Bejan Alexander Budanitsky and.
Predicting the Semantic Orientation of Adjectives
Queensland University of Technology An Ontology-based Mining Approach for User Search Intent Discovery Yan Shen, Yuefeng Li, Yue Xu, Renato Iannella, Abdulmohsen.
Designing clustering methods for ontology building: The Mo’K workbench Authors: Gilles Bisson, Claire Nédellec and Dolores Cañamero Presenter: Ovidiu Fortu.
Article by: Feiyu Xu, Daniela Kurz, Jakub Piskorski, Sven Schmeier Article Summary by Mark Vickers.
Word Sense Disambiguation for Automatic Taxonomy Construction from Text-Based Web Corpora 12th International Conference on Web Information System Engineering.
Mining and Summarizing Customer Reviews
Mining the Peanut Gallery: Opinion Extraction and Semantic Classification of Product Reviews K. Dave et al, WWW 2003, citations Presented by Sarah.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics – Bag of concepts – Semantic distance between two words.
Automatic Extraction of Opinion Propositions and their Holders Steven Bethard, Hong Yu, Ashley Thornton, Vasileios Hatzivassiloglou and Dan Jurafsky Department.
Learning Information Extraction Patterns Using WordNet Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield,
Information Retrieval and Web Search Relevance Feedback. Query Expansion Instructor: Rada Mihalcea Class web page:
Computational Lexical Semantics Lecture 8: Selectional Restrictions Linguistic Institute 2005 University of Chicago.
PropBank, VerbNet & SemLink Edward Loper. PropBank 1M words of WSJ annotated with predicate- argument structures for verbs. –The location & type of each.
An Integrated Approach to Extracting Ontological Structures from Folksonomies Huairen Lin, Joseph Davis, Ying Zhou ESWC 2009 Hyewon Lim October 9 th, 2009.
RuleML-2007, Orlando, Florida1 Towards Knowledge Extraction from Weblogs and Rule-based Semantic Querying Xi Bai, Jigui Sun, Haiyan Che, Jin.
COMP423.  Query expansion  Two approaches ◦ Relevance feedback ◦ Thesaurus-based  Most Slides copied from ◦
Automatic Lexical Annotation Applied to the SCARLET Ontology Matcher Laura Po and Sonia Bergamaschi DII, University of Modena and Reggio Emilia, Italy.
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Using Text Mining and Natural Language Processing for Health Care Claims Processing Cihan ÜNAL
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Jennie Ning Zheng Linda Melchor Ferhat Omur. Contents Introduction WordNet Application – WordNet Data Structure - WordNet FrameNet Application – FrameNet.
Annotating Words using WordNet Semantic Glosses Julian Szymański Department of Computer Systems Architecture, Faculty of Electronics, Telecommunications.
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
1 Query Operations Relevance Feedback & Query Expansion.
WORD SENSE DISAMBIGUATION STUDY ON WORD NET ONTOLOGY Akilan Velmurugan Computer Networks – CS 790G.
SYMPOSIUM ON SEMANTICS IN SYSTEMS FOR TEXT PROCESSING September 22-24, Venice, Italy Combining Knowledge-based Methods and Supervised Learning for.
10/22/2015ACM WIDM'20051 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis Voutsakis.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
Modelling Human Thematic Fit Judgments IGK Colloquium 3/2/2005 Ulrike Padó.
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li and Li Fei-Fei Dept. of Computer Science, Princeton University, USA CVPR ImageNet1.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
Authors: Marius Pasca and Benjamin Van Durme Presented by Bonan Min Weakly-Supervised Acquisition of Open- Domain Classes and Class Attributes from Web.
© 2004 Chris Staff CSAW’04 University of Malta of 15 Expanding Query Terms in Context Chris Staff and Robert Muscat Department of.
WordNet Enhancements: Toward Version 2.0 WordNet Connectivity Derivational Connections Disambiguated Definitions Topical Connections.
Query Suggestion. n A variety of automatic or semi-automatic query suggestion techniques have been developed  Goal is to improve effectiveness by matching.
1 Masters Thesis Presentation By Debotosh Dey AUTOMATIC CONSTRUCTION OF HASHTAGS HIERARCHIES UNIVERSITAT ROVIRA I VIRGILI Tarragona, June 2015 Supervised.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Information Retrieval and Web Search Relevance Feedback. Query Expansion Instructor: Rada Mihalcea.
Commonsense Reasoning in and over Natural Language Hugo Liu, Push Singh Media Laboratory of MIT The 8 th International Conference on Knowledge- Based Intelligent.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Knowledge Structure Vijay Meena ( ) Gaurav Meena ( )
2/10/2016Semantic Similarity1 Semantic Similarity Methods in WordNet and Their Application to Information Retrieval on the Web Giannis Varelas Epimenidis.
Towards Semi-Automated Annotation for Prepositional Phrase Attachment Sara Rosenthal William J. Lipovsky Kathleen McKeown Kapil Thadani Jacob Andreas Columbia.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Semantic Interoperability in GIS N. L. Sarda Suman Somavarapu.
Semantic search-based image annotation Petra Budíková, FI MU CEMI meeting, Plzeň,
Sentiment Analysis Using Common- Sense and Context Information Basant Agarwal 1,2, Namita Mittal 2, Pooja Bansal 2, and Sonal Garg 2 1 Department of Computer.
Query expansion COMP423. Menu Query expansion Two approaches Relevance feedback Thesaurus-based Most Slides copied from
COMP423: Intelligent Agent Text Representation. Menu – Bag of words – Phrase – Semantics Semantic distance between two words.
Multimedia Information Retrieval
WordNet: A Lexical Database for English
CS 620 Class Presentation Using WordNet to Improve User Modelling in a Web Document Recommender System Using WordNet to Improve User Modelling in a Web.
Automatic Detection of Causal Relations for Question Answering
Relevance Feedback & Query Expansion
Giannis Varelas Epimenidis Voutsakis Paraskevi Raftopoulou
Precision and Recall.
Presentation transcript:

1 Data Collection and Normalization for the Scenario- Based Lexical Knowledge Resource of a Text-to-Scene Conversion System Margit Bowler

2 Who I Am  Rising senior at Reed College in Portland  Linguistics major, concentration in Russian  Rising senior at Reed College in Portland  Linguistics major, concentration in Russian

3 Overview  WordsEye & Scenario-Based Lexical Knowledge Resource (SBLR)  Use of Amazon’s Mechanical Turk (AMT) for data collection  Manual normalization of the AMT data and definition of semantic relations  Automatic normalization techniques of AMT data with respect to building the SBLR  Future automatic normalization techniques  WordsEye & Scenario-Based Lexical Knowledge Resource (SBLR)  Use of Amazon’s Mechanical Turk (AMT) for data collection  Manual normalization of the AMT data and definition of semantic relations  Automatic normalization techniques of AMT data with respect to building the SBLR  Future automatic normalization techniques

4 WordsEye Text-to-Scene Conversion the humongous white shiny bear is on the american mountain range. the mountain range is 100 feet tall. the ground is water. the sky is partly cloudy. the airplane is 90 feet in front of the nose of the bear. the airplane is facing right.

5 Scenario-Based Lexical Knowledge Resource (SBLR)  Information on semantic categories of words  Semantic relations between predicates (verbs, nouns, adjectives, prepositions) and their arguments  Contextual, common-sense knowledge about the visual scenes various actions and items occur in  Information on semantic categories of words  Semantic relations between predicates (verbs, nouns, adjectives, prepositions) and their arguments  Contextual, common-sense knowledge about the visual scenes various actions and items occur in

6 How to build the SBLR… efficiently?  Manual construction of the SBLR is time- consuming and expensive  Past methods have included mining information from external semantic resources (e.g. WordNet, FrameNet, PropBank) & information extraction techniques from other corpora  Manual construction of the SBLR is time- consuming and expensive  Past methods have included mining information from external semantic resources (e.g. WordNet, FrameNet, PropBank) & information extraction techniques from other corpora

7 Amazon’s Mechanical Turk (AMT)  Online marketplace for work  Anyone can work on AMT, however:  It is possible to screen workers by various criteria. We screened ours by:  Located in the USA  99%+ approval rating  Online marketplace for work  Anyone can work on AMT, however:  It is possible to screen workers by various criteria. We screened ours by:  Located in the USA  99%+ approval rating

8 AMT Tasks  In each task, we asked for up to 10 responses. A comment box was provided for >10 responses.  Task 1: Given the object X, name 10 locations where you would find X. (Locations)  Task 2: Given the object X, name 10 objects found near X. (Nearby Objects)  Task 3: Given the object X, list 10 parts of X. (Part- Whole)  In each task, we asked for up to 10 responses. A comment box was provided for >10 responses.  Task 1: Given the object X, name 10 locations where you would find X. (Locations)  Task 2: Given the object X, name 10 objects found near X. (Nearby Objects)  Task 3: Given the object X, list 10 parts of X. (Part- Whole)

9 AMT Task Results  17,200 total responses  Spent $ for all three tasks  It took approximately 5 days to complete each task  17,200 total responses  Spent $ for all three tasks  It took approximately 5 days to complete each task Target Words User Inputs Reward Locations 3426,850$0.05 Objects 2453,500$0.07 Parts 3426,850$0.05

10 Goal: How to automatically normalize data collected from AMT in such a way that AMT would be useful for building the Scenario-Based Lexical Knowledge Resource (SBLR)?

11 Manual Normalization of AMT Data  Removal of uninformative target item- response item pairs between which no relevant semantic relationship was held  Definition of the semantic relations held between the remaining target item-response item pairs  This manually normalized set of data was used as the standard against which we measured various automatic normalization techniques.  Removal of uninformative target item- response item pairs between which no relevant semantic relationship was held  Definition of the semantic relations held between the remaining target item-response item pairs  This manually normalized set of data was used as the standard against which we measured various automatic normalization techniques.

12 Rejected Target-Response Pairs  Misinterpretation of ambiguous target item (e.g. mobile)  Viable interpretation of target item was not contained within the SBLR (e.g. crawfish as food rather than a living animal)  Too generic responses (e.g. store in response to turntable)  Misinterpretation of ambiguous target item (e.g. mobile)  Viable interpretation of target item was not contained within the SBLR (e.g. crawfish as food rather than a living animal)  Too generic responses (e.g. store in response to turntable)

13 Examples of Approved AMT Responses  Locations: mural - gallery lizard - desert  Nearby Objects: ambulance - stretcher cauldron - fire  Part-Whole: scissors - blade monument - granite  Locations: mural - gallery lizard - desert  Nearby Objects: ambulance - stretcher cauldron - fire  Part-Whole: scissors - blade monument - granite

14 Semantic Relations  Defined a total of 34 relations  Focused on defining concrete, graphically depictable relationships  “Generic” relations accounted for most of the labeled pairs (e.g. containing.r, next-to.r)  Finer distinctions were made within these generic semantic relations (e.g. habitat.r, residence.r within the overarching containing.r relation)  Defined a total of 34 relations  Focused on defining concrete, graphically depictable relationships  “Generic” relations accounted for most of the labeled pairs (e.g. containing.r, next-to.r)  Finer distinctions were made within these generic semantic relations (e.g. habitat.r, residence.r within the overarching containing.r relation)

15 Example Semantic Relations  Locations: mural - gallery - containing.r lizard - desert - habitat.r  Nearby Objects: ambulance - stretcher - next-to.r cauldron - fire - above.r  Part-Whole: scissors - blade - object-part.r monument - granite - stuff-object.r  Locations: mural - gallery - containing.r lizard - desert - habitat.r  Nearby Objects: ambulance - stretcher - next-to.r cauldron - fire - above.r  Part-Whole: scissors - blade - object-part.r monument - granite - stuff-object.r

16 Semantic Relations within Locations Task  We collected 6850 locations for 342 target objects from our 3D library. Relation Number of occurrences Percentage of total scored pairs containing.r % habitat.r % on-surface.r % geographical- location.r % group.r %

17 Semantic Relations within Nearby Objects Task  We collected 6850 nearby objects for 342 target objects from our 3D library. Relation Number of occurrences Percentage of total scored pairs next-to.r % on-surface.r % containing.r % habitat.r % object-part.r %

18 Semantic Relations within Part- Whole Task  We collected 3500 parts of 245 objects. Relation Number of occurrences Percentage of total scored pairs object-part.r % stuff-object.r % containing.r501.48% habitat.r361.06% stuff-mass.r170.5%

19 Automatic Normalization Techniques  Collected AMT data was classified into higher- scoring versus lower-scoring sets by:  Log-likelihood and log-odds of sentential co- occurrences in the Gigaword English corpus  WordNet path similarity  Resnik similarity  WordNet average pair-wise similarity  WordNet matrix similarity  Accuracy evaluated by comparison against manually normalized data  Collected AMT data was classified into higher- scoring versus lower-scoring sets by:  Log-likelihood and log-odds of sentential co- occurrences in the Gigaword English corpus  WordNet path similarity  Resnik similarity  WordNet average pair-wise similarity  WordNet matrix similarity  Accuracy evaluated by comparison against manually normalized data

20 Precision & Recall  AMT data is quite cheap to collect, so we were concerned predominantly with precision (obtaining highly accurate data) rather than recall (avoiding loss of some data).  In order to achieve more accurate data (high precision), we will lose a portion of our AMT data (low recall)  AMT data is quite cheap to collect, so we were concerned predominantly with precision (obtaining highly accurate data) rather than recall (avoiding loss of some data).  In order to achieve more accurate data (high precision), we will lose a portion of our AMT data (low recall)

21 Locations Task  Achieved best precision with log-odds.  Within high-scoring set, responses that were too general (e.g. turntable - store) were rejected.  Within low-scoring set, extremely specific locations that were unlikely to occur within a corpus or WordNet’s synsets were approved (e.g. caliper - architect’s briefcase)  Achieved best precision with log-odds.  Within high-scoring set, responses that were too general (e.g. turntable - store) were rejected.  Within low-scoring set, extremely specific locations that were unlikely to occur within a corpus or WordNet’s synsets were approved (e.g. caliper - architect’s briefcase) Base- line Log- likel. Log- odds WN Path Sim. Resnik WN Avg. PW WN Matrix Sim. Precision Recall

22 Nearby Objects Task  Relatively few target-response pairs were discarded, resulting in high recall.  High precision due to open-ended nature of task; responses often fell under a relation, if not next-to.r.  Relatively few target-response pairs were discarded, resulting in high recall.  High precision due to open-ended nature of task; responses often fell under a relation, if not next-to.r. Base- line Log- likel. Log- odds WN Path Sim. Resnik WN Avg. PW WN Matrix Sim. Precision Recall

23 Part-Whole Task  Rejected target-response pairs from the high-scoring set were often due to responses that named attributes, rather than parts, of the target item (e.g. croissant - flaky)  Approved pairs from the low-scoring set were mainly due to obvious, “common sense” responses that would usually be inferred, not explicitly stated (e.g. bunny - brain)  Rejected target-response pairs from the high-scoring set were often due to responses that named attributes, rather than parts, of the target item (e.g. croissant - flaky)  Approved pairs from the low-scoring set were mainly due to obvious, “common sense” responses that would usually be inferred, not explicitly stated (e.g. bunny - brain) Base- line Log- likel. Log- odds WN Path Sim. Resnik WN Avg. PW WN Matrix Sim. Precision Recall

24 Future Automatic Normalization Techniques  Computing word association measures on much larger corpora (e.g. Google’s 1 trillion word corpus)  WordNet synonyms and hypernyms  Latent Semantic Analysis to build word similarity matrices  Computing word association measures on much larger corpora (e.g. Google’s 1 trillion word corpus)  WordNet synonyms and hypernyms  Latent Semantic Analysis to build word similarity matrices

25 In Summary…  WordsEye & Scenario-Based Lexical Knowledge Resource (SBLR)  Amazon’s Mechanical Turk & our tasks  Manual normalization of AMT data  Automatic normalization techniques used on AMT data and results  Possible future automatic normalization methods  WordsEye & Scenario-Based Lexical Knowledge Resource (SBLR)  Amazon’s Mechanical Turk & our tasks  Manual normalization of AMT data  Automatic normalization techniques used on AMT data and results  Possible future automatic normalization methods

26 Thanks to…  Richard Sproat  Masoud Rouhizadeh  All the CSLU interns  Richard Sproat  Masoud Rouhizadeh  All the CSLU interns

27 Questions?