Overview of the KBP 2012 Slot-Filling Tasks Hoa Trang Dang (National Institute of Standards and Technology Javier Artiles (Rakuten Institute of Technology)

Slides:



Advertisements
Similar presentations
Search in Source Code Based on Identifying Popular Fragments Eduard Kuric and Mária Bieliková Faculty of Informatics and Information.
Advertisements

Information Extraction Lecture 4 – Named Entity Recognition II CIS, LMU München Winter Semester Dr. Alexander Fraser, CIS.
Knowledge Base Completion via Search-Based Question Answering
Overview of the TAC2013 Knowledge Base Population Evaluation: Temporal Slot Filling Mihai Surdeanu with a lot help from: Hoa Dang, Joe Ellis, Heng Ji,
Text Analysis Conference Knowledge Base Population 2013 Hoa Trang Dang National Institute of Standards and Technology Sponsored by:
Overview of the TAC2013 Knowledge Base Population Evaluation: English Slot Filling Mihai Surdeanu with a lot help from: Hoa Dang, Joe Ellis, Heng Ji, and.
Large-Scale Entity-Based Online Social Network Profile Linkage.
Distant Supervision for Knowledge Base Population Mihai Surdeanu, David McClosky, John Bauer, Julie Tibshirani, Angel Chang, Valentin Spitkovsky, Christopher.
Beyond TREC-QA Ling573 NLP Systems and Applications May 28, 2013.
Tri-lingual EDL Planning Heng Ji (RPI) Hoa Trang Dang (NIST) WORRY, BE HAPPY!
Overview of the KBP 2013 Slot Filler Validation Track Hoa Trang Dang National Institute of Standards and Technology.
NYU ANLP-00 1 Automatic Discovery of Scenario-Level Patterns for Information Extraction Roman Yangarber Ralph Grishman Pasi Tapanainen Silja Huttunen.
Linguistic Resources for the 2013 TAC KBP Sentiment SF Evaluation Joe Ellis (presenter), Jeremy Getman, Jonathan Wright, Stephanie Strassel Linguistic.
Linguistic Resources for the 2013 TAC KBP Slot Filling Evaluations Joe Ellis (presenter), Jeremy Getman, Jonathan Wright, Stephanie Strassel Linguistic.
Explorations in Tag Suggestion and Query Expansion Jian Wang and Brian D. Davison Lehigh University, USA SSM 2008 (Workshop on Search in Social Media)
Ang Sun Ralph Grishman Wei Xu Bonan Min November 15, 2011 TAC 2011 Workshop Gaithersburg, Maryland USA.
NaLIX: A Generic Natural Language Search Environment for XML Data Presented by: Erik Mathisen 02/12/2008.
Event Extraction: Learning from Corpora Prepared by Ralph Grishman Based on research and slides by Roman Yangarber NYU.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Yin Yang (Hong Kong University of Science and Technology) Nilesh Bansal (University of Toronto) Wisam Dakka (Google) Panagiotis Ipeirotis (New York University)
Title Extraction from Bodies of HTML Documents and its Application to Web Page Retrieval Microsoft Research Asia Yunhua Hu, Guomao Xin, Ruihua Song, Guoping.
TREC 2009 Review Lanbo Zhang. 7 tracks Web track Relevance Feedback track (RF) Entity track Blog track Legal track Million Query track (MQ) Chemical IR.
Reyyan Yeniterzi Weakly-Supervised Discovery of Named Entities Using Web Search Queries Marius Pasca Google CIKM 2007.
A Two Tier Framework for Context-Aware Service Organization & Discovery Wei Zhang 1, Jian Su 2, Bin Chen 2,WentingWang 2, Zhiqiang Toh 2, Yanchuan Sim.
Outline Quick review of GS Current problems with GS Our solutions Future work Discussion …
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
1 Technologies for (semi-) automatic metadata creation Diana Maynard.
CROSSMARC Web Pages Collection: Crawling and Spidering Components Vangelis Karkaletsis Institute of Informatics & Telecommunications NCSR “Demokritos”
Data Mining Chapter 1 Introduction -- Basic Data Mining Tasks -- Related Concepts -- Data Mining Techniques.
Structured Use of External Knowledge for Event-based Open Domain Question Answering Hui Yang, Tat-Seng Chua, Shuguang Wang, Chun-Keat Koh National University.
Enhanced Infrastructure for Creation & Collection of Translation Resources Zhiyi Song, Stephanie Strassel (speaker), Gary Krug, Kazuaki Maeda.
Towards Natural Question-Guided Search Alexander Kotov ChengXiang Zhai University of Illinois at Urbana-Champaign.
Wei Xu, Ralph Grishman, Le Zhao (CMU) New York University Novmember 24, 2011.
21/11/2002 The Integration of Lexical Knowledge and External Resources for QA Hui YANG, Tat-Seng Chua Pris, School of Computing.
1 Automating Slot Filling Validation to Assist Human Assessment Suzanne Tamang and Heng Ji Computer Science Department and Linguistics Department, Queens.
Linguistic Resources for the 2013 TAC KBP Entity Linking Evaluation Joe Ellis (presenter), Justin Mott, Xuansong Li, Jeremy Getman, Jonathan Wright, Stephanie.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
PRIS at Slot Filling in KBP 2012: An Enhanced Adaboost Pattern-Matching System Yan Li Beijing University of Posts and Telecommunications
Ang Sun Director of Research, Principal Scientist, inome
LOGO 1 Corroborate and Learn Facts from the Web Advisor : Dr. Koh Jia-Ling Speaker : Tu Yi-Lang Date : Shubin Zhao, Jonathan Betz (KDD '07 )
LING 573 Deliverable 3 Jonggun Park Haotian He Maria Antoniak Ron Lockwood.
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
An Iterative Approach to Extract Dictionaries from Wikipedia for Under-resourced Languages G. Rohit Bharadwaj Niket Tandon Vasudeva Varma Search and Information.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
August 17, 2005Question Answering Passage Retrieval Using Dependency Parsing 1/28 Question Answering Passage Retrieval Using Dependency Parsing Hang Cui.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
RESEARCH POSTER PRESENTATION DESIGN © Triggers in Extraction 5. Experiments Data Development set: KBP SF 2012 corpus.
Information Retrieval
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Linguistic Resources for the 2013 TAC KBP Cold Start Evaluation Joe Ellis (presenter), Jeremy Getman, Jonathan Wright, Stephanie Strassel Linguistic Data.
4. Relationship Extraction Part 4 of Information Extraction Sunita Sarawagi 9/7/2012CS 652, Peter Lindes1.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
AQUAINT AQUAINT Evaluation Overview Ellen M. Voorhees.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Linguistic Resources for the 2013 TAC KBP Temporal SF Evaluation Joe Ellis (presenter), Jeremy Getman, Jonathan Wright, Stephanie Strassel Linguistic Data.
Cold-Start KBP Something from Nothing Sean Monahan, Dean Carpenter Language Computer.
MSM 2013 Challenge: Annotowatch Stefan Dlugolinsky, Peter Krammer, Marek Ciglan, Michal Laclavik Institute of Informatics, Slovak Academy of Sciences.
WePS2 Attribute Extraction Task Sekine and Artiles WWW 2009 Workshop.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Automatically Labeled Data Generation for Large Scale Event Extraction
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Introduction Task: extracting relational facts from text
Reading Report on Question Answering
CS246: Information Retrieval
Topic: Semantic Text Mining
Presentation transcript:

Overview of the KBP 2012 Slot-Filling Tasks Hoa Trang Dang (National Institute of Standards and Technology Javier Artiles (Rakuten Institute of Technology) James Mayfield (Johns Hopkins University) Joe Ellis, Xuansong Li, Kira Griffitt, Stephanie Strassel, Jonathan Wright (Linguistic Data Consortium)

Slot-filling Tasks Goal: Augment a reference knowledge base (KB) with info about target entities as found in a diverse collection of documents Reference KB: Oct 2008 Wikipedia snapshot. Each KB node corresponds to a Wikipedia and contains: ▫Infobox ▫Wiki_text (free text not in infobox) English source documents: ▫2.3 M news docs (1.2 M docs in 2011) ▫1.5 M Web and other docs (0.5 M docs in 2011) [Spanish source documents] Diagnostic task: Slot Filler Validation

Slots derived from Wikipedia infobox PersonOrganization per:alternate_namesper:member_oforg:alternate_names per:date_of_birthper:employee_oforg:political_religious_affiliation per:ageper:religionorg:top_members_employees per:country_of_birthper:spouseorg:number_of_employees per:stateorprovince_of_birthper:childrenorg:members per:city_of_birthper:parentsorg:member_of per:date_of_deathper:siblingsorg:subsidiaries per:country_of_deathper:other_familyorg:parents per:stateorprovince_of_deathper:chargesorg:founded_by per:city_of_deathorg:date_founded per:cause_of_deathorg:date_dissolved per:countries_of_residenceorg:country_of_headquarters per:statesorprovinces_of_residenceorg:stateorprovince_of_headquarters per:cities_of_residenceorg:city_of_headquarters per:schools_attendedorg:shareholders per:titleorg:website

Slot-Filling Task Requirements Task: given target entity and predefined slots for each entity type (PER, ORG), return all new slot fillers for that entity that can be found in the source documents, and a supporting document for each filler Non-redundant ▫Don’t return a slot filler if it’s already in the KB ▫Don’t return more than one instance of a slot filler Exact boundaries of filler string, as found in supporting document ▫Text is complete (e.g., “John Doe” rather than “John”) ▫No extraneous text (e.g., “John Doe” rather than “John Doe’s house” Evaluation based on TREC-QA pooling methodology, combine ▫Candidate slot fillers from non-exhaustive manual search ▫Candidate slot fillers from fully automatic systems Answer “key” is incomplete, coverage depends on number, quality, and diversity of contributing systems.

Differences from KBP 2011 Offsets provided for target entity mention in query Increased number of submissions (up to 5) Require normalization of slot fillers that are dates (“yesterday” -> “ ”) Request each proposed slot filler to include ▫A confidence value ▫Offsets for justification (usually a sentence) ▫Offsets for the raw (unnormalized) slot filler in the document Move toward more precise justifications ▫Improved usability (for humans) in end applications ▫Improved training data for systems Offsets and confidence values did not affect official scores ▫But confidence values were used to rank and truncate extremely lengthy submissions

Slot-Filling Evaluation Pool responses from submitted runs and from manual search -> ▫Set of [docid, answer-string] pairs for each target entity and slot Assessment: ▫Each pair judged as one of correct, redundant, inexact, or wrong (credit given only for correct responses) ▫Correct pairs grouped into equivalence classes (entities); each single- valued slot has at most one equivalence class for a given target entity Scoring: ▫Recall: number of correct equivalence classes returned / number of known equivalence classes ▫Precision: number of correct equivalence classes returned / number of [docid, answer-string] pairs returned ▫F1 = (P*R)/(R+P)

Slot Filling Participants TeamOrganization ADVIS_UIC* University of Illinois at Chicago GDUFS* Guangdong University of Foreign Affairs IIRGUniversity College Dublin lsvSaarland University NLPCompThe Hong Kong Polytechnic University NYUNew York University papelo* NEC Laboratories America PRISBeijing University of Posts and Telecommunications Siel_12International Institute of Information Technology, Hyderabad sweat2012* Chinese Academy of Sciences TALP_UPC* Technical University of Catalonia, UPC * first-time slot-filling team

Top 6 KBP 2012 Slot-Filling teams

Top 4 KBP 2012 Slot-Filling teams

Slot-Filling Approaches IIRG: (+ling, -ML) ▫Stanford CoreNLP for POS, NER, parse. ▫Sentence retrieval by exact match with named mention of target entity. ▫Rule-based pattern matching and keyword matching to identify slot fillers. lsv: (-ling, +ML) ▫Shallow approach – no parse or coref ▫Query expansion via Wikipedia redirect links ▫SVM and Freebase for distant supervision NYU: (+ling, +ML) ▫POS, parse, NER, time expression tagging, coref ▫Query expansion via small set of handcrafted rules, Wikipedia redirect links ▫MaxEnt and Freebase for distant supervision. ▫Combination of: hand-coded rules, patterns generated by bootstrapping and then manually reviewed, and classifier trained by distant supervision PRIS: (+ling, +ML) ▫Stanford CoreNLP for POS, NER, SUTime, parse, coref; ▫Query expansion via small set of handcrafted rules, coref’d names; ▫Adaboost for finding new extraction patterns (word sequence patterns and dependency path patterns)

Distribution of slots in answer key

Slot productivity per:title14%per:title21%per:title14% org:top_members/e mployees 12%org:top_members/em ployees 12%org:top_members_ employees 11% per:employee_of7%org:alternate_names10%per:member_of6% org:alternate_names5%per:employee_of7%per:children6% org:subsidiaries4%per:member_of5%org:alternate_names6% per:member_of4%per:alternate_names5%per:employee_of4% per:cities_of_ residence 4%org:subsidiaries3%per:cities_of_ residence 4%

Slot filler Validation (SFV) Goals ▫Improve precision of full slot-filling systems (without reducing recall) ▫Allow teams without a full slot-filling system to participate, focus on answer validation rather than document retrieval SFV input: ▫All input to slot-filling task ▫Submission files from all Slot Filling runs, containing candidate slot fillers ▫No information about “past performance” of each slot filling system SFV output: ▫Binary classification (Correct / Incorrect) of each candidate slot filler Evaluation: ▫Filter out “Incorrect” slot fillers from each run, and score; compare to score for original run Submissions: 1 team (Blender_CUNY)

Filtering candidate slot fillers

Answer Justification Goals ▫Improve training data for systems – narrow down location of answer patterns ▫Reduce assessment effort (for correct answers with correct justifications) ▫Improve usability (for humans) in end applications Task guidelines: ▫For each slot filler, provide start and end offsets for the sentence or clause that provides justification for the relation. For example, for query per:spouse of “Michelle Obama” and the sentence “He is married to Michelle Obama” (“He” referring to Barack Obama mentioned earlier in the document), the filler … should be “Barack Obama”, the offsets for filler must point to “He” and the offsets for justification must point to “He is married to Michelle Obama”. Slight mismatch with LDC assessment guidelines (require antecedent of relevant pronouns in justification, otherwise judged as inexact) ▫Need additional discussion/refinement of guidelines

LDC Data, Annotation, and Assessment