Liberal Event Extraction and Event Schema Induction

Slides:



Advertisements
Similar presentations
Machine Learning Approaches to the Analysis of Large Corpora : A Survey Xunlei Rose Hu and Eric Atwell University of Leeds.
Advertisements

Progress update Lin Ziheng. System overview 2 Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self.
NYU ANLP-00 1 Automatic Discovery of Scenario-Level Patterns for Information Extraction Roman Yangarber Ralph Grishman Pasi Tapanainen Silja Huttunen.
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
Automatic Metaphor Interpretation as a Paraphrasing Task Ekaterina Shutova Computer Lab, University of Cambridge NAACL 2010.
CS Word Sense Disambiguation. 2 Overview A problem for semantic attachment approaches: what happens when a given lexeme has multiple ‘meanings’?
Event Extraction: Learning from Corpora Prepared by Ralph Grishman Based on research and slides by Roman Yangarber NYU.
July 9, 2003ACL An Improved Pattern Model for Automatic IE Pattern Acquisition Kiyoshi Sudo Satoshi Sekine Ralph Grishman New York University.
4/14/20051 ACE Annotation Ralph Grishman New York University.
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
Word sense induction using continuous vector space models
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Ontology Learning and Population from Text: Algorithms, Evaluation and Applications Chapters Presented by Sole.
Extracting Opinions, Opinion Holders, and Topics Expressed in Online News Media Text Soo-Min Kim and Eduard Hovy USC Information Sciences Institute 4676.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Learning Narrative Schemas Nate Chambers, Dan Jurafsky Stanford University IBM Watson Research Center Visit.
Learning Information Extraction Patterns Using WordNet Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield,
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Automatic Lexical Annotation Applied to the SCARLET Ontology Matcher Laura Po and Sonia Bergamaschi DII, University of Modena and Reggio Emilia, Italy.
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Annotating Words using WordNet Semantic Glosses Julian Szymański Department of Computer Systems Architecture, Faculty of Electronics, Telecommunications.
SYMPOSIUM ON SEMANTICS IN SYSTEMS FOR TEXT PROCESSING September 22-24, Venice, Italy Combining Knowledge-based Methods and Supervised Learning for.
Modelling Human Thematic Fit Judgments IGK Colloquium 3/2/2005 Ulrike Padó.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
A Semantic Approach to IE Pattern Induction Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield, UK.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Minimally Supervised Event Causality Identification Quang Do, Yee Seng, and Dan Roth University of Illinois at Urbana-Champaign 1 EMNLP-2011.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
Distribution of information in biomedical abstracts and full- text publications M. J. Schuemie et al. Dept. of Medical Informatics, Erasmus University.
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
Named Entity Disambiguation on an Ontology Enriched by Wikipedia Hien Thanh Nguyen 1, Tru Hoang Cao 2 1 Ton Duc Thang University, Vietnam 2 Ho Chi Minh.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Supertagging CMSC Natural Language Processing January 31, 2006.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
SALSA-WS 09/05 Approximating Textual Entailment with LFG and FrameNet Frames Aljoscha Burchardt, Anette Frank Computational Linguistics Department Saarland.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Second Language Learning From News Websites Word Sense Disambiguation using Word Embeddings.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
A Database of Narrative Schemas A 2010 paper by Nathaniel Chambers and Dan Jurafsky Presentation by Julia Kelly.
RELATION EXTRACTION, SYMBOLIC SEMANTICS, DISTRIBUTIONAL SEMANTICS Heng Ji Oct13, 2015 Acknowledgement: distributional semantics slides from.
Identifying Expressions of Opinion in Context Eric Breck and Yejin Choi and Claire Cardie IJCAI 2007.
Trends in NL Analysis Jim Critz University of New York in Prague EurOpen.CZ 12 December 2008.
Unsupervised Sparse Vector Densification for Short Text Similarity
Automatically Labeled Data Generation for Large Scale Event Extraction
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
Linguistic Graph Similarity for News Sentence Searching
Concept Grounding to Multiple Knowledge Bases via Indirect Supervision
Deep Learning for Bacteria Event Identification
Web News Sentence Searching Using Linguistic Graph Similarity
On Dataless Hierarchical Text Classification
Relation Extraction CSCI-GA.2591
NYU Coreference CSCI-GA.2591 Ralph Grishman.
Wei Wei, PhD, Zhanglong Ji, PhD, Lucila Ohno-Machado, MD, PhD
Two Discourse Driven Language Models for Semantics
(Entity and) Event Extraction CSCI-GA.2591
Improving a Pipeline Architecture for Shallow Discourse Parsing
Vector-Space (Distributional) Lexical Semantics
HowToKB: Mining HowTo Knowledge from Online Communities
Category-Based Pseudowords
Lei Sha, Jing Liu, Chin-Yew Lin, Sujian Li, Baobao Chang, Zhifang Sui
Introduction Task: extracting relational facts from text
Automatic Detection of Causal Relations for Question Answering
Text-to-Text Generation
Effective Entity Recognition and Typing by Relation Phrase-Based Clustering
Word embeddings (continued)
Rachit Saluja 03/20/2019 Relation Extraction with Matrix Factorization and Universal Schemas Sebastian Riedel, Limin Yao, Andrew.
Unsupervised Learning of Narrative Schemas and their Participants
Embedding based entity summarization
Presenter: Yu Chen Computer Science Department
Presentation transcript:

Liberal Event Extraction and Event Schema Induction Lifu Huang1, Taylor Cassidy2, Xiaocheng Feng3, Heng Ji1, Clare R. Voss2, Jiawei Han4, Avirup Sil5 1 Rensselaer Polytechnic Institute, 2 US Army Research Lab, 3 Harbin Institute of Technology, 4 University of Illinois at Urbana-Champaign, 5 IBM T.J. Watson Research Center

Task Definition Event Mention: a string of words in text denoting a particular event; Event Trigger: the word in the event mention that holds the bulk of its semantic content; Event Argument: a concept that serves as a participant or attribute with a specific role in an event mention; Argument Role: the function or purpose of the argument with respect to the corresponding event; e.g.:

Comparison Traditional Event Extraction vs. Liberal Event Extraction

Motivating Examples - Top 8 most similar words based on lexical embedding

Hypothesis 1 Event triggers that occur in similar contexts and share the same sense tend to have similar types. WSD (word sense disambiguation) to learn a distinct embedding for each sense;

Motivating Examples

Hypothesis 2 Beyond the lexical semantics of a particular event trigger, its type is also dependent on its arguments and their roles, as well as other words contextually connected to the trigger; Event Structure;

Approach Overview Hypothesis 2 Hypothesis 1

Identification Trigger Identification: Arguments Identification: AMR Parsing: map concepts to OntoNotes; FrameNet: war, theft, pickpocket; Arguments Identification: AMR Parsing: Most arguments can be captured by rich semantic parsing.

Identification Trigger Identification: Arguments Identification: AMR Parsing: map concepts to OntoNotes; FrameNet: war, theft, pickpocket; Arguments Identification: AMR Parsing: Most arguments can be captured by rich semantic parsing.

Representation WSD based Lexical Embedding: Preprocess: IMS (Zhong and Ng. 2010) for word sense disambiguation; Skip-gram Word2Vec (Mikolov et al., 2013); Top 10 most similar words: Fire-1 Score firing cannon grenades grenade gun arm explosive point-blank 0.829 0.774 0.767 0.760 0.757 0.755 0.742 0.740 Fire-2 Score rehired hire-1 resign-1 rehire sacked quit-1 sack-1 quits 0.790 0.626 0.618 0.596 0.591 0.565 0.563

Representation Event Structure Representation: Recursive Neural Tensor AutoEncoder

Joint Constraint Clustering Joint Optimized Trigger and Argument Clusters: Spectral Clustering; Internal Measures: cohesion (Dintra) and separation(Dinter);

Naming Trigger Type Naming: Argument Role Naming: The trigger which is nearest to the centroid; Argument Role Naming: Borrow from existing linguistic resource: map AMR roles to the roles in FrameNet, VerbNet, PropBank;

Naming Trigger Type Naming: Argument Role Naming: The trigger which is nearest to the centroid; Argument Role Naming:

Experiments and Evaluation Data Set: ACE (Automatic Content Extraction); ERE (Entity Relation Event);

Experiments and Evaluation Schema Discovery: Event Types and Arguments:

Experiments and Evaluation Schema Discovery:

Experiments and Evaluation Schema Discovery:

Experiments and Evaluation Schema Discovery: Coverage Comparison with ACE and ERE: New Event types and Arguments: e.g., Arguments for Attack Event: Attacker, Target, Instrument, Time, Place The Dutch government, facing strong public anti-war pressure, said it would not commit fighting forces to the war against Iraq but added it supported the military campaign to disarm Saddam. Attack New Argument: Purpose

Experiments and Evaluation Event Extraction for All Types: Fully annotate 100 sentences with inter-annotator agreement: 83% for triggers and 79% for arguments; Missing Triggers and Arguments:

Experiments and Evaluation Event Extraction for All Types: Fully annotate 100 sentences with inter-annotator agreement: 83% for triggers and 79% for arguments; Missing Triggers and Arguments: Multi-word expressions (e.g., took office) or not verb or noun concepts:

Experiments and Evaluation Event Extraction for All Types: Fully annotate 100 sentences with inter-annotator agreement: 83% for triggers and 79% for arguments; Missing Triggers and Arguments: Multi-word expressions (e.g., took office) or not verb or noun concepts: e.g.: As well as previously holding senior positions at Barclays Bank, BZW and Kleinwort Benson, McCarthy was formerly a top civil servant at the Department of Trade and Industry.

Experiments and Evaluation Event Extraction for All Types: Fully annotate 100 sentences with inter-annotator agreement: 83% for triggers and 79% for arguments; Missing Triggers and Arguments: Multi-word expressions (e.g., took office) or not verb or noun concepts Arguments that are not direct semantic related with event triggers:

Experiments and Evaluation Event Extraction for All Types: Fully annotate 100 sentences with inter-annotator agreement: 83% for triggers and 79% for arguments; Missing Triggers and Arguments: Multi-word expressions (e.g., took office) or not verb or noun concepts Arguments that are not direct semantic related with event triggers: e.g., : Anti-corruption judge Saul Pena stated Montesinos has admitted to the abuse of authority charge.

Experiments and Evaluation Impact of Representations: AMR vs. Dependency: AMR: Instrument Dependency Parsing: compound modifier e.g., Approximately 25 kilometers southwest of Sringar 2 militants were killed in a second gun battle.

Experiments and Evaluation Event Extraction on ACE and ERE Types Manually asses triggers to ACE/ERE; Comparison Supervised Methods: DMCNN: A dynamic multi-pooling convolutional neural network based on distributed word representations (Chen et al., 2015); Joint IE: A Structured perceptron model based on symbolic semantic features (Li et al., 2013); LSTM: A long short-term memory neural network (Hochreiter and Schmidhuber, 1997) based on distributed semantic features;

Experiments and Evaluation Event Extraction on ACE and ERE Types Manually asses triggers to ACE/ERE; Comparison Supervised Methods:

Experiments and Evaluation Event Extraction on ACE and ERE Types Manually asses triggers to ACE/ERE; Comparison Supervised Methods: Supervised methods heavily rely on the quality and quantity of the training data: ERE training documents contain 1,068 events and 2,448 arguments, while ACE training documents contain more than 4,700 events and 9,700 arguments;

Experiments and Evaluation Event Extraction on Biomedical Data set: 14 biomedical articles (755 sentences) with perfect AMR annotations; Randomly sample 100 sentences and manually asses the correctness of each event and argument; 83.1% precision on trigger labeling (619 events); 78.4% precision on argument labeling (1,124 arguments);

Experiments and Evaluation Event Extraction on Biomedical Data set: 14 biomedical articles (755 sentences) with perfect AMR annotations; Randomly sample 100 sentences and manually asses the correctness of each event and argument;

Questions and Comments? Thanks!