SRL using complete syntactic analysis Mihai Surdeanu and Jordi Turmo TALP Research Center Universitat Politècnica de Catalunya.

Slides:



Advertisements
Similar presentations
SEMANTIC ROLE LABELING BY TAGGING SYNTACTIC CHUNKS
Advertisements

Progress update Lin Ziheng. System overview 2 Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self.
CONSTRAINED CONDITIONAL MODELS TUTORIAL Jingyu Chen, Xiao Cheng.
Page 1 SRL via Generalized Inference Vasin Punyakanok, Dan Roth, Wen-tau Yih, Dav Zimak, Yuancheng Tu Department of Computer Science University of Illinois.
A Structured Model for Joint Learning of Argument Roles and Predicate Senses Yotaro Watanabe Masayuki Asahara Yuji Matsumoto ACL 2010 Uppsala, Sweden July.
Recovering Human Body Configurations: Combining Segmentation and Recognition Greg Mori, Xiaofeng Ren, and Jitentendra Malik (UC Berkeley) Alexei A. Efros.
E XTRACTING SEMANTIC ROLE INFORMATION FROM UNSTRUCTURED TEXTS Diana Trandab ă 1 and Alexandru Trandab ă 2 1 Faculty of Computer Science, University “Al.
Max-Margin Matching for Semantic Role Labeling David Vickrey James Connor Daphne Koller Stanford University.
计算机科学与技术学院 Chinese Semantic Role Labeling with Dependency-driven Constituent Parse Tree Structure Hongling Wang, Bukang Wang Guodong Zhou NLP Lab, School.
Dependency Parsing Some slides are based on:
Semantic Role Labeling Abdul-Lateef Yussiff
A Joint Model For Semantic Role Labeling Aria Haghighi, Kristina Toutanova, Christopher D. Manning Computer Science Department Stanford University.
Semantic Role Chunking Combining Complementary Syntactic Views Sameer Pradhan, Kadri Hacioglu, Wayne Ward, James H. Martin, Daniel Jurafsky  Center for.
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
Reranking Parse Trees with a SRL system Charles Sutton and Andrew McCallum University of Massachusetts June 30, 2005.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
Ch 10 Part-of-Speech Tagging Edited from: L. Venkata Subramaniam February 28, 2002.
Two-Phase Semantic Role Labeling based on Support Vector Machines Kyung-Mi Park Young-Sook Hwang Hae-Chang Rim NLP Lab. Korea Univ.
Page 1 Generalized Inference with Multiple Semantic Role Labeling Systems Peter Koomen, Vasin Punyakanok, Dan Roth, (Scott) Wen-tau Yih Department of Computer.
Boosting Applied to Tagging and PP Attachment By Aviad Barzilai.
A transformation-based approach to argument labeling Derrick Higgins Educational Testing Service
Dept. of Computer Science & Engg. Indian Institute of Technology Kharagpur Part-of-Speech Tagging and Chunking with Maximum Entropy Model Sandipan Dandapat.
A Memory-Based Approach to Semantic Role Labeling Beata Kouchnir Tübingen University 05/07/04.
A Framework for Named Entity Recognition in the Open Domain Richard Evans Research Group in Computational Linguistics University of Wolverhampton UK
Semantic Role Labeling using Maximum Entropy Model Joon-Ho Lim NLP Lab. Korea Univ.
SI485i : NLP Set 9 Advanced PCFGs Some slides from Chris Manning.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
STRUCTURED PERCEPTRON Alice Lai and Shi Zhi. Presentation Outline Introduction to Structured Perceptron ILP-CRF Model Averaged Perceptron Latent Variable.
3.0 State Space Representation of Problems 3.1 Graphs 3.2 Formulating Search Problems 3.3 The 8-Puzzle as an example 3.4 State Space Representation using.
Some Advances in Transformation-Based Part of Speech Tagging
1 Trees Tree nomenclature Implementation strategies Traversals –Depth-first –Breadth-first Implementing binary search trees.
Researcher affiliation extraction from homepages I. Nagy, R. Farkas, M. Jelasity University of Szeged, Hungary.
Illinois-Coref: The UI System in the CoNLL-2012 Shared Task Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Mark Sammons, and Dan Roth Supported by ARL,
Semantic Role Labeling for Arabic using Kernel Methods Mona Diab Alessandro Moschitti Daniele Pighin.
AQUAINT Workshop – June 2003 Improved Semantic Role Parsing Kadri Hacioglu, Sameer Pradhan, Valerie Krugler, Steven Bethard, Ashley Thornton, Wayne Ward,
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
INSTITUTE OF COMPUTING TECHNOLOGY Forest-based Semantic Role Labeling Hao Xiong, Haitao Mi, Yang Liu and Qun Liu Institute of Computing Technology Academy.
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
Multi-Relational Data Mining: An Introduction Joe Paulowskey.
Prototype-Driven Learning for Sequence Models Aria Haghighi and Dan Klein University of California Berkeley Slides prepared by Andrew Carlson for the Semi-
Combining Lexical Resources: Mapping Between PropBank and VerbNet Edward Loper,Szu-ting Yi, Martha Palmer September 2006.
Quick Speech Synthesis CMSC Natural Language Processing April 29, 2003.
Emotions from text: machine learning for text-based emotion prediction Cecilia Alm, Dan Roth, Richard Sproat UIUC, Illinois HLT/EMPNLP 2005.
Supertagging CMSC Natural Language Processing January 31, 2006.
School of Computer Science 1 Information Extraction with HMM Structures Learned by Stochastic Optimization Dayne Freitag and Andrew McCallum Presented.
Problem Reduction So far we have considered search strategies for OR graph. In OR graph, several arcs indicate a variety of ways in which the original.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
NLP. Introduction to NLP Last week, Min broke the window with a hammer. The window was broken with a hammer by Min last week With a hammer, Min broke.
1 Holistic Twig Joins: Optimal XML Pattern Matching Nicolas Bruno, Nick Koudas, Divesh Srivastava ACM SIGMOD 2002 Presented by Jun-Ki Min.
Department of Computer Science The University of Texas at Austin USA Joint Entity and Relation Extraction using Card-Pyramid Parsing Rohit J. Kate Raymond.
Chapter 6 – Trees. Notice that in a tree, there is exactly one path from the root to each node.
Julia Hockenmaier and Mark Steedman.   The currently best single-model statistical parser (Charniak, 1999) achieves Parseval scores of over 89% on the.
Dependency Parsing Niranjan Balasubramanian March 24 th 2016 Credits: Many slides from: Michael Collins, Mausam, Chris Manning, COLNG 2014 Dependency Parsing.
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Relation Extraction (RE) via Supervised Classification See: Jurafsky & Martin SLP book, Chapter 22 Exploring Various Knowledge in Relation Extraction.
Semantic Role Labelling Using Chunk Sequences Ulrike Baldewein Katrin Erk Sebastian Padó Saarland University Saarbrücken Detlef Prescher Amsterdam University.
Language Identification and Part-of-Speech Tagging
COSC 6336: Natural Language Processing
Decision Trees DEFINITION: DECISION TREE A decision tree is a tree in which the internal nodes represent actions, the arcs represent outcomes of an action,
CSC 594 Topics in AI – Natural Language Processing
The Resolution of Speculation and Negation
INAGO Project Automatic Knowledge Base Generation from Text for Interactive Question Answering.
Bidirectional CRF for NER
Improving a Pipeline Architecture for Shallow Discourse Parsing
Compilers Principles, Techniques, & Tools Taught by Jing Zhang
Heuristic Search Methods
Parsing I: CFGs & the Earley Parser
Johns Hopkins 2003 Summer Workshop on Syntax and Statistical Machine Translation Chapters 5-8 Ethan Phelps-Goodman.
Progress report on Semantic Role Labeling
Presentation transcript:

SRL using complete syntactic analysis Mihai Surdeanu and Jordi Turmo TALP Research Center Universitat Politècnica de Catalunya

Analysis of argument to constituent mappings One-to-one (shared parent)(different parents) Train 96.06% 2.49% 1.45% Dev 91.36% 4.83% 3.81% One-to-many

The approach One-to-one (shared parent)(different parents) Model learnedSimple heuristics One-to-many

One-to-one model: Features Argument structure features Argument context features Predicate structure features Predicate context features Distance features Most of them inspired from other works

Argument structure features: –syntactic label –head word, suffixes of 2, 3 and 4, lemma, and POS tag –content word computed as in [Surdeanu et al. 03], suffixes of 2, 3 and 4, lemma, and POS tag –first and last constituent words, POS tags –NE labels –binary features for temporal cue words (i.e., words that often occur in AM-TMP phrases in training) –number of occurrences of each possible syntactic label within the constituent –sequence of syntactic labels of the children constituents One-to-one model: Features

Argument context features : –phrase label, head word and POS tag of the constituent parent, and left and right immediate siblings Predicate structure features: –word and lemma –voice (active, passive, copulative, infinitive and progressive) –binary feature to indicate if the predicate is frequent (>2) Predicate context features: –sub-categorization rule: the rule that expands the predicate immediate parent One-to-one model: Features

Distance features : –syntactic path between argument and predicate (chain of labels) –length of the path –number of clauses in the path –number of verb phrases in the path –binary feature to indicate if the argument starts with a predicate particle –surface distance (set of features used in [Collins 99]) –generalized syntactic paths built using templates Arg ↑ Ancestor ↓ X ↓ Pred Arg ↑ X ↑ Ancestor ↓ Pred –subsumption count (difference between the depths in the tree of the predicate and the argument) One-to-one model: Features

One-to-one model:Classifiers Trained one-vs-all classifiers for the top 24 most common arguments Each classifier is developed using AdaBoost with confidence rated predictions [Shapire & Singer,99] Classifiers are combined using a greedy strategy: All predictions sorted in descending order of confidence Add prediction to the global solution if prediction does not violate the constraints related to the previously added predictions Arguments can not overlap No duplicates allowed for A0-A5 Only numeric arguments from PropBank frames are considered

One-to-many: Argument expansion heuristics 1.Argument mapped to one terminal phrase → extend its boundary to the right to include all terminal phrases sharing the same parent and not labeled for the same predicate 2.Argument preceeded/followed by quotes within the same constituent → extend its boundaries to include quotes

Results Trained with all positive examples and the first negative ones. PrecisionRecallF1 Dev.79.14%71.57%75.17% Test WSJ80.32%72.95%76.46% Test Brown72.41%59.67%65.42% Test WSJ+Brown79.35%71.17%75.04% Surpass with almost 6% the results of the best SRL that used partial syntax in the CoNLL’04 Only 0.14% out of the 75.17% in F are due to the argument expansion heuristics

Conclusions This is a simple approach –Basically one-to-one mappings –Greedy combination strategy Given the simplicity and the promising results obtained, it can be taken as a lower- limit for combination methods

thanks