Progress update Lin Ziheng. System overview 2 Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self.

Slides:



Advertisements
Similar presentations
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Advertisements

Using Syntax to Disambiguate Explicit Discourse Connectives in Text Source: ACL-IJCNLP 2009 Author: Emily Pitler and Ani Nenkova Reporter: Yong-Xiang Chen.
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
Tracking L2 Lexical and Syntactic Development Xiaofei Lu CALPER 2010 Summer Workshop July 14, 2010.
A Machine Learning Approach to Coreference Resolution of Noun Phrases By W.M.Soon, H.T.Ng, D.C.Y.Lim Presented by Iman Sen.
Albert Gatt Corpora and Statistical Methods Lecture 11.
Automatically Evaluating Text Coherence Using Discourse Relations Ziheng Lin, Hwee Tou Ng and Min-Yen Kan Department of Computer Science National University.
Sequence Classification: Chunking Shallow Processing Techniques for NLP Ling570 November 28, 2011.
Event Extraction Using Distant Supervision Kevin Reschke, Martin Jankowiak, Mihai Surdeanu, Christopher D. Manning, Daniel Jurafsky 30 May 2014 Language.
A Joint Model For Semantic Role Labeling Aria Haghighi, Kristina Toutanova, Christopher D. Manning Computer Science Department Stanford University.
Recognizing Implicit Discourse Relations in the Penn Discourse Treebank Ziheng Lin, Min-Yen Kan, and Hwee Tou Ng Department of Computer Science National.
Discourse Parsing in the Penn Discourse Treebank: Using Discourse Structures to Model Coherence and Improve User Tasks Ziheng Lin Ph.D. Thesis Proposal.
Shallow Parsing CS 4705 Julia Hirschberg 1. Shallow or Partial Parsing Sometimes we don’t need a complete parse tree –Information extraction –Question.
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
Reranking Parse Trees with a SRL system Charles Sutton and Andrew McCallum University of Massachusetts June 30, 2005.
SRL using complete syntactic analysis Mihai Surdeanu and Jordi Turmo TALP Research Center Universitat Politècnica de Catalunya.
1 Annotation Guidelines for the Penn Discourse Treebank Part B Eleni Miltsakaki, Rashmi Prasad, Aravind Joshi, Bonnie Webber.
1 I256: Applied Natural Language Processing Marti Hearst Sept 25, 2006.
SI485i : NLP Set 9 Advanced PCFGs Some slides from Chris Manning.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Na-Rae Han (University of Pittsburgh), Joel Tetreault (ETS), Soo-Hwa Lee (Chungdahm Learning, Inc.), Jin-Young Ha (Kangwon University) May , LREC.
Robert Hass CIS 630 April 14, 2010 NP NP↓ Super NP tagging JJ ↓
Andreea Bodnari, 1 Peter Szolovits, 1 Ozlem Uzuner 2 1 MIT, CSAIL, Cambridge, MA, USA 2 Department of Information Studies, University at Albany SUNY, Albany,
Title Extraction from Bodies of HTML Documents and its Application to Web Page Retrieval Microsoft Research Asia Yunhua Hu, Guomao Xin, Ruihua Song, Guoping.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
The CoNLL-2013 Shared Task on Grammatical Error Correction Hwee Tou Ng, Yuanbin Wu, and Christian Hadiwinoto 1 Siew.
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
Syntax The study of how words are ordered and grouped together Key concept: constituent = a sequence of words that acts as a unit he the man the short.
Ling 570 Day 17: Named Entity Recognition Chunking.
Mining and Analysis of Control Structure Variant Clones Guo Qiao.
Open Information Extraction using Wikipedia
1 Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Cornell University Department of Computer Science.
Training dependency parsers by jointly optimizing multiple objectives Keith HallRyan McDonaldJason Katz- BrownMichael Ringgaard.
Automatic classification for implicit discourse relations Lin Ziheng.
Triplet Extraction from Sentences Lorand Dali Blaž “Jožef Stefan” Institute, Ljubljana 17 th of October 2008.
1 Boosting-based parse re-ranking with subtree features Taku Kudo Jun Suzuki Hideki Isozaki NTT Communication Science Labs.
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
Transformation-Based Learning Advanced Statistical Methods in NLP Ling 572 March 1, 2012.
A Systematic Exploration of the Feature Space for Relation Extraction Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois,
Prototype-Driven Learning for Sequence Models Aria Haghighi and Dan Klein University of California Berkeley Slides prepared by Andrew Carlson for the Semi-
Improving Morphosyntactic Tagging of Slovene by Tagger Combination Jan Rupnik Miha Grčar Tomaž Erjavec Jožef Stefan Institute.
Dependency Parser for Swedish Project for EDA171 by Jonas Pålsson Marcus Stamborg.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Supertagging CMSC Natural Language Processing January 31, 2006.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
In silico immune response prediction based on peptide array data Mitja Luštrek Institute for Biostatistics and Informatics in Medicine and Aging Research.
4. Relationship Extraction Part 4 of Information Extraction Sunita Sarawagi 9/7/2012CS 652, Peter Lindes1.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Department of Computer Science The University of Texas at Austin USA Joint Entity and Relation Extraction using Card-Pyramid Parsing Rohit J. Kate Raymond.
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
A Syntax-Driven Bracketing Model for Phrase-Based Translation Deyi Xiong, et al. ACL 2009.
Dependency Parsing Niranjan Balasubramanian March 24 th 2016 Credits: Many slides from: Michael Collins, Mausam, Chris Manning, COLNG 2014 Dependency Parsing.
Relation Extraction (RE) via Supervised Classification See: Jurafsky & Martin SLP book, Chapter 22 Exploring Various Knowledge in Relation Extraction.
Roadmap Probabilistic CFGs –Handling ambiguity – more likely analyses –Adding probabilities Grammar Parsing: probabilistic CYK Learning probabilities:
Natural Language Processing Information Extraction Jim Martin (slightly modified by Jason Baldridge)
Language Identification and Part-of-Speech Tagging
The 2nd Workshop on Argumentation Mining
CRF &SVM in Medication Extraction
Erasmus University Rotterdam
Improving a Pipeline Architecture for Shallow Discourse Parsing
Aspect-based sentiment analysis
Constraining Chart Parsing with Partial Tree Bracketing
Basic Text Processing: Sentence Segmentation
Johns Hopkins 2003 Summer Workshop on Syntax and Statistical Machine Translation Chapters 5-8 Ethan Phelps-Goodman.
Extracting Why Text Segment from Web Based on Grammar-gram
Presentation transcript:

Progress update Lin Ziheng

System overview 2

Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self category: IN – Parent category: SBAR – Left sibling category: none – Right sibling category: S – Right sibling contains a VP: yes 3

Components – Connective classifier New features – Conn POS – Prev word + conn: even though, particularly since – Prev word POS – Prev word POS + conn POS – Conn + Next word – Next word POS – Conn POS + Next word POS – All lemmatized verbs in the sentence containing conn 4

Components – Argument labeler 5

Argument labeler – Argument position classifier Relative positions of Arg1 – Arg1 and Arg2 in the same sentence: SS (60.9%) – Arg1 in the immediately previous sentence: IPS (30.1%) – Arg1 in some non-adjacent previous sentence: NAPS (9.0%) – Arg1 in some following sentence: FS (0%, only 8 instances) FS ignored 6

Argument labeler – Argument position classifier Features: – Connective string – Conn POS – Conn position in the sentence: first, second, third, third last, second last, or last – Prev word – Prev word POS – Prev word + conn – Prev word POS + conn POS – Second prev word – Second prev word POS – Second prev word + conn – Second prev word POS + conn POS 7

Argument labeler – Argument extractor SS cases: handcrafted a set of syntactically motivated rules to extract Arg1 and Arg2 8

Argument labeler – Argument extractor An example: 9

Argument labeler – Argument extractor IPS cases: label the sentence containing the connective as Arg2 and the immediately previous sentence as Arg1 NAPS cases: – Arg1 locates in the second previous sentence in 45.8% of the NAPS cases – Use the majority decision and assume Arg1 is always in the second previous sentence 10

Components – Explicit classifier Prasad et al. (2008) reported human agreements of 94% on Level 1 classes and 84% on Level 2 types A baseline using only connectives as features gives 95.7% and 86% on Sec. 23 – Difficult to improve acc. on testing section 3 types of features: – Connective string – Conn POS – Conn + prev word 11

Components – Non-explicit classifier Non-explicit: Implicit, AltLex, EntRel, NoRel – 11 Level 2 types for Implicit/AltLex, plus EntRel and NoRel  13 types 4 feature sets from Lin et al. (2009) – Contextual features – Constituent parse features – Dependency parse features – Word-pair features 3 features to capture AltLex: Arg2_word1, Arg2_word2, Arg2_word3 12

Components – Attribution span labeler Two steps: split the text into clauses, and decide which clauses are attribution spans Rule-based clause splitter: – first split a sentence into clauses by punctuations – for each clause, we further split it if one of the following production links if found: VP  SBAR, S  SINV, S  S, SINV  S, S  SBAR, VP  S 13

Components – Attribution span labeler Attr span classifier features: (curr, prev and next clauses) – Unigrams of curr – Lowercased and lemmatized vers in curr – The first and last terms of curr – The last term of prev – The first term of next – The last term of prev + the first term of curr – The last term of curr + the first term of next – The position of curr in the sentence – Punctuations rules extracted from curr 14

Evaluation Train: 02-21, dev: 22, test: 23 Each component is tested – without and with error propagation (EP) from previous component – with gold standard (GS) parse trees and sentence boundaries, and with automatic (Auto) parser and sentence splitter 15

Evaluation – Connective classifier GS: increased acc and F1 by 2.05% and 3.05% Auto: increased acc and F1 by 1.71% and 2.54% Contextual info is helpful 16

Evaluation – Argument position classifier Able to accurately label SS But performs badly on the NAPS class – Due to the similarity between IPS and NAPS classes 17

Evaluation – Argument extractor Human agreements on partial and exact matches: 94.5% and 90.2% Exact F1 much lower than partial F1 – Due to small portions of text deleted 18

Evaluation – Explicit classifier Baseline: using only connective strings – 86% GS + no EP F1 increased by 0.44% 19

Evaluation – Non-explicit classifier Majority baseline: all classified as EntRel Adding EP degrades F1 by ~13%, but still outperforms baseline by ~6% 20

Evaluation – Attribution span labeler When EP added: the decrease of F1 is largely due to the drop in precision When Auto added: the decrease of F1 is largely due the drop in recall 21

Evaluation – The whole pipeline Definition: a relation is correct if its relation type is classified correctly, and both Arg1 and Arg2 are partially or exactly matched GS + EP – Partial: 46.38% F1 – Exact: 31.72% F1 22

On-going changes Joint learning Change rule-based argument extractor to a machine learning approach 23