Normalized alignment of dependency trees for detecting textual entailment Erwin Marsi & Emiel Krahmer Tilburg University Wauter Bosma & Mariët Theune University.

Slides:



Advertisements
Similar presentations
COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Advertisements

COGEX at the Second RTE Marta Tatu, Brandon Iles, John Slavick, Adrian Novischi, Dan Moldovan Language Computer Corporation April 10 th, 2006.
Progress update Lin Ziheng. System overview 2 Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self.
QA-LaSIE Components The question document and each candidate answer document pass through all nine components of the QA-LaSIE system in the order shown.
Recognizing Textual Entailment Challenge PASCAL Suleiman BaniHani.
Sequence Classification: Chunking Shallow Processing Techniques for NLP Ling570 November 28, 2011.
Baselines for Recognizing Textual Entailment Ling 541 Final Project Terrence Szymanski.
Made with OpenOffice.org 1 Sentiment Classification using Word Sub-Sequences and Dependency Sub-Trees Pacific-Asia Knowledge Discovery and Data Mining.
Recognizing Implicit Discourse Relations in the Penn Discourse Treebank Ziheng Lin, Min-Yen Kan, and Hwee Tou Ng Department of Computer Science National.
In Search of a More Probable Parse: Experiments with DOP* and the Penn Chinese Treebank Aaron Meyers Linguistics 490 Winter 2009.
Robust Textual Inference via Graph Matching Aria Haghighi Andrew Ng Christopher Manning.
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
Syntactic Contributions in the Entailment Task Lucy Vanderwende, Arul Menezes, Rion Snow (Stanford)
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
1 RTE April 10, 2006 An approach based on Logic Forms and WordNet relationships to Textual Entailment performance O. Ferrández, R. M. Terol, R. Muñoz,
Automatic Classification of Semantic Relations between Facts and Opinions Koji Murakami, Eric Nichols, Junta Mizuno, Yotaro Watanabe, Hayato Goto, Megumi.
Using Maximal Embedded Subtrees for Textual Entailment Recognition Sophia Katrenko & Pieter Adriaans Adaptive Information Disclosure project Human Computer.
Dependency Parsing with Reference to Slovene, Spanish and Swedish Simon Corston-Oliver Anthony Aue Microsoft Research.
A Framework for Named Entity Recognition in the Open Domain Richard Evans Research Group in Computational Linguistics University of Wolverhampton UK
Parsing SLP Chapter 13. 7/2/2015 Speech and Language Processing - Jurafsky and Martin 2 Outline  Parsing with CFGs  Bottom-up, top-down  CKY parsing.
A Confidence Model for Syntactically-Motivated Entailment Proofs Asher Stern & Ido Dagan ISCOL June 2011, Israel 1.
NATURAL LANGUAGE TOOLKIT(NLTK) April Corbet. Overview 1. What is NLTK? 2. NLTK Basic Functionalities 3. Part of Speech Tagging 4. Chunking and Trees 5.
Page 1 Relation Alignment for Textual Entailment Recognition Department of Computer Science University of Illinois at Urbana-Champaign Mark Sammons, V.G.Vinod.
Outline P1EDA’s simple features currently implemented –And their ablation test Features we have reviewed from Literature –(Let’s briefly visit them) –Iftene’s.
Knowledge and Tree-Edits in Learnable Entailment Proofs Asher Stern and Ido Dagan (earlier partial version by Roy Bar-Haim) Download at:
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
Fabio Massimo Zanzotto
An Integrated Approach to Extracting Ontological Structures from Folksonomies Huairen Lin, Joseph Davis, Ying Zhou ESWC 2009 Hyewon Lim October 9 th, 2009.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
Knowledge and Tree-Edits in Learnable Entailment Proofs Asher Stern, Amnon Lotan, Shachar Mirkin, Eyal Shnarch, Lili Kotlerman, Jonathan Berant and Ido.
Syntactic and semantic models and algorithms in Question Answering Alexander Solovyev Bauman Moscow Sate Technical University RCDL.
The CoNLL-2013 Shared Task on Grammatical Error Correction Hwee Tou Ng, Yuanbin Wu, and Christian Hadiwinoto 1 Siew.
Automatic Detection of Tags for Political Blogs Khairun-nisa Hassanali Vasileios Hatzivassiloglou The University.
Mastering the Pipeline CSCI-GA.2590 Ralph Grishman NYU.
2010/2/4Yi-Ting Huang Pennacchiotti, M., & Zanzotto, F. M. Learning Shallow Semantic Rules for Textual Entailment. Recent Advances in Natural Language.
Inductive Dependency Parsing Joakim Nivre
Training dependency parsers by jointly optimizing multiple objectives Keith HallRyan McDonaldJason Katz- BrownMichael Ringgaard.
Malay-English Bitext Mapping and Alignment Using SIMR/GSA Algorithms Mosleh Al-Adhaileh Tang Enya Kong Mosleh Al-Adhaileh and Tang Enya Kong Computer Aided.
Approximating a Deep-Syntactic Metric for MT Evaluation and Tuning Matouš Macháček, Ondřej Bojar; {machacek, Charles University.
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
1 Intelligente Analyse- und Informationssysteme Frank Reichartz, Hannes Korte & Gerhard Paass Fraunhofer IAIS, Sankt Augustin, Germany Dependency Tree.
Element Level Semantic Matching Pavel Shvaiko Meaning Coordination and Negotiation Workshop, ISWC 8 th November 2004, Hiroshima, Japan Paper by Fausto.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Automatic Grammar Induction and Parsing Free Text - Eric Brill Thur. POSTECH Dept. of Computer Science 심 준 혁.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
1 Measuring the Semantic Similarity of Texts Author : Courtney Corley and Rada Mihalcea Source : ACL-2005 Reporter : Yong-Xiang Chen.
Emotion Recognition from Text Using Situational Information and a Personalized Emotion Model Yong-soo Seol 1, Han-woo Kim 1, and Dong-joo Kim 2 1 Department.
Nirmalya Roy School of Electrical Engineering and Computer Science Washington State University Cpt S 223 – Advanced Data Structures Math Review 1.
SALSA-WS 09/05 Approximating Textual Entailment with LFG and FrameNet Frames Aljoscha Burchardt, Anette Frank Computational Linguistics Department Saarland.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
NATURAL LANGUAGE PROCESSING
Correcting Misuse of Verb Forms John Lee, Stephanie Seneff Computer Science and Artificial Intelligence Laboratory, MIT, Cambridge ACL 2008.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Towards an Extractive Summarization System Using Sentence Vectors and Clustering John Cadigan, David Ellison, Ethan Roday.
Mastering the Pipeline CSCI-GA.2590 Ralph Grishman NYU.
Linguistic Graph Similarity for News Sentence Searching
Web News Sentence Searching Using Linguistic Graph Similarity
Learning Textual Entailment from Examples
The Resolution of Speculation and Negation
Natural Language Processing (NLP)
Recognizing Partial Textual Entailment
Review-Level Aspect-Based Sentiment Analysis Using an Ontology
CS246: Information Retrieval
Natural Language Processing (NLP)
Word embeddings (continued)
Johns Hopkins 2003 Summer Workshop on Syntax and Statistical Machine Translation Chapters 5-8 Ethan Phelps-Goodman.
Natural Language Processing (NLP)
Presentation transcript:

Normalized alignment of dependency trees for detecting textual entailment Erwin Marsi & Emiel Krahmer Tilburg University Wauter Bosma & Mariët Theune University of Twente

April RTE2 Workshop2 Basic idea A true hypothesis is included in the text, allowing omission and rephrasing Text: The Rolling Stones kicked off their latest tour on Sunday with a concert at Boston's Fenway Park. Hypothesis: The Rolling Stones have begun their latest tour with a concert in Boston. Entailment: True Omissions: –on Sunday –Fenway Park Paraphrases: –kicked off  begun –Boston's Fenway Park  Boston

April RTE2 Workshop3 Matching surface words alone is not sufficient... Variation in surface realization  perfect word match is no guarantee for entailment Using syntactic analysis –for syntactic normalization –to match on hierarchical relations among constituents Example: “He became a boxing referee in 1964, and became well- known […]” “He became well-known in 1964”

April RTE2 Workshop4 Preprocessing Input: T-H pairs in XML Processing pipeline: 1.Sentence splitting, MXTERMINATOR (Reynar & Ratnaparkhi, 1997) 2.Tokenization, Penn Treebank SED script 3.POS tagging with PTB POS tags using Mbt (van den Bosch et al) 4.Lemmatizing using Memory-based learning (van den Bosch et al) 5.Dependency parsing using Maltparser trained on PTB (Nivre & Scholz, 2004) 6.Syntactic normalization Output: T-H dependency tree(s) pairs in XML

April RTE2 Workshop5 Syntactic Normalization Three types of syntactic normalization: –Auxiliary reduction –Passive to active form –Copula reduction

April RTE2 Workshop6 Auxiliary Reduction Auxiliaries of progressive and perfective tense are removed Their children are attached to the remaining content verb The same goes for modal verbs, and for do in the do-support function. Example: “demand for ivory has dropped”  “demand for ivory dropped” Example: “legalization does not solve any social problems”  “legalization not solves any social problems”

April RTE2 Workshop7 Passive to Active Form The passive form auxiliary is removed The original subject becomes object Where possible, a by-phrase becomes the subject Example: “Ahmedinejad was attacked by the US”  “the US attacked Ahmedinejad”

April RTE2 Workshop8 Copula Reduction Copular verbs are removed by attaching the predicate as a daughter to the subject Example: “Microsoft Corp. is a partner of Intel Corp.”  “Microsoft Corp., a partner of Intel Corp.”

April RTE2 Workshop9 Alignment of Dependency Trees Tree alignment algorithm based on (Meyers, Yangarbar and Grishman, 1996) Searches for an optimal alignment of the nodes of the text tree to the nodes of the hypothesis tree Tree alignment is a function of: 1.how well the words of the two nodes match 2.recursively, the weighted alignment score for each of the aligned daughter nodes

April RTE2 Workshop10 Word Matching function WordMatch(w t,w h ) -> [0,1] maps text-hypothesis word pairs to a similarity score returns 1 if –w t is identical to w h –the lemma of w t is identical to the lemma of w h –w t is a synonym of w h (lookup in EuroWordnet with lemma & POS) –w h is a hypernym of w t (idem) returns similarity from automatically derived thesaurus if > 0.1 (Lin’s dependency-based thesaurus) otherwise returns 0 also match on phrasal verbs –e.g. “kick off“ is a synonym of “begin“

April RTE2 Workshop11 Alignment example Text: The development of agriculture by early humans, roughly 10,000 years ago, was also harmful to many natural ecosystems as they were systematically destroyed and replaced with artificial versions. Hypothesis: Humans existed 10,000 years ago. Entailment: True

April RTE2 Workshop12 Alignment example (cont’d)

April RTE2 Workshop13 Entailment prediction Prediction rule: IF top node of the hypothesis is aligned AND score > threshold THEN entailment = true ELSE entailment = false Threshold and parameters of tree alignment algorithm (skip penalty) optimized per task

April RTE2 Workshop14 Results Percentage entailment accuracy (n=800)

April RTE2 Workshop15 Problems Many parses contain errors due to syntactic ambiguity and propagation of –Spelling errors –Tokenization errors –POS errors –broken dependency trees Consequently, syntactic normalization & alignment failed Dependency relations did not help

April RTE2 Workshop16 Discussion & Conclusion There are many forms of textual entailment that we cannot recognize automatically... –Paraphrasing –Co-reference resolution –Ellipsis –Condition/modality –Inference –Common sense / world knowledge RTE requires a combination of deep NLP, common sense knowledge and reasoning