Improving Machine Learning Approaches to Coreference Resolution Vincent Ng and Claire Cardie Cornell Univ. ACL 2002 slides prepared by Ralph Grishman.

Slides:



Advertisements
Similar presentations
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Advertisements

The Software Infrastructure for Electronic Commerce Databases and Data Mining Lecture 4: An Introduction To Data Mining (II) Johannes Gehrke
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Conceptualization, Operationalization, and Measurement
Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
A Machine Learning Approach to Coreference Resolution of Noun Phrases By W.M.Soon, H.T.Ng, D.C.Y.Lim Presented by Iman Sen.
CONSTRAINED CONDITIONAL MODELS TUTORIAL Jingyu Chen, Xiao Cheng.
K-NEAREST NEIGHBORS AND DECISION TREE Nonparametric Supervised Learning.
Chapter 18: Discourse Tianjun Fu Ling538 Presentation Nov 30th, 2006.
From last time What’s the real point of using vector spaces?: A user’s query can be viewed as a (very) short document. Query becomes a vector in the same.
Faculty Of Applied Science Simon Fraser University Cmpt 825 presentation Corpus Based PP Attachment Ambiguity Resolution with a Semantic Dictionary Jiri.
CS 4705 Algorithms for Reference Resolution. Anaphora resolution Finding in a text all the referring expressions that have one and the same denotation.
Generic Object Detection using Feature Maps Oscar Danielsson Stefan Carlsson
July 9, 2003ACL An Improved Pattern Model for Automatic IE Pattern Acquisition Kiyoshi Sudo Satoshi Sekine Ralph Grishman New York University.
CS 4705 Lecture 21 Algorithms for Reference Resolution.
Supervised models for coreference resolution Altaf Rahman and Vincent Ng Human Language Technology Research Institute University of Texas at Dallas 1.
OntoNotes project Treebank Syntax Training Data Decoders Propositions Verb Senses and verbal ontology links Noun Senses and targeted nominalizations Coreference.
A Memory-Based Approach to Semantic Role Labeling Beata Kouchnir Tübingen University 05/07/04.
Anaphora Resolution Sanghoon Kwak Takahiro Aoyama.
SI485i : NLP Set 14 Reference Resolution. 2 Kraken, also called the Crab-fish, which is not that huge, for heads and tails counted, he is no larger than.
SI485i : NLP Set 9 Advanced PCFGs Some slides from Chris Manning.
A Global Relaxation Labeling Approach to Coreference Resolution Coling 2010 Emili Sapena, Llu´ıs Padr´o and Jordi Turmo TALP Research Center Universitat.
A Light-weight Approach to Coreference Resolution for Named Entities in Text Marin Dimitrov Ontotext Lab, Sirma AI Kalina Bontcheva, Hamish Cunningham,
Andreea Bodnari, 1 Peter Szolovits, 1 Ozlem Uzuner 2 1 MIT, CSAIL, Cambridge, MA, USA 2 Department of Information Studies, University at Albany SUNY, Albany,
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Automatic Detection of Tags for Political Blogs Khairun-nisa Hassanali and Vasileios Hatzivassiloglou Human Language Technology Research Institute The.
Illinois-Coref: The UI System in the CoNLL-2012 Shared Task Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Mark Sammons, and Dan Roth Supported by ARL,
On the Issue of Combining Anaphoricity Determination and Antecedent Identification in Anaphora Resolution Ryu Iida, Kentaro Inui, Yuji Matsumoto Nara Institute.
Incorporating Extra-linguistic Information into Reference Resolution in Collaborative Task Dialogue Ryu Iida Shumpei Kobayashi Takenobu Tokunaga Tokyo.
Combining terminology resources and statistical methods for entity recognition: an evaluation Angus Roberts, Robert Gaizauskas, Mark Hepple, Yikun Guo.
Using Semantic Relations to Improve Information Retrieval Tom Morton.
1 Exploiting Syntactic Patterns as Clues in Zero- Anaphora Resolution Ryu Iida, Kentaro Inui and Yuji Matsumoto Nara Institute of Science and Technology.
Coreference Resolution
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
A Cross-Lingual ILP Solution to Zero Anaphora Resolution Ryu Iida & Massimo Poesio (ACL-HLT 2011)
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
Reference Resolution. Sue bought a cup of coffee and a donut from Jane. She met John as she left. He looked at her enviously as she drank the coffee.
An Entity-Mention Model for Coreference Resolution with Inductive Logic Programming Xiaofeng Yang 1 Jian Su 1 Jun Lang 2 Chew Lim Tan 3 Ting Liu 2 Sheng.
1 Toward Opinion Summarization: Linking the Sources Veselin Stoyanov and Claire Cardie Department of Computer Science Cornell University Ithaca, NY 14850,
Error Analysis for Learning-based Coreference Resolution Olga Uryupina
Sets of Digital Data CSCI 2720 Fall 2005 Kraemer.
Inference Protocols for Coreference Resolution Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Nick Rizzolo, Mark Sammons, and Dan Roth This research.
Evaluation issues in anaphora resolution and beyond Ruslan Mitkov University of Wolverhampton Faro, 27 June 2002.
Measuring the Influence of Errors Induced by the Presence of Dialogs in Reference Clustering of Narrative Text Alaukik Aggarwal, Department of Computer.
Support Vector Machines and Kernel Methods for Co-Reference Resolution 2007 Summer Workshop on Human Language Technology Center for Language and Speech.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
Copyright  2004 limsoon wong Using WEKA for Classification (without feature selection)
Learning Event Durations from Event Descriptions Feng Pan, Rutu Mulkar, Jerry R. Hobbs University of Southern California ACL ’ 06.
ANAPHORA RESOLUTION SYSTEM FOR NATURAL LANGUAGE REQUIREMENTS DOCUMENT IN KOREAN 課程 : 自然語言與應用 課程老師 : 顏國郎 報告者 : 鄭冠瑀.
Dependency Parsing Niranjan Balasubramanian March 24 th 2016 Credits: Many slides from: Michael Collins, Mausam, Chris Manning, COLNG 2014 Dependency Parsing.
Relation Extraction (RE) via Supervised Classification See: Jurafsky & Martin SLP book, Chapter 22 Exploring Various Knowledge in Relation Extraction.
10. Decision Trees and Markov Chains for Gene Finding.
Automatic Writing Evaluation
Criterial features If you have examples of language use by learners (differentiated by L1 etc.) at different levels, you can use that to find the criterial.
ACL 2002, Univ. of Pennsylvania, Philadelphia, PA (July 2002) Session: Anaphora and Coreference Session Chair: Lillian Lee Improving Machine Learning.
Linguistic Graph Similarity for News Sentence Searching
k-Nearest neighbors and decision tree
Data Science Algorithms: The Basic Methods
Simone Paolo Ponzetto University of Heidelberg Massimo Poesio
Prepared by: Mahmoud Rafeek Al-Farra
Relation Extraction CSCI-GA.2591
NYU Coreference CSCI-GA.2591 Ralph Grishman.
K Nearest Neighbor Classification
Algorithms for Reference Resolution
A Machine Learning Approach to Coreference Resolution of Noun Phrases
Learning to Classify Documents Edwin Zhang Computer Systems Lab
Automatic Detection of Causal Relations for Question Answering
A Machine Learning Approach to Coreference Resolution of Noun Phrases
The Winograd Schema Challenge Hector J. Levesque AAAI, 2011
Presentation transcript:

Improving Machine Learning Approaches to Coreference Resolution Vincent Ng and Claire Cardie Cornell Univ. ACL 2002 slides prepared by Ralph Grishman

Goal Improve on Soon et al. by 4 better preprocessing (chunking, names, …) 4 better search procedure for antecedent 4 better selection of positive examples 4 more features 4 more features...

Better search for antecedent 4 Soon et al. Use decision tree as binary classifier, take nearest antecedent classified as +ve 4 Ng&Cardie use same sort of classifier, but count +ve and -ve examples at each leaf, and use that to compute a probability 4 Ng&Cardie then take highest ranking antecedent (if probability > 0.5)

Better choice of positive examples 4 Soon et al. always use most recent antecedent 4 For Ng&Cardie, if anaphor is not a pronoun, they use most recent antecedent that is not a pronoun

More features #1 4 Soon et al. Have a ‘same string’ feature 4 Ng&Cardie split this up into 3 features, for pronominals, nominals, and names

First improvements: F scores

More features Added 41 more features: 4 lexical 4 grammatical 4 semantic

Lexical features (examples) 4 Non-empty overlap of words of two NPs 4 Prenominal modifiers of one NP are a subset of prenominal modifiers of other

Grammatical features (examples) 4 NPs are in predicate nominal construct 4 One NP spans the other 4 NP1 is a quoted string 4 One of the NPs is a title

Semantic features (examples) For nominals with different heads 4 direct or indirect hypernym relation in WordNet 4 distance of hypernym relation 4 sense number for hypernym relation

Selecting features 4 Full feature set yielded very low precision on nominal anaphors overtraining: too many features for too little data 4 So they (manually) eliminated many features which led to low precision (on training data) no ‘development set’ separate from training and test sets

Adding features: F scores