An Entity-Mention Model for Coreference Resolution with Inductive Logic Programming Xiaofeng Yang 1 Jian Su 1 Jun Lang 2 Chew Lim Tan 3 Ting Liu 2 Sheng.

Slides:



Advertisements
Similar presentations
CS460/IT632 Natural Language Processing/Language Technology for the Web Lecture 2 (06/01/06) Prof. Pushpak Bhattacharyya IIT Bombay Part of Speech (PoS)
Advertisements

Specialized models and ranking for coreference resolution Pascal Denis ALPAGE Project Team INRIA Rocquencourt F Le Chesnay, France Jason Baldridge.
A Machine Learning Approach to Coreference Resolution of Noun Phrases By W.M.Soon, H.T.Ng, D.C.Y.Lim Presented by Iman Sen.
Coreference Based Event-Argument Relation Extraction on Biomedical Text Katsumasa Yoshikawa 1), Sebastian Riedel 2), Tsutomu Hirao 3), Masayuki Asahara.
Large-Scale Entity-Based Online Social Network Profile Linkage.
1 Relational Learning of Pattern-Match Rules for Information Extraction Presentation by Tim Chartrand of A paper bypaper Mary Elaine Califf and Raymond.
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
Page 1 Generalized Inference with Multiple Semantic Role Labeling Systems Peter Koomen, Vasin Punyakanok, Dan Roth, (Scott) Wen-tau Yih Department of Computer.
Empirical Methods in Information Extraction - Claire Cardie 자연어처리연구실 한 경 수
Supervised models for coreference resolution Altaf Rahman and Vincent Ng Human Language Technology Research Institute University of Texas at Dallas 1.
Improving Machine Learning Approaches to Coreference Resolution Vincent Ng and Claire Cardie Cornell Univ. ACL 2002 slides prepared by Ralph Grishman.
STRUCTURED PERCEPTRON Alice Lai and Shi Zhi. Presentation Outline Introduction to Structured Perceptron ILP-CRF Model Averaged Perceptron Latent Variable.
A Global Relaxation Labeling Approach to Coreference Resolution Coling 2010 Emili Sapena, Llu´ıs Padr´o and Jordi Turmo TALP Research Center Universitat.
807 - TEXT ANALYTICS Massimo Poesio Lecture 7: Coreference (Anaphora resolution)
Andreea Bodnari, 1 Peter Szolovits, 1 Ozlem Uzuner 2 1 MIT, CSAIL, Cambridge, MA, USA 2 Department of Information Studies, University at Albany SUNY, Albany,
Recognition of Multi-sentence n-ary Subcellular Localization Mentions in Biomedical Abstracts G. Melli, M. Ester, A. Sarkar Dec. 6, 2007
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Inductive learning Simplest form: learn a function from examples
Authors: Ting Wang, Yaoyong Li, Kalina Bontcheva, Hamish Cunningham, Ji Wang Presented by: Khalifeh Al-Jadda Automatic Extraction of Hierarchical Relations.
A Two Tier Framework for Context-Aware Service Organization & Discovery Wei Zhang 1, Jian Su 2, Bin Chen 2,WentingWang 2, Zhiqiang Toh 2, Yanchuan Sim.
Illinois-Coref: The UI System in the CoNLL-2012 Shared Task Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Mark Sammons, and Dan Roth Supported by ARL,
On the Issue of Combining Anaphoricity Determination and Antecedent Identification in Anaphora Resolution Ryu Iida, Kentaro Inui, Yuji Matsumoto Nara Institute.
Incorporating Extra-linguistic Information into Reference Resolution in Collaborative Task Dialogue Ryu Iida Shumpei Kobayashi Takenobu Tokunaga Tokyo.
Part-Of-Speech Tagging using Neural Networks Ankur Parikh LTRC IIIT Hyderabad
Opinion Sentence Search Engine on Open-domain Blog Osamu Furuse, Nobuaki Hiroshima, Setsuo Yamada, Ryoji Kataoka NTT Cyber Solutions Laboratories, NTT.
Using Semantic Relations to Improve Information Retrieval Tom Morton.
Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.
EMNLP’01 19/11/2001 ML: Classical methods from AI –Decision-Tree induction –Exemplar-based Learning –Rule Induction –TBEDL ML: Classical methods from AI.
Coreference Resolution
Recognizing Names in Biomedical Texts: a Machine Learning Approach GuoDong Zhou 1,*, Jie Zhang 1,2, Jian Su 1, Dan Shen 1,2 and ChewLim Tan 2 1 Institute.
Markov Logic and Deep Networks Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
A Cross-Lingual ILP Solution to Zero Anaphora Resolution Ryu Iida & Massimo Poesio (ACL-HLT 2011)
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
Expert Systems with Applications 34 (2008) 459–468 Multi-level fuzzy mining with multiple minimum supports Yeong-Chyi Lee, Tzung-Pei Hong, Tien-Chin Wang.
NYU: Description of the Proteus/PET System as Used for MUC-7 ST Roman Yangarber & Ralph Grishman Presented by Jinying Chen 10/04/2002.
COLING 2012 Extracting and Normalizing Entity-Actions from Users’ comments Swapna Gottipati, Jing Jiang School of Information Systems, Singapore Management.
CS 6998 NLP for the Web Columbia University 04/22/2010 Analyzing Wikipedia and Gold-Standard Corpora for NER Training William Y. Wang Computer Science.
Multi-Relational Data Mining: An Introduction Joe Paulowskey.
1 Toward Opinion Summarization: Linking the Sources Veselin Stoyanov and Claire Cardie Department of Computer Science Cornell University Ithaca, NY 14850,
807 - TEXT ANALYTICS Massimo Poesio Lecture 7: Anaphora resolution (Coreference)
Date : 2013/03/18 Author : Jeffrey Pound, Alexander K. Hudek, Ihab F. Ilyas, Grant Weddell Source : CIKM’12 Speaker : Er-Gang Liu Advisor : Prof. Jia-Ling.
Using Semantic Relations to Improve Passage Retrieval for Question Answering Tom Morton.
Error Analysis for Learning-based Coreference Resolution Olga Uryupina
Extracting Keyphrases to Represent Relations in Social Networks from Web Junichiro Mori and Mitsuru Ishizuka Universiry of Tokyo Yutaka Matsuo National.
For Monday Finish chapter 19 Take-home exam due. Program 4 Any questions?
A Generalized Version Space Learning Algorithm for Noisy and Uncertain Data T.-P. Hong, S.-S. Tseng IEEE Transactions on Knowledge and Data Engineering,
Multilingual Opinion Holder Identification Using Author and Authority Viewpoints Yohei Seki, Noriko Kando,Masaki Aono Toyohashi University of Technology.
Named Entity Disambiguation on an Ontology Enriched by Wikipedia Hien Thanh Nguyen 1, Tru Hoang Cao 2 1 Ton Duc Thang University, Vietnam 2 Ho Chi Minh.
Inference Protocols for Coreference Resolution Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Nick Rizzolo, Mark Sammons, and Dan Roth This research.
Evaluation issues in anaphora resolution and beyond Ruslan Mitkov University of Wolverhampton Faro, 27 June 2002.
Support Vector Machines and Kernel Methods for Co-Reference Resolution 2007 Summer Workshop on Human Language Technology Center for Language and Speech.
1 Fine-grained and Coarse-grained Word Sense Disambiguation Jinying Chen, Hoa Trang Dang, Martha Palmer August 22, 2003.
Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore
Exploiting Named Entity Taggers in a Second Language Thamar Solorio Computer Science Department National Institute of Astrophysics, Optics and Electronics.
Advisor: Hsin-Hsi Chen Reporter: Chi-Hsin Yu Date: From Word Representations:... ACL2010, From Frequency... JAIR 2010 Representing Word... Psychological.
Department of Computer Science The University of Texas at Austin USA Joint Entity and Relation Extraction using Card-Pyramid Parsing Rohit J. Kate Raymond.
Chap. 10 Learning Sets of Rules 박성배 서울대학교 컴퓨터공학과.
Solving Hard Coreference Problems Haoruo Peng, Daniel Khashabi and Dan Roth Problem Description  Problems with Existing Coref Systems Rely heavily on.
Static model noOverlaps :: ArgumentCandidate[] candidates -> discrete[] types for (i : (0.. candidates.size() - 1)) for (j : (i candidates.size()
ACL 2002, Univ. of Pennsylvania, Philadelphia, PA (July 2002) Session: Anaphora and Coreference Session Chair: Lillian Lee Improving Machine Learning.
A Deep Memory Network for Chinese Zero Pronoun Resolution
Part 2 Applications of ILP Formulations in Natural Language Processing
Simone Paolo Ponzetto University of Heidelberg Massimo Poesio
Recovering from Errors during Programming by Demonstration
Social Knowledge Mining
Clustering Algorithms for Noun Phrase Coreference Resolution
Discriminative Frequent Pattern Analysis for Effective Classification
An Interactive Approach to Collectively Resolving URI Coreference
Presentation transcript:

An Entity-Mention Model for Coreference Resolution with Inductive Logic Programming Xiaofeng Yang 1 Jian Su 1 Jun Lang 2 Chew Lim Tan 3 Ting Liu 2 Sheng Li 2 Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen 1 Institute for Infocomm Research 2 Harbin Institute of Technology 3 National University of Singapore ACL 2008

22 Introduction Coreference resolution : The process of linking multiple mentions that refer to the same entity coreference; anaphor 同指涉 antecedent 先行詞 Inductive logic programming : Supervised learning Inductive rule from positive cases 2 2

33 Related Work 1. Mention pair model Aone and Bennett (1995); McCarthy and Lehnert (1995); Soon et al. (2001); Ng and Cardie (2002)) Individual mention usually lacks adequate descriptive information of the referred entity (ex: Powell vs he) 2. Entity-mention model Luo et al., 2004; Yang et al., 2004 Can’t describing each individual mention in an entity 3. Inductive logic programming in NLP Parsing (Mooney, 1997) POS disambiguation (Cussens, 1996) Lexicon construction (Claveau et al., 2003) WSD (Specia et al., 2007) 3

4 Modeling Coreference Resolution The probability that a mention belongs to an entity Example e1 : Microsoft Corp. - its - The company e2 : its new CEO - he e3 : yesterday 4

5 Mention-Pair Model Soon et al. (2001) and Ng and Cardie (2002) Instance i{m k, m j } m j is an active mention & m k is a preceding mention Positive: m j and its closest antecedent (only one for m j ) Negative: every intervening mentions between mj and its closest antecedent m j is linked with the mention that is classified as positive (if any) with the highest confidence value

6 Feature Set for Coreference Resolution 同位語 述詞

7 Entity-Mention Model Mention-pair model error: Lack adequate descriptive information “ Mr. Powell ”, “ Powell ”, and “s he ” Instance i{m k, m j } Positive: m j and the entity to which m j belongs. Negative: every entity whose last mention occurs between m j and its closest antecedent If no positive entity exists, the m j forms a new entity entity features: first-order features Any-X, Most-X, All-X Distance feature :the minimum distance between the mentions in the entity and the active mention.

8 Entity-mention Model with ILP (1/3) Tool: ALEPH by Srinivasan (2000) (Oxford) Input: positive example E + negative example E - background knowledge K Output: hypotheses h e 1_6 denotes the part of e 1 before m 6, example representation: link(e 1_6, m 6 ) 8

9 Entity-mention Model with ILP (2/3) background knowledge K predicates 1. Information related to e i_j and m j 2. Relations between e i_j and its mentions has_mention(e 1_6, m 6 ) 3. Information related to m j and each mention m k in e i_j 9

10 Entity-mention Model with ILP (3/3) Hypothesis rule link(A,B) :- has mention(A,C), numAgree(B,C,1), strMatch Head(B,C,1), bareNP(C,1).

11 Experiments and Result(1/4) Corpus: ACE-2 V1.0 corpus (NIST, 2003) Modify ILP tool, ALEPH: Rule accuracy 100% to 50% 3 predicates to 10 predicates 11

12 Baseline model: C4.5 algorithm Preprocessing Tokenizer Part-of-Speech tagger accuracy of 97% on Penn WSJ TreeBank NP chunker (Zhou and Su, 2000) F-measure above 94% on Penn WSJ TreeBank Named-Entity Recognizer (Zhou and Su, 2002) F-measure of 96.6% (MUC-6) and 94.1%(MUC-7) 12 Experiments and Result(2/4)

13 Experiments and Result(3/4) F-measure is 2-4% lower than the state-of-the- art, which utilized sophisticated semantic or real world knowledge Significant under 2-tailed t test (p < 0.05) 13

14 Experiments and Result(4/4) Multiple non-instantiated arguments (i.e. C and D) could possibly appear in the same rule

15 Conclusion & Future Work The model can express the relations between an entity and its mentions, and to automatically learn the first-order rules ILP based entity-model performs better than the mention-pair model (with up to 2.3% increase in F- measure) Future work: Investigate more sophisticated clustering methods that would lead to global optimization keeping a large search space (Luo et al., 2004) using integer programming (Denis and Baldridge, 2007) 15

16 Thank you! 16