Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)

Slides:



Advertisements
Similar presentations
A probabilistic model for retrospective news event detection
Advertisements

Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
Query Chain Focused Summarization Tal Baumel, Rafi Cohen, Michael Elhadad Jan 2014.
Multi-Document Person Name Resolution Michael Ben Fleischman (MIT), Eduard Hovy (USC) From Proceedings of ACL-42 Reference Resolution workshop 2004.
Improved TF-IDF Ranker
Comparing Twitter Summarization Algorithms for Multiple Post Summaries David Inouye and Jugal K. Kalita SocialCom May 10 Hyewon Lim.
LEDIR : An Unsupervised Algorithm for Learning Directionality of Inference Rules Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From EMNLP.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
DIMENSIONALITY REDUCTION BY RANDOM PROJECTION AND LATENT SEMANTIC INDEXING Jessica Lin and Dimitrios Gunopulos Ângelo Cardoso IST/UTL December
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
1 Lecture 8 Measures of association: chi square test, mutual information, binomial distribution and log likelihood ratio.
Cover Coefficient based Multidocument Summarization CS 533 Information Retrieval Systems Özlem İSTEK Gönenç ERCAN Nagehan PALA.
Extracting Interest Tags from Twitter User Biographies Ying Ding, Jing Jiang School of Information Systems Singapore Management University AIRS 2014, Kuching,
1 Multi-document Summarization and Evaluation. 2 Task Characteristics  Input: a set of documents on the same topic  Retrieved during an IR search 
Artificial Intelligence Research Centre Program Systems Institute Russian Academy of Science Pereslavl-Zalessky Russia.
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
CHAMELEON : A Hierarchical Clustering Algorithm Using Dynamic Modeling
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Learning Information Extraction Patterns Using WordNet Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield,
Exploiting Wikipedia as External Knowledge for Document Clustering Sakyasingha Dasgupta, Pradeep Ghosh Data Mining and Exploration-Presentation School.
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
Adaptive News Access Daniel Billsus Presented by Chirayu Wongchokprasitti.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
1 Text Summarization: News and Beyond Kathleen McKeown Department of Computer Science Columbia University.
Processing of large document collections Part 7 (Text summarization: multi- document summarization, knowledge- rich approaches, current topics) Helena.
Incident Threading for News Passages (CIKM 09) Speaker: Yi-lin,Hsu Advisor: Dr. Koh, Jia-ling. Date:2010/06/14.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
LexRank: Graph-based Centrality as Salience in Text Summarization
A Machine Learning Approach to Sentence Ordering for Multidocument Summarization and Its Evaluation D. Bollegala, N. Okazaki and M. Ishizuka The University.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
From Social Bookmarking to Social Summarization: An Experiment in Community-Based Summary Generation Oisin Boydell, Barry Smyth Adaptive Information Cluster,
LexPageRank: Prestige in Multi- Document Text Summarization Gunes Erkan and Dragomir R. Radev Department of EECS, School of Information University of Michigan.
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
1 Web-Page Summarization Using Clickthrough Data* JianTao Sun, Yuchang Lu Dept. of Computer Science TsingHua University Beijing , China Dou Shen,
Event-Centric Summary Generation Lucy Vanderwende, Michele Banko and Arul Menezes One Microsoft Way, WA, USA DUC 2004.
Processing of large document collections Part 5 (Text summarization) Helena Ahonen-Myka Spring 2005.
DOCUMENT UPDATE SUMMARIZATION USING INCREMENTAL HIERARCHICAL CLUSTERING CIKM’10 (DINGDING WANG, TAO LI) Advisor: Koh, Jia-Ling Presenter: Nonhlanhla Shongwe.
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
1 A Web Search Engine-Based Approach to Measure Semantic Similarity between Words Presenter: Guan-Yu Chen IEEE Trans. on Knowledge & Data Engineering,
Boundary Detection in Tokenizing Network Application Payload for Anomaly Detection Rachna Vargiya and Philip Chan Department of Computer Sciences Florida.
A New Multi-document Summarization System Yi Guo and Gorge Stylios Heriot-Watt University, Scotland, U.K. (DUC2003)
1 Generating Comparative Summaries of Contradictory Opinions in Text (CIKM09’)Hyun Duk Kim, ChengXiang Zhai 2010/05/24 Yu-wen,Hsu.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Probabilistic Text Structuring: Experiments with Sentence Ordering Mirella Lapata Department of Computer Science University of Sheffield, UK (ACL 2003)
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
1 Centroid Based multi-document summarization: Efficient sentence extraction method Presenter: Chen Yi-Ting.
Acquisition of Categorized Named Entities for Web Search Marius Pasca Google Inc. from Conference on Information and Knowledge Management (CIKM) ’04.
1 Measuring the Semantic Similarity of Texts Author : Courtney Corley and Rada Mihalcea Source : ACL-2005 Reporter : Yong-Xiang Chen.
26/01/20161Gianluca Demartini Ranking Categories for Faceted Search Gianluca Demartini L3S Research Seminars Hannover, 09 June 2006.
LexPageRank: Prestige in Multi-Document Text Summarization Gunes Erkan, Dragomir R. Radev (EMNLP 2004)
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
MMM2005The Chinese University of Hong Kong MMM2005 The Chinese University of Hong Kong 1 Video Summarization Using Mutual Reinforcement Principle and Shot.
The P YTHY Summarization System: Microsoft Research at DUC 2007 Kristina Toutanova, Chris Brockett, Michael Gamon, Jagadeesh Jagarlamudi, Hisami Suzuki,
A Survey on Automatic Text Summarization Dipanjan Das André F. T. Martins Tolga Çekiç
Sample paper in APA style Sample paper in APA style.
1 Summarization CENG 784, Fall 2013 Aykut Erdem. 2 Today Summarization  Motivation  Kinds of summaries  Approaches and paradigms  Evaluating summaries.
2016/9/301 Exploiting Wikipedia as External Knowledge for Document Clustering Xiaohua Hu, Xiaodan Zhang, Caimei Lu, E. K. Park, and Xiaohua Zhou Proceeding.
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
CRF &SVM in Medication Extraction
Challenges in Creating an Automated Protein Structure Metaserver
Summarizing Entities: A Survey Report
Automatic Detection of Causal Relations for Question Answering
Presentation transcript:

Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)

2/23 Abstract Most approaches to extractive summarization define a set of features upon which selection of sentences is based, using algorithms independent of the features We proposed a new set of features based on low- level atomic events that describe relationships between important actors Our experimental results indicate that not only the event-based features offer an improvement in summary quality over words as features, but that this effect is more pronounced for more sophisticated summarization methods

3/23 Introduction Identify what information is important and should be included into the summary Break the input text into textual units (sentences, clauses, etc.) Score every textual unit according to what information is covered in it textual unitsinformation features Choose the textual unit that should be added to the summary repeat until we reach the desired length rescore the textual units based on what information is already covered by the summary

4/23 General Summarization Model T i : text unit I C i : concept i * Each concept c i has an associated weight w i Indicating the importance

5/23 General Summarization Model Using the above matrix it is possible to formulate the extractive summarization problem as extracting the minimal amount of text units which cover all the concepts that are interesting or important To account for the cost of long summaries, we can constrain the total length of the summary or balance it against the total weight of covered concepts

6/23 Associating Concepts with Features Lexical features: Words: tf*idf weights show what words are important Words used in titles and section headings (Luhn’59) Presence of cue phrases in the textual unit: in conclusion, significant (Kupiec et al’95) Co-occurrence of some particular terms: lexical chains (Barzilay & Elhadad’97), topic signatures (Lin & Hovy’2000) Non-lexical features: Textual unit’s position in the input text: head-line, first sentence in the paragraph (Baxedale’58) Rhetorical representation of the source text (Marcu’97) We suggest to use atomic events as features signaling out the important sentences

7/23 Atomic Events Atomic events = Relation + Connector (potential label for the relation) Relation is a pair of Named Entities or significant nouns For the input text, get all possible pairs of named entities within one sentence For every relation analyze all the verbs and action defining nouns in-between the named entities in the relation, these verbs/nouns can be used as labels for the extracted relations Some important words are not marked by name entities but are highly likely to be among the most frequently used nouns Top ten most frequent nouns are added to the relation list

8/23 Atomic Events Algorithm for atomic event extraction Analyze each input sentence one at a time; ignore sentences that do not contain at least two name entities or frequent nouns Extract all the possible pairs (relations) of named entities / frequent nouns in the sentence and words in-between the name entities (connectors) For each relation, count how many times this relation is used in the input texts Keep only connectors that are content verbs or action nouns, according to Word-Net’s noun hierarchy. For each connector calculate how many times it is used for the extracted relations

9/23 Atomic Events Calculate normalized frequencies for all relations Normalized frequency of a relation = n/N, where n – frequency of the current relation in a topic N – overall frequency of all relations in a topic Calculate normalized frequencies for all connectors Normalized frequency of a connector = c/S, where c – frequency of the current connector in a relation S – overall frequency of all connectors for a relation

10/23 Atomic Events Norm. Rel. Frequency First Element Second Element China AirlinesTaiwan China AirlinesTaipei China AirlinesMonday TaiwanMonday BaliTaipei TaipeiTaiwan BaliTaiwan TaipeiMonday International AirportTaiwan

11/23 Atomic Events RelationConnectorNorm. Conn. Freq. China Airlines - Taiwancrashed/VBD trying/VBG burst/VBP land/VB China Airlines - Taipeiburst/VBP crashed/VBD crashed/VBN

12/23 Atomic Events Atomic event score: The score of the atomic event predicts how well the important this atomic event for the collection of texts is Atomic Event Score = Normalized freq. of the relation () * ( Normalized freq. of the connector )

13/23 Textual Unit Selection Static Greedy Algorithm For every textual units, calculate the weight of this text unit as the sum of the weights of all the concepts covered by this textual unit Choose the text unit with maximum weight and add it to the final output Continue extracting other textual units in order of total weight till we get the summary of the desired length

14/23 Textual Unit Selection Adaptive Greedy Algorithm For each textual calculate its weight as the sum of weights of all concepts it covers Choose the textual unit with the maximum weight and add it to the output. Add the concepts covered by this textual unit to the list of concepts covered in the final output. Recalculate the weights of text units: subtract from each unit’s weight the weight of all concepts in it that are already covered in the output Continue extracting text units in order of their total weight (back to step 2) until the summary is of the desired length

15/23 Textual Unit Selection Modified Adaptive Greedy Algorithm For every textual unit calculate its weight as the sum of weights of all concepts it covers Consider only those textual units that contain the concept with the highest weight that has not yet been covered. Out of these, choose the one with highest total weight and add it to the final output. Add the concepts which are covered by this textual unit to the list of concepts covered in the final output. Recalculate the weights of text units: subtract from each unit’s weight the weight of all concepts in it that are already covered in the output Continue extracting text units in order of their total weight (back to step 2) until the summary is of the desired length

16/23 Experiment Input Data The document sets used in the evaluation of multi-document summarization during the first Document Understanding Conference (DUC2001) 30 test document sets, each with approximately 10 news stories on different events For each document set three human-constructed summaries are provided for each of the target lengths 50, 100, 200 and 400 words

17/23 Experiment Evaluation Metric ROUGE (Lin and Hovy, 2003) Recall-based measure Summary Length For each document set, four summaries of length 50, 100, 200 and 400 are created Rouge evaluation has not yet been tested extensively, and ROUGE scores are difficult to interpret as they are not absolute and not comparable across source document sets

18/23 Experiment Results: Static Greedy Algorithm

19/23 Experiment Results: Adaptive Greedy Algorithm

20/23 Experiment

21/23 Experiment Results: Modified Greedy Algorithm

22/23 Experiment Results: Comparison with DUC systems In DUC2003 the task was to create summaries only of length 100 Events as features and adaptive greedy algorithm In 14 out of 30 cases our system outperforms the median of the ROUGE scores of all the 15 participating systems over that specific document set The suitability of the event-based summarizer varies according to the type of documents being summarized

23/23 Conclusion Our experimental results indicate that events are indeed an effective features, at least in comparison with words With all three of our summarization algorithms, we achieved a gain in performance when using events Our approach to defining and extracting events can be improved in many ways Matching connectors that are similar in meaning Representing paraphrases of the same event Methods for detecting and prioritizing special event components such as time and location phrases Merging information across related atomic events Partial matches between atomic events and input sentences