Methods for Automatic Evaluation of Sentence Extract Summaries * G.Ravindra +, N.Balakrishnan +, K.R.Ramakrishnan * Supercomputer Education & Research.

Slides:



Advertisements
Similar presentations
An Ontology Creation Methodology: A Phased Approach
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
Towards Twitter Context Summarization with User Influence Models Yi Chang et al. WSDM 2013 Hyewon Lim 21 June 2013.
Comparing Twitter Summarization Algorithms for Multiple Post Summaries David Inouye and Jugal K. Kalita SocialCom May 10 Hyewon Lim.
Automatically Annotating and Integrating Spatial Datasets Chieng-Chien Chen, Snehal Thakkar, Crail Knoblock, Cyrus Shahabi Department of Computer Science.
Common Core State Standards (CCSS) Nevada Joint Union High School District Nevada Union High School September 23, 2013 Louise Johnson, Ed.D. Superintendent.
In Search of a More Probable Parse: Experiments with DOP* and the Penn Chinese Treebank Aaron Meyers Linguistics 490 Winter 2009.
Evaluating Texts 2D summary... précis... explication...
A New Suffix Tree Similarity Measure for Document Clustering Hung Chim, Xiaotie Deng City University of Hong Kong WWW 2007 Session: Similarity Search April.
Evaluating Search Engine
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
A Flexible Workbench for Document Analysis and Text Mining NLDB’2004, Salford, June Gulla, Brasethvik and Kaada A Flexible Workbench for Document.
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
FACE RECOGNITION, EXPERIMENTS WITH RANDOM PROJECTION
1 Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang, Assistant Professor Dept. of Computer Science & Information Engineering National Central.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Presented by Zeehasham Rasheed
Article by: Feiyu Xu, Daniela Kurz, Jakub Piskorski, Sven Schmeier Article Summary by Mark Vickers.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Flash talk by: Aditi Garg, Xiaoran Wang Authors: Sarah Rastkar, Gail C. Murphy and Gabriel Murray.
Indexing Overview Approaches to indexing Automatic indexing Information extraction.
On Error Preserving Encryption Algorithms for Wireless Video Transmission Ali Saman Tosun and Wu-Chi Feng The Ohio State University Department of Computer.
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
Introduction.  Classification based on function role in classroom instruction  Placement assessment: administered at the beginning of instruction 
Mining and Summarizing Customer Reviews
Title Extraction from Bodies of HTML Documents and its Application to Web Page Retrieval Microsoft Research Asia Yunhua Hu, Guomao Xin, Ruihua Song, Guoping.
1 Wikification CSE 6339 (Section 002) Abhijit Tendulkar.
Using Text Mining and Natural Language Processing for Health Care Claims Processing Cihan ÜNAL
Scott Duvall, Brett South, Stéphane Meystre A Hands-on Introduction to Natural Language Processing in Healthcare Annotation as a Central Task for Development.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
 Text Representation & Text Classification for Intelligent Information Retrieval Ning Yu School of Library and Information Science Indiana University.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
From Social Bookmarking to Social Summarization: An Experiment in Community-Based Summary Generation Oisin Boydell, Barry Smyth Adaptive Information Cluster,
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
Research Topics CSC Parallel Computing & Compilers CSC 3990.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Introduction Use machine learning and various classifying techniques to be able to create an algorithm that can decipher between spam and ham s. .
Translation Memory System (TMS)1 Translation Memory Systems Presentation by1 Melina Takanen & Julianna Ekert CAT Prof. Thorsten Trippel University.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Algorithmic Detection of Semantic Similarity WWW 2005.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Processing of large document collections Part 5 (Text summarization) Helena Ahonen-Myka Spring 2005.
2005/12/021 Content-Based Image Retrieval Using Grey Relational Analysis Dept. of Computer Engineering Tatung University Presenter: Tienwei Tsai ( 蔡殿偉.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Collocations and Terminology Vasileios Hatzivassiloglou University of Texas at Dallas.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Finding frequent and interesting triples in text Janez Brank, Dunja Mladenić, Marko Grobelnik Jožef Stefan Institute, Ljubljana, Slovenia.
A Classification-based Approach to Question Answering in Discussion Boards Liangjie Hong, Brian D. Davison Lehigh University (SIGIR ’ 09) Speaker: Cho,
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
Generating Query Substitutions Alicia Wood. What is the problem to be solved?
Intelligent Database Systems Lab N.Y.U.S.T. I. M. 1 Mining knowledge from natural language texts using fuzzy associated concept mapping Presenter : Wu,
A Patent Document Retrieval System Addressing Both Semantic and Syntactic Properties Liang Chen*,Naoyuki Tokuda+, Hisahiro Adachi+ *University of Northern.
8 December 1997Industry Day Applications of SuperTagging Raman Chandrasekar.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
LexPageRank: Prestige in Multi-Document Text Summarization Gunes Erkan, Dragomir R. Radev (EMNLP 2004)
Pastra and Saggion, EACL 2003 Colouring Summaries BLEU Katerina Pastra and Horacio Saggion Department of Computer Science, Natural Language Processing.
Refined Online Citation Matching and Adaptive Canonical Metadata Construction CSE 598B Course Project Report Huajing Li.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Enhancing Text Clustering by Leveraging Wikipedia Semantics.
A Survey on Automatic Text Summarization Dipanjan Das André F. T. Martins Tolga Çekiç
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
MCAS-Alt Teacher Consultants
Dr Anie Attan 26 April 2017 Language Academy UTMJB
Linguistic Graph Similarity for News Sentence Searching
A German Corpus for Similarity Detection
Web News Sentence Searching Using Linguistic Graph Similarity
Presentation transcript:

Methods for Automatic Evaluation of Sentence Extract Summaries * G.Ravindra +, N.Balakrishnan +, K.R.Ramakrishnan * Supercomputer Education & Research Center + Department of Electrical Engineering * Indian Institute of Science Bangalore-INDIA

Agenda Introduction to Text Summarization Need for summarization, types of summaries Evaluating Extract Summaries Challenges in manual and automatic evaluation Fuzzy Summary Evaluation Complexity Scores

What is Text Summarization Reductive transformation of source text to summary text by content generalization and/or selection Loss of information What can be lost and what should not be lost How much can be lost What is the size of the summary Types of Summaries Extracts and Abstracts Influence of genre on the performance of a summarization algorithm Newswire stories are favorable to sentence position

Need for Summarization Explosive growth in availability of digital textual data Books in digital libraries, mailing-list archives, on-line news portals Duplication of textual segments in books E.g.: 10 introductory books on quantum physics have a number of paragraphs common to all of them (syntactically different but semantically the same) Hand-held devices Small screens and limited memory Low power devices and hence limited processing capability E.g.: Stream a book from a digital library to a hand-held device Production of information is faster than consumption

Types of Summaries Extracts Text selection E.g: Paragraphs from books, sentences from editorials, phrases from s Application of statistical techniques Abstracts Text selection followed by generalization Need for linguistic processing E.g.: Convert a sentence to a phrase Generic Summaries Independent of genre Indicative Summaries Gives a general idea as to the topic of discussion in the text being summarized Informational Summaries Serves as a surrogate to the original text

Evaluating Extract Summaries Manual evaluation Human judges are allowed to score a summary on a well defined scale based on a well defined criteria Subject to judge’s understanding of the subject Depends on judge’s opinions Guidelines constrain opinions Individual judges’ scores are combined to generate the final score Re-evaluation might result in different scores Logistic problems for researchers

Automatic Evaluation Machine-based evaluation Consistent over multiple runs Fast, avoids logistic problems Suitable for researchers experimenting with new algorithms Flip-side Not as accurate as human evaluation Should be used as precursor to a detailed human evaluation Algorithmically handles various sentence constructs and linguistic variants

Fuzzy Summary Evaluation: FuSE Proposing the use of Fuzzy union theory to quantify the similarity of two extract summaries Similarity between the reference (human generated) summary and candidate (machine generated) summary is evaluated Each sentence is a fuzzy set Each sentence in the reference summary has a membership grade in every sentence of the candidate machine generated summary Membership grade of a reference summary sentence in the candidate summary is the union of membership grades across all candidate summary sentences Use membership grades to compute an f-score value Membership grade is the hamming distance between two sentences based on collocations

Fuzzy F-score Candidate summary sentence set Union function Reference summary sentence set Membership grade of candidate sentence in reference sentence Fuzzy Precision Fuzzy Recall

Choice of Union operator Propose the use of Frank’s S-norm operator Allows combining partial matches non- linearly Membership grade of a sentence in a summary is dependent on its length Automatically includes brevity-bonus into the scheme

Frank’s S-norm operator Damping Coefficient Mean of non-zero membership grades for a sentence Sentence length Length of the longest sentence

Characteristics of Frank’s base

Performance of FuSE for various sentence lengths

Dictionary-enhanced Fuzzy Summary Evaluation:DeFuSE FuSE does not understand sentence similarity based on synonymy and hypernymy Identifying synonymous words makes evaluation more accurate Identifying hypernymous word relationships allows consideration of “gross information” during evaluation Note: Very deep hypernymy trees could result in topic drift and hence improper evaluation

Use of Word Net

Example: Use of hypernymy HURRICANE GILBERT DEVASTATED DOMINICAN REPUBLIC AND PARTS OF CUBA (PHYSICAL PHENOMENON) GILBERT (DESTROY,RUIN) (REGION) AND PARTS OF (REGION) TROPICAL STORM GILBERT DESTROYED PARTS OF HAVANA TROPICAL (PHYSICAL PHENOMENON) GILBERT DESTROYED PARTS OF (REGION)

Complexity Score Attempts to quantify the summarization algorithm based on the difficulty in generating a summary of a particular accuracy Generating a 9 sentence summary from a 10 sentence document is very easy. An algorithm which randomly selects 9 sentences will have a worst case accuracy of 90% A complicated AI+NLP based algorithm cannot do any better If a 2 sentence summary is to be generated from a 10 sentence document, we have 45 possible candidates out of which one is accurate

Computing Complexity Score Probability of generating a summary of a length m1 with accurate sentences l 1 when human summary has h sentences and the document being summarized has n sentences

Complexity Score (Cont..) To compare two summaries of equal length the performance of one relative to the baseline is given by

Complexity Score (Cont..) Complexity in generating a 10% extract with 12 correct sentences is higher than generating a 30% extract with 12 correct sentences

Conclusion Summary evaluation is as complicated as summary generation Fuzzy schemes are ideal for evaluating extract summaries Use of synonymy and hypernymy relations improve evaluation accuracy Complexity score is a new way of looking at summary evaluation