1 I256: Applied Natural Language Processing Marti Hearst Oct 2, 2006.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Academic Writing Writing an Abstract.
Spelling Correction for Search Engine Queries Bruno Martins, Mario J. Silva In Proceedings of EsTAL-04, España for Natural Language Processing Presenter:
Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Comparing Twitter Summarization Algorithms for Multiple Post Summaries David Inouye and Jugal K. Kalita SocialCom May 10 Hyewon Lim.
Evaluating Search Engine
Information Retrieval in Practice
Text Specificity and Impact on Quality of News Summaries Annie Louis & Ani Nenkova University of Pennsylvania June 24, 2011.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Cover Coefficient based Multidocument Summarization CS 533 Information Retrieval Systems Özlem İSTEK Gönenç ERCAN Nagehan PALA.
By : asef poormasoomi autumn Introduction summary: brief but accurate representation of the contents of a document 2.
Overview of Search Engines
Indexing Overview Approaches to indexing Automatic indexing Information extraction.
Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
IMSS005 Computer Science Seminar
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
Newsjunkie: Providing Personalized Newsfeeds via Analysis of Information Novelty Gabrilovich et.al WWW2004.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
1 Text Summarization: News and Beyond Kathleen McKeown Department of Computer Science Columbia University.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
Processing of large document collections Part 7 (Text summarization: multi- document summarization, knowledge- rich approaches, current topics) Helena.
Features and Algorithms Paper by: XIAOGUANG QI and BRIAN D. DAVISON Presentation by: Jason Bender.
Processing of large document collections Part 5 (Text summarization) Helena Ahonen-Myka Spring 2006.
Similar Document Search and Recommendation Vidhya Govindaraju, Krishnan Ramanathan HP Labs, Bangalore, India JOURNAL OF EMERGING TECHNOLOGIES IN WEB INTELLIGENCE.
A Machine Learning Approach to Sentence Ordering for Multidocument Summarization and Its Evaluation D. Bollegala, N. Okazaki and M. Ishizuka The University.
Weighting and Matching against Indices. Zipf’s Law In any corpus, such as the AIT, we can count how often each word occurs in the corpus as a whole =
1 Reference Julian Kupiec, Jan Pedersen, Francine Chen, “A Trainable Document Summarizer”, SIGIR’95 Seattle WA USA, Xiaodan Zhu, Gerald Penn, “Evaluation.
Presenter: Shanshan Lu 03/04/2010
Introduction to Digital Libraries hussein suleman uct cs honours 2003.
LexPageRank: Prestige in Multi- Document Text Summarization Gunes Erkan and Dragomir R. Radev Department of EECS, School of Information University of Michigan.
Comparing and Ranking Documents Once our search engine has retrieved a set of documents, we may want to Rank them by relevance –Which are the best fit.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
1 Web-Page Summarization Using Clickthrough Data* JianTao Sun, Yuchang Lu Dept. of Computer Science TsingHua University Beijing , China Dou Shen,
Information Extraction and Automatic Summarisation *
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
Processing of large document collections Part 5 (Text summarization) Helena Ahonen-Myka Spring 2005.
Summarizing Encyclopedic Term Descriptions on the Web from Coling 2004 Atsushi Fujii and Tetsuya Ishikawa Graduate School of Library, Information and Media.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
A Critique and Improvement of an Evaluation Metric for Text Segmentation A Paper by Lev Pevzner (Harvard University) Marti A. Hearst (UC, Berkeley) Presented.
1 Generating Comparative Summaries of Contradictory Opinions in Text (CIKM09’)Hyun Duk Kim, ChengXiang Zhai 2010/05/24 Yu-wen,Hsu.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Principals of Research Writing. What is Research Writing? Process of communicating your research  Before the fact  Research proposal  After the fact.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
LexPageRank: Prestige in Multi-Document Text Summarization Gunes Erkan, Dragomir R. Radev (EMNLP 2004)
Pastra and Saggion, EACL 2003 Colouring Summaries BLEU Katerina Pastra and Horacio Saggion Department of Computer Science, Natural Language Processing.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
The P YTHY Summarization System: Microsoft Research at DUC 2007 Kristina Toutanova, Chris Brockett, Michael Gamon, Jagadeesh Jagarlamudi, Hisami Suzuki,
A Survey on Automatic Text Summarization Dipanjan Das André F. T. Martins Tolga Çekiç
Knowledge and Information Retrieval Dr Nicholas Gibbins 32/4037.
Abstract  An abstract is a concise summary of a larger project (a thesis, research report, performance, service project, etc.) that concisely describes.
1 Summarization CENG 784, Fall 2013 Aykut Erdem. 2 Today Summarization  Motivation  Kinds of summaries  Approaches and paradigms  Evaluating summaries.
Natural Language Processing Vasile Rus
Information Retrieval in Practice
Queensland University of Technology
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Applying Key Phrase Extraction to aid Invalidity Search
CS 430: Information Discovery
Presentation transcript:

1 I256: Applied Natural Language Processing Marti Hearst Oct 2, 2006

2 From lecture notes by Nachum Dershowitz & Dan Cohen Contents Introduction and Applications Types of summarization tasks Basic paradigms Single document summarization Evaluation methods

3 From lecture notes by Nachum Dershowitz & Dan Cohen Introduction The problem – Information overload –4 Billion URLs indexed by Google –200 TB of data on the Web [Lyman and Varian 03] –Information is created every day in enormous amounts One solution – summarization Abstracts promote current awareness save reading time facilitate selection facilitate literature searches aid in the preparation of reviews But what is an abstract??

4 From lecture notes by Nachum Dershowitz & Dan Cohen abstract: brief but accurate representation of the contents of a document goal: take an information source, extract the most important content from it and present it to the user in a condensed form and in a manner sensitive to the user’s needs. compression: the amount of text to present or the length of the summary to the length of the source. Introduction

5 From lecture notes by Nachum Dershowitz & Dan Cohen The problem has been addressed since the 50’s[Luhn 58] Numerous methods are currently being suggested Most methods still rely on 50’s-70’s algorithms Problem is still hard yet there are some applications: MS Word, by Drago Radev’s research group History

6 From lecture notes by Nachum Dershowitz & Dan Cohen

7 MSWord AutoSummarize

8 From lecture notes by Nachum Dershowitz & Dan Cohen Applications Abstracts for Scientific and other articles News summarization (mostly multiple document summarization) Classification of articles and other written data Web pages for search engines Web access from PDAs, Cell phones Question answering and data gathering

9 From lecture notes by Nachum Dershowitz & Dan Cohen Types of Summaries Indicative vs Informative Informative: a substitute for the entire document Indicative: give an idea of what is there Background Does the reader have the needed prior knowledge? Expert reader vs Novice reader Query based or General Query based – a form is being filled, answers should be answered General – General purpose summarization

10 From lecture notes by Nachum Dershowitz & Dan Cohen Types of Summaries (input) Single document vs multiple documents Domain specific (chemistry) or general Genre specific (newspaper items) of general

11 From lecture notes by Nachum Dershowitz & Dan Cohen Types of Summaries (output) extract vs abstract Extracts – representative paragraphs/sentences/ phrases/words, fragments of the original text Abstracts – a concise summary of the central subjects in the document. Research shows that sometimes readers prefer Extracts! language chosen for summarization format of the resulting summary (table/paragraph/key words)

12 From lecture notes by Nachum Dershowitz & Dan Cohen Methods Quantitative heuristics, manually scored Machine-learning based statistical scoring methods Higher semantic/syntactic structures Network (graph) based methods Other methods (rhetorical analysis, lexical chains, co-reference chains) AI methods

13 From lecture notes by Nachum Dershowitz & Dan Cohen Quantitative Heuristics General method: score each entity (sentence, word) ; combine scores; choose best sentence(s) Scoring techniques: Word frequencies throughout the text (Luhn 58) Position in the text (Edmunson 69, Lin&Hovy 97) Title method (Edmunson 69) Cue phrases in sentences (Edmunson 69)

14 From lecture notes by Nachum Dershowitz & Dan Cohen Using Word Frequencies (Luhn 58) Very first work in automated summarization Assumptions: Frequent words indicate the topic Frequent means with reference to the corpus frequency Clusters of frequent words indicate summarizing sentence Stemming based on similar prefix characters Very common words and very rare words are ignored

15 Ranked Word Frequency Zipf’s curve

16 From lecture notes by Nachum Dershowitz & Dan Cohen Find consecutive sequences of high- weight keywords Allow a certain number of gaps of low-weight terms Sentences with highest sum of cluster weights are chosen Word frequencies (Luhn 58)

17 From lecture notes by Nachum Dershowitz & Dan Cohen Claim : Important sentences occur in specific positions “lead-based” summary inverse of position in document works well for the “news” Important information occurs in specific sections of the document (introduction/conclusion) Position in the text (Edmunson 69)

18 From lecture notes by Nachum Dershowitz & Dan Cohen Claim : title of document indicates its content Unless editors are being cute Not true for novels usually What about blogs …? words in title help find relevant content create a list of title words, remove “stop words” Use those as keywords in order to find important sentences (for example with Luhn’s methods) Title method (Edmunson 69)

19 From lecture notes by Nachum Dershowitz & Dan Cohen method (Edmunson 69) Cue phrases method (Edmunson 69) Claim : Important sentences contain cue words/indicative phrases “The main aim of the present paper is to describe…” (IND) “The purpose of this article is to review…” (IND) “In this report, we outline…” (IND) “Our investigation has shown that…” (INF) Some words are considered bonus others stigma bonus: comparatives, superlatives, conclusive expressions, etc. stigma: negatives, pronouns, etc. Claim : Important sentences contain cue words/indicative phrases “The main aim of the present paper is to describe…” (IND) “The purpose of this article is to review…” (IND) “In this report, we outline…” (IND) “Our investigation has shown that…” (INF) Some words are considered bonus others stigma bonus: comparatives, superlatives, conclusive expressions, etc. stigma: negatives, pronouns, etc.

20 From lecture notes by Nachum Dershowitz & Dan Cohen Linear contribution of 4 features title, cue, keyword, position the weights are adjusted using training data with any minimization technique Evaluated on a corpus of 200 chemistry articles Length ranged from 100 to 3900 words Judges were told to extract 25% of the sentences, to maximize coherence, minimize redundancy. Features –Position (sensitive to types of headings for sections) –cue –title –keyword Best results obtained with: –cue + title + position Feature combination ( Edmundson ’69)

21 From lecture notes by Nachum Dershowitz & Dan Cohen Statistical learning method Feature set sentence length –|S| > 5 fixed phrases –26 manually chosen paragraph –sentence position in paragraph thematic words –binary: whether sentence is included in manual extract uppercase words –not common acronyms Corpus –188 document + summary pairs from scientific journals Bayesian Classifier (Kupiec at el 95)

22 From lecture notes by Nachum Dershowitz & Dan Cohen Uses Bayesian classifier: Assuming statistical independence: Bayesian Classifier (Kupiec at el 95)

23 From lecture notes by Nachum Dershowitz & Dan Cohen Each Probability is calculated empirically from a corpus Higher probability sentences are chosed to be in the summary Performance: For 25% summaries, 84% precision Bayesian Classifier (Kupiec at el 95)

24 From lecture notes by Nachum Dershowitz & Dan Cohen When a manual summary is available: 1. choose a granularity (clause; sentence; paragraph), 2. create a similarity measure for that granularity (word overlap; multi-word overlap, perfect match), 3. measure the similarity of each unit in the new to the most similar unit(s) 4. measure Recall and Precision. Otherwise 1. Intrinsic –how good is the summary as a summary? 2. Extrinsic – how well does the summary help the user? Evaluation methods

25 From lecture notes by Nachum Dershowitz & Dan Cohen Intrinsic measures (glass-box): how good is the summary as a summary? Problem: how do you measure the goodness of a summary? Studies: compare to ideal (Edmundson, 69; Kupiec et al., 95; Salton et al., 97; Marcu, 97) or supply criteria— fluency, informativeness, coverage, etc. (Brandow et al., 95). Summary evaluated on its own or comparing it with the source Is the text cohesive and coherent? Does it contain the main topics of the document? Are important topics omitted? Intrinsic measures

26 From lecture notes by Nachum Dershowitz & Dan Cohen (Black box): how well does the summary help a user with a task? Problem: does summary quality correlate with performance? Studies: GMAT tests (Morris et al., 92); news analysis (Miike et al. 94); IR (Mani and Bloedorn, 97); text categorization (SUMMAC 98; Sundheim, 98). Evaluation in an specific task Can the summary be used instead of the document? Can the document be classified by reading the summary? Can we answer questions by reading the summary? Extrinsic measures

27 The Document Understanding Conference (DUC) This is really the Text Summarization Competition Started in 2001 Task and Evaluation (for ): Various target sizes were used ( words) Both single and multiple-document summaries assessed Summaries were manually judged for both content and readability. Each peer (human or automatic) summary was compared against a single model summary –using SEE ( cyl/SEE/) –estimates the percentage of information in the model thatwas covered in the peer. –Also used ROUGE (Lin ’04) in 2004  Recall-Oriented Understudy for Gisting Evaluation  Uses counts of n-gram overlap between candidate and gold-standard summary, assumes fixed-length summaries

28 The Document Understanding Conference (DUC) Made a big change in 2005 Extrinsic evaluation proposed but rejected (write a natural disaster summary) Instead: a complex question-focused summarization task that required summarizers to piece together information from multiple documents to answer a question or set of questions as posed in a DUC topic. Also indicated a desired granularity of information

29 The Document Understanding Conference (DUC) Evaluation metrics for new task: Grammaticality Non-redundancy Referential clarity Focus Structure and Coherence Responsiveness (content-based evaluation) This was a difficult task to do well in.

30 Let’s make a summarizer! Each person (or pair) write code for one small part of the problem, using Kupiec et al’s method. We’ll combine the parts in class.

31 Next Time More on Bayesian classification Other summarization approaches (Marcu paper) Multi-document summarization (Goldstein et al. paper) In-class summarizer!