Statistical NLP Spring 2010 Lecture 22: Summarization Dan Klein – UC Berkeley Includes slides from Aria Haghighi, Dan Gillick.

Slides:



Advertisements
Similar presentations
Ani Nenkova Lucy Vanderwende Kathleen McKeown SIGIR 2006.
Advertisements

Query Chain Focused Summarization Tal Baumel, Rafi Cohen, Michael Elhadad Jan 2014.
Sequential Minimal Optimization Advanced Machine Learning Course 2012 Fall Semester Tsinghua University.
Greedy Algorithms.
Unsupervised Modeling of Twitter Conversations
Mauro Sozio and Aristides Gionis Presented By:
1 Lecture 5 Support Vector Machines Large-margin linear classifier Non-separable case The Kernel trick.
Problem Semi supervised sarcasm identification using SASI
Sequence Clustering and Labeling for Unsupervised Query Intent Discovery Speaker: Po-Hsien Shih Advisor: Jia-Ling Koh Source: WSDM’12 Date: 1 November,
Minimum Redundancy and Maximum Relevance Feature Selection
Five Problems CSE 421 Richard Anderson Winter 2009, Lecture 3.
CMPUT 466/551 Principal Source: CMU
Language Model based Information Retrieval: University of Saarland 1 A Hidden Markov Model Information Retrieval System Mahboob Alam Khalid.
A Statistical Model for Domain- Independent Text Segmentation Masao Utiyama and Hitoshi Isahura Presentation by Matthew Waymost.
The ICSI Summarization System Dan Gillick, Benoit Favre, and Dilek Hakkani-Tür {dgillick, favre, International Computer Science.
Lecture 6 Image Segmentation
Search and Retrieval: More on Term Weighting and Document Ranking Prof. Marti Hearst SIMS 202, Lecture 22.
SLIDE 1IS 240 – Spring 2010 Logistic Regression The logistic function: The logistic function is useful because it can take as an input any.
Abstract We present a model of curvilinear grouping using piecewise linear representations of contours and a conditional random field to capture continuity.
Web Projections Learning from Contextual Subgraphs of the Web Jure Leskovec, CMU Susan Dumais, MSR Eric Horvitz, MSR.
Essay assessors CPI 494, April 14, 2009 Kurt VanLehn.
Learning Subjective Adjectives from Corpora Janyce M. Wiebe Presenter: Gabriel Nicolae.
Adapted by Doug Downey from Machine Learning EECS 349, Bryan Pardo Machine Learning Clustering.
Maximum Entropy Model LING 572 Fei Xia 02/07-02/09/06.
Distributed Representations of Sentences and Documents
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Maximum Entropy Model LING 572 Fei Xia 02/08/07. Topics in LING 572 Easy: –kNN, Rocchio, DT, DL –Feature selection, binarization, system combination –Bagging.
Clustering Unsupervised learning Generating “classes”
Information Retrieval in Practice
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
Learning with Positive and Unlabeled Examples using Weighted Logistic Regression Wee Sun Lee National University of Singapore Bing Liu University of Illinois,
1 Data-Driven Dependency Parsing. 2 Background: Natural Language Parsing Syntactic analysis String to (tree) structure He likes fish S NP VP NP VNPrn.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
An Introduction to Machine Learning and Natural Language Processing Tools Presented by: Mark Sammons, Vivek Srikumar (Many slides courtesy of Nick Rizzolo)
Diversity in Ranking via Resistive Graph Centers Avinava Dubey IBM Research India Soumen Chakrabarti IIT Bombay Chiranjib Bhattacharyya IISc Bangalore.
A Machine Learning Approach to Sentence Ordering for Multidocument Summarization and Its Evaluation D. Bollegala, N. Okazaki and M. Ishizuka The University.
Statistical NLP Spring 2011 Lecture 25: Summarization Dan Klein – UC Berkeley TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
LexPageRank: Prestige in Multi- Document Text Summarization Gunes Erkan and Dragomir R. Radev Department of EECS, School of Information University of Michigan.
LARGE MARGIN CLASSIFIERS David Kauchak CS 451 – Fall 2013.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
BLAST: Basic Local Alignment Search Tool Altschul et al. J. Mol Bio CS 466 Saurabh Sinha.
CS654: Digital Image Analysis
Prototype-Driven Learning for Sequence Models Aria Haghighi and Dan Klein University of California Berkeley Slides prepared by Andrew Carlson for the Semi-
QoS Supported Clustered Query Processing in Large Collaboration of Heterogeneous Sensor Networks Debraj De and Lifeng Sang Ohio State University Workshop.
Inference Protocols for Coreference Resolution Kai-Wei Chang, Rajhans Samdani, Alla Rozovskaya, Nick Rizzolo, Mark Sammons, and Dan Roth This research.
Threshold Setting and Performance Monitoring for Novel Text Mining Wenyin Tang and Flora S. Tsai School of Electrical and Electronic Engineering Nanyang.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
KNN & Naïve Bayes Hongning Wang Today’s lecture Instance-based classifiers – k nearest neighbors – Non-parametric learning algorithm Model-based.
Machine Learning Concept Learning General-to Specific Ordering
Probabilistic Text Structuring: Experiments with Sentence Ordering Mirella Lapata Department of Computer Science University of Sheffield, UK (ACL 2003)
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
CSE332: Data Abstractions Lecture 12: Introduction to Sorting Dan Grossman Spring 2010.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
The P YTHY Summarization System: Microsoft Research at DUC 2007 Kristina Toutanova, Chris Brockett, Michael Gamon, Jagadeesh Jagarlamudi, Hisami Suzuki,
Relevance Feedback Prof. Marti Hearst SIMS 202, Lecture 24.
Neural Networks Lecture 4 out of 4. Practical Considerations Input Architecture Output.
A Survey on Automatic Text Summarization Dipanjan Das André F. T. Martins Tolga Çekiç
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Mixture Densities Maximum Likelihood Estimates.
SOFTWARE TESTING LECTURE 9. OBSERVATIONS ABOUT TESTING “ Testing is the process of executing a program with the intention of finding errors. ” – Myers.
Web News Sentence Searching Using Linguistic Graph Similarity
Intro to NLP and Deep Learning
Prototype-Driven Learning for Sequence Models
Data Integration with Dependent Sources
Contributors Jeremy Brown, Bryan Winters, and Austin Ray
CS224N: Query Focused Multi-Document Summarization
Panagiotis G. Ipeirotis Luis Gravano
Presentation transcript:

Statistical NLP Spring 2010 Lecture 22: Summarization Dan Klein – UC Berkeley Includes slides from Aria Haghighi, Dan Gillick

Selection Maximum Marginal Relevance mid-‘90s present Maximize similarity to the query Minimize redundancy [Carbonell and Goldstein, 1998] s1s1s1s1 s3s3s3s3 s2s2s2s2 s4s4s4s4 Q Greedy search over sentences

Maximum Marginal Relevance Graph algorithms [Mihalcea 05++] mid-‘90s present Selection

Maximum Marginal Relevance Graph algorithms mid-‘90s present s1s1s1s1 s3s3s3s3 s2s2s2s2 s4s4s4s4 Nodes are sentences Selection

Maximum Marginal Relevance Graph algorithms mid-‘90s present s1s1s1s1 s3s3s3s3 s2s2s2s2 s4s4s4s4 Nodes are sentences Edges are similarities Selection

Maximum Marginal Relevance Graph algorithms mid-‘90s present s1s1s1s1 s3s3s3s3 s2s2s2s2 s4s4s4s4 Nodes are sentences Edges are similarities Stationary distribution represents node centrality Selection

Maximum Marginal Relevance Graph algorithms Word distribution models mid-‘90s present Input document distributionSummary distribution ~w P A (w) Obama? speech? health? Montana?w P D (w) Obama0.017 speech0.024 health0.009 Montana0.002 Selection

Maximum Marginal Relevance Graph algorithms Word distribution models mid-‘90s present SumBasic [Nenkova and Vanderwende, 2005] Value(w i ) = P D (w i ) Value(s i ) = sum of its word values Choose s i with largest value Adjust P D (w) Repeat until length constraint Selection

Maximum Marginal Relevance Graph algorithms Word distribution models Regression models mid-‘90s present s1s1s1s1 s2s2s2s2 s3s3s3s3 word values positionlength s2s2s2s2 s3s3s3s3 s1s1s1s1 F(x) frequency is just one of many features Selection

Maximum Marginal Relevance Graph algorithms Word distribution models Regression models Topic model-based [Haghighi and Vanderwende, 2009] mid-‘90s present Selection

H & V 09PYTHY

Maximum Marginal Relevance Graph algorithms Word distribution models Regression models Topic models Globally optimal search mid-‘90s present [McDonald, 2007] s1s1s1s1 s3s3s3s3 s2s2s2s2 s4s4s4s4 Q Optimal search using MMR Integer Linear Program Selection

[Gillick and Favre, 2008] Universal health care is a divisive issue. Obama addressed the House on Tuesday. President Obama remained calm. The health care bill is a major test for the Obama administration. s1s1s1s1 s2s2s2s2 s3s3s3s3 s4s4s4s4 conceptvalue Selection

[Gillick and Favre, 2008] Universal health care is a divisive issue. Obama addressed the House on Tuesday. President Obama remained calm. The health care bill is a major test for the Obama administration. conceptvalue obama3 s1s1s1s1 s2s2s2s2 s3s3s3s3 s4s4s4s4 Selection

[Gillick and Favre, 2008] Universal health care is a divisive issue. Obama addressed the House on Tuesday. President Obama remained calm. The health care bill is a major test for the Obama administration. conceptvalue obama3 health2 s1s1s1s1 s2s2s2s2 s3s3s3s3 s4s4s4s4 Selection

[Gillick and Favre, 2008] Universal health care is a divisive issue. Obama addressed the House on Tuesday. President Obama remained calm. The health care bill is a major test for the Obama administration. conceptvalue obama3 health2 house1 s1s1s1s1 s2s2s2s2 s3s3s3s3 s4s4s4s4 Selection

[Gillick and Favre, 2008] Universal health care is a divisive issue. Obama addressed the House on Tuesday. President Obama remained calm. conceptvalue obama3 health2 house1 s1s1s1s1 s2s2s2s2 s3s3s3s3 s4s4s4s4 The health care bill is a major test for the Obama administration.summarylengthvalue {s 1, s 3 }175 {s 2, s 3, s 4 }176 Length limit: 18 words greedy optimal Selection

[Gillick, Riedhammer, Favre, Hakkani-Tur, 2008] total concept value summary length limit maintain consistency between selected sentences and concepts Integer Linear Program for the maximum coverage model Selection

[Gillick and Favre, 2009] This ILP is tractable for reasonable problems Selection

How to include sentence position? First sentences are unique Selection

How to include sentence position? Only allow first sentences in the summary Up-weight concepts appearing in first sentences Identify more sentences that look like first sentences surprisingly strong baseline included in TAC 2009 system first sentence classifier is not reliable enough yet Selection

Some interesting work on sentence ordering [Barzilay et. al., 1997; 2002] But choosing independent sentences is easier First sentences usually stand alone well Sentences without unresolved pronouns Classifier trained on OntoNotes: <10% error rate Baseline ordering module (chronological) is not obviously worse than anything fancier Selection

52 submissions 27 teams 44 topics 10 input docs 100 word summaries Gillick & Favre Rating scale: 1-10 Humans in [8.3, 9.3] Rating scale: 1-10 Humans in [8.5, 9.3] Rating scale: 0-1 Humans in [0.62, 0.77] Rating scale: 0-1 Humans in [0.11, 0.15] Results [G & F, 2009]

[Gillick and Favre, 2008] Error Breakdown?

Sentence extraction is limiting... and boring! But abstractive summaries are much harder to generate… in 25 words? Beyond Extraction?

Brown first saw the new planet Jan. 8 using the 48-inch Samuel Oschin Telescope Sample Errors:... news that the U.K. Panel on Takeover and Mergers... outside the U.S. Amoco already has shed such operations Preprocessing

Sentence Boundary Detection (SBD) is non-trivial because periods are ambiguous (sentence boundaries, abbreviations, or both). Wall Street Journal Corpus

Preprocessing Local context (one word on each side of the period) is sufficient. 80% of sentence-ending abbrs. in the WSJ test set Key innovations: Using an abbreviation list is too coarse.Abbr. Ends sentence TotalRatio Inc Co Corp U.S Calif Ltd