1 Towards Automated Related Work Summarization (ReWoS) HOANG Cong Duy Vu 03/12/2010.

Slides:



Advertisements
Similar presentations
Academic Writing Writing an Abstract.
Advertisements

HOW TO WRITE AN ACADEMIC PAPER
Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser ICL, U. Heidelberg CIS, LMU München Statistical Machine Translation.
Towards Twitter Context Summarization with User Influence Models Yi Chang et al. WSDM 2013 Hyewon Lim 21 June 2013.
Project Proposal.
Computational Models of Discourse Analysis Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Instructions for completing the ES089g term paper.
Playing the Telephone Game: Determining the Hierarchical Structure of Perspective and Speech Expressions Eric Breck and Claire Cardie Department of Computer.
Made with OpenOffice.org 1 Sentiment Classification using Word Sub-Sequences and Dependency Sub-Trees Pacific-Asia Knowledge Discovery and Data Mining.
Vikas BhardwajColumbia University NLP for the Web – Spring 2010 Improving QA Accuracy by Question Inversion Prager et al. IBM T.J. Watson Res. Ctr. 02/18/2010.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
MANISHA VERMA, VASUDEVA VARMA PATENT SEARCH USING IPC CLASSIFICATION VECTORS.
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Advanced Research Methodology
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
An Automatic Segmentation Method Combined with Length Descending and String Frequency Statistics for Chinese Shaohua Jiang, Yanzhong Dang Institute of.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Title Extraction from Bodies of HTML Documents and its Application to Web Page Retrieval Microsoft Research Asia Yunhua Hu, Guomao Xin, Ruihua Song, Guoping.
Learning Information Extraction Patterns Using WordNet Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield,
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
IMSS005 Computer Science Seminar
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Presented by Tienwei Tsai July, 2005
National Institute of Informatics Kiyoko Uchiyama 1 A Study for Introductory Terms in Logical Structure of Scientific Papers.
Advanced Signal Processing 05/06 Reinisch Bernhard Statistical Machine Translation Phrase Based Model.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
Leveraging Reusability: Cost-effective Lexical Acquisition for Large-scale Ontology Translation G. Craig Murray et al. COLING 2006 Reporter Yong-Xiang.
Processing of large document collections Part 7 (Text summarization: multi- document summarization, knowledge- rich approaches, current topics) Helena.
Querying Structured Text in an XML Database By Xuemei Luo.
Experimental Research Methods in Language Learning Chapter 16 Experimental Research Proposals.
April 14, 2003Hang Cui, Ji-Rong Wen and Tat- Seng Chua 1 Hierarchical Indexing and Flexible Element Retrieval for Structured Document Hang Cui School of.
A Machine Learning Approach to Sentence Ordering for Multidocument Summarization and Its Evaluation D. Bollegala, N. Okazaki and M. Ishizuka The University.
CS 533 Information Retrieval Systems.  Introduction  Connectivity Analysis  Kleinberg’s Algorithm  Problems Encountered  Improved Connectivity Analysis.
Summarizing Conversations with Clue Words Giuseppe Carenini Raymond T. Ng Xiaodong Zhou Department of Computer Science Univ. of British Columbia.
LexPageRank: Prestige in Multi- Document Text Summarization Gunes Erkan and Dragomir R. Radev Department of EECS, School of Information University of Michigan.
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
Opinion Holders in Opinion Text from Online Newspapers Youngho Kim, Yuchul Jung and Sung-Hyon Myaeng Reporter: Chia-Ying Lee Advisor: Prof. Hsin-Hsi Chen.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
LOGO Summarizing Conversations with Clue Words Giuseppe Carenini, Raymond T. Ng, Xiaodong Zhou (WWW ’07) Advisor : Dr. Koh Jia-Ling Speaker : Tu.
1 KINDS OF PARAGRAPH. There are at least seven types of paragraphs. Knowledge of the differences between them can facilitate composing well-structured.
Processing of large document collections Part 5 (Text summarization) Helena Ahonen-Myka Spring 2005.
DOCUMENT UPDATE SUMMARIZATION USING INCREMENTAL HIERARCHICAL CLUSTERING CIKM’10 (DINGDING WANG, TAO LI) Advisor: Koh, Jia-Ling Presenter: Nonhlanhla Shongwe.
Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
A Critique and Improvement of an Evaluation Metric for Text Segmentation A Paper by Lev Pevzner (Harvard University) Marti A. Hearst (UC, Berkeley) Presented.
AN EFFECTIVE STATISTICAL APPROACH TO BLOG POST OPINION RETRIEVAL Ben He Craig Macdonald Iadh Ounis University of Glasgow Jiyin He University of Amsterdam.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Research Methodology Class.   Your report must contains,  Abstract  Chapter 1 - Introduction  Chapter 2 - Literature Review  Chapter 3 - System.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Principals of Research Writing. What is Research Writing? Process of communicating your research  Before the fact  Research proposal  After the fact.
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
LexPageRank: Prestige in Multi-Document Text Summarization Gunes Erkan, Dragomir R. Radev (EMNLP 2004)
Abstracting.  An abstract is a concise and accurate representation of the contents of a document, in a style similar to that of the original document.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Meta-Path-Based Ranking with Pseudo Relevance Feedback on Heterogeneous Graph for Citation Recommendation By: Xiaozhong Liu, Yingying Yu, Chun Guo, Yizhou.
Sample paper in APA style Sample paper in APA style.
Glossary of Terms Used in Science Papers AS
Presentation transcript:

1 Towards Automated Related Work Summarization (ReWoS) HOANG Cong Duy Vu 03/12/2010

2 Outline Recall A Motivating Example The Proposed Approach –General Content Summarization (GCSum) –Specific Content Summarization (SCSum) –Generation Experiments & Results Future Work Conclusion

3 Recall A set of articles Topic hierarchy tree RW Summarizer A desired lengthUser [] [,] [] A RW summary assumption RW: related work

4 A Motivating Example A related work section extracted from “Bilingual Topic Aspect Classification with A few Training Examples” (Wu et al., 2008)

5 The Proposed Approach The ReWoS architecture, Decision edges are labeled as (T)rue, (F)alse or (R)elevant. For leaf nodes For internal nodes

6 The Proposed Approach Pre-Processing –Based on heuristic rules of sentence length and lexical clues Sentences with token-based length is too short ( 80) Sentences referring to future tenses Sentences containing obviously redundant clues such as: “in the section...”, “figure XXX shows...”, “for instance” …

7 The Proposed Approach Agent-based rule –Attempts to distinguish whether the sentence describes an author’s own work or not. –Based on the presence of tokens that signals work done by the author, such as “we”, “our”, “us”, “this approach”, and “this method” … –Says that if a sentence does not satisfy this rule, route for GCSum, otherwise for SCSum

8 General Content Summarization (GCSum) The objective of GCSum is to extract sentences containing useful background information on the topics of the internal node in focus.

9 General Content Summarization (GCSum) General content informativeindicative 1)Text classification is a task that assigns a certain number of pre-defined labels for a given text. 2)Statistical machine translation (SMT) seeks to develop mathematical models of the translation process whose parameters can be automatically estimated from a parallel corpus. 1)Many previous studies have approached the problem of mono-lingual text classification. 2)This paper refers to the problem of sentiment analysis.

10 General Content Summarization (GCSum) Informative sentences –Give detail on a specific aspect of the problem, e.g. definitions, purpose or application of the topic Indicative sentences –simpler, inserted to make the topic transition explicit and rhetorically sound Summarization issue –Given a topic: For indicative sentences, using pre-defined templates For informative sentences, extract from input articles

11 General Content Summarization (GCSum) GCSum first checks the subject of each candidate sentence, filtering ones whose subjects do not contain at least one topic keyword. (Subject-based rule) Or GCSum checks whether stock verb phrases (i.e., “based on”, “make use of” and 23 other patterns) are used as the main verb. (Verb-based rule) Or GCSum checks for the presence of at least one citation – general sentences may list a set of citations as examples. (Citation-based rule) Importantly note that if cannot find out any informative sentences from input articles, generate indicative sentences instead!

12 General Content Summarization (GCSum) Topic relevance computation (GCSum) –ranks sentences based on keyword content –states that the topic of an internal node is affected by its surrounding nodes – ancestor, descendants and others - score S is the final relevance score - score S QA, score S Q, and score S QR mean the component relevance score of the sentence S with respect to the ancestor, current or other remaining nodes,respectively.

13 General Content Summarization (GCSum) Topic relevance computation (GCSum) The linear combination: S’( ) = S( ) + S( ) – S(5 x ) ancestors others 2 The maximum number of sentences for each intermediate node is itself ancestors itself others

14 General Content Summarization (GCSum) To obtain each component relevance score, we employ TF×ISF relevance computation

15 Specific Content Summarization (SCSum) Sentences that are marked with author-as-agent are input to the Specific Content Summarization (SCSum) module. SCSum aims to extract sentences that contain detailed information about a specific author’s work that is relevant to the input leaf nodes’ topic.

16 Specific Content Summarization (SCSum) Topic relevance computation (SCSum) The linear combination: S’( ) = S( + ) + S( ) – S( ) 2 1 ancestors siblings Initially, the number of sentences for each leaf node is assigned equivalently. The relevance score is computed using the formula similar to GCSum presented earlier. 2 1 itself ancestors itself siblings 4 423

17 Specific Content Summarization (SCSum) Context modeling –Motivation: single sentences occasionally do not contain enough context to clearly express the idea mentioned in original articles –Try to use the contexts to increase the confidence of agent-based sentences score(sentence) score(contexts) + topic final_score(sentence)

18 SCSum - Context modeling ***We evaluated the accuracy of each of the paraphrases that was extracted from the manually aligned data, as well as the top ranked paraphrases from the experimental conditions detailed below in Section 3.3. ***Because the accuracy of paraphrases can vary depending on context, we substituted each set of candidate paraphrases into between 2-10 sentences which contained the original phrase. ***Figure 4 shows the paraphrases for under control substituted into one of the sentences in which it occurred. ***We created a total of 289 such evaluation sets, with a total of 1366 unique sentences created through substitution. ***We had two native English speakers produce judgments as to whether the new sentences preserved the meaning of the original phrase and as to whether they remained grammatical. ***Paraphrases that were judged to preserve both meaning and grammaticality were considered to be correct, and examples which failed on either judgment were considered to be incorrect. Example extracted from (Bannard and Callison-Burch 2005) Adjacent sentences Agent- based sentence ***(Bannard and Callison-Burch 2005) replaced phrases with paraphrases in a number of sentences and asked judges whether the substitutions “preserved meaning and remained grammatical.” Summary sentence

19 Specific Content Summarization (SCSum) Context modeling –Choose nearby sentences within a contextual window (size 5) after the agent-based sentence to represent more for given topic.

20 Specific Content Summarization (SCSum) Weighting –The observation is that the presence of one or more of current, ancestor and sibling nodes may affect the final score from the computation –Add a new weighting coefficient for the score computed from the topic relevance computation (SCSum) a weighting coefficient that takes on differing values based on the presence of keywords in the sentence Values as follows: If sentence contains no keywords in siblings: + Keywords in both ancestors & itself  1 + Keywords in itself only  Keywords in ancestors only  0.25 If sentence contains keywords in siblings  0.1 (penalty)

21 Specific Content Summarization (SCSum) Ranking & Re-ranking –Sentences are ranked descendingly according to their relevance scores –Then, simplified MMR (SimRank) is performed: A sentence X is removed if it has the maximum cosine similarity value exceeding a pre-defined threshold (0.75) with any sentence Y which is already chosen at previous steps of SimRank.

22 Post-Proccessing Two steps: –First, replace agentive forms (e.g., “we”, “our”, “this study”,...) with a citation to the articles –Second, resolves abbreviations found in the extracted sentences E.g. SMT  Statistical Machine Translation

23 Generation In this work, we only generate the related work summaries by using depth-first traversals to form the ordering of topic nodes in a topic tree 1 − 4 −2 − 3 − 5 − 6 − 7 Node ordering

24 Experiments & Results Dataset –Use RWSData described before, including 20 sets 10 out of 20 sets were evaluated automatically and manually. Baselines –LEAD (title + abstract – based RW) –MEAD (centroid + cosine similarity) Proposed systems –ReWoS-WCM (ReWoS without context modeling) –ReWoS-CM (ReWoS with context modeling)

25 Experiments & Results Automatic evaluation –Use ROUGE variants (ROUGE-1, ROUGE-2, ROUGE-S4, ROUGE-SU4) Manual evaluation (measure over 5-point scale of 1 (very poor) to 5 (very good) –Correctness: Is the summary content actually relevant to the hierarchical topics given? –Novelty: Does the summary introduce novel information that is significant in comparison with the human created summary? –Fluency: Does the summary’s exposition flow well, in terms of syntax as well as discourse? –Usefulness: Is the summary acceptable in terms of its usefulness in supporting the researchers to quickly grasp the related works relevant to hierarchical topics given? Summary length: 1% of the original relevant articles, measured in sentences

26 Experiments & Results - ROUGE evaluation seems to work unreasonably when dealing with verbose summaries, often produced by MEAD. - Related work summaries are multi-topic summaries of multi-article references. This may cause miscalculation from overlapping n-grams that occur across multiple topics or references.

27 Experiments & Results - The table shows that both ReWoS–WCM and ReWoS-CM perform significantly better than baseline in terms of correctness, novelty, and usefulness. - Comparing with LEAD, showing that necessary information is not only located in titles or abstracts, but also in relevant portions of the research article body. - ReWoS–CM (with context modeling) performed equivalent to ReWoS–WCM (without it) in terms of correctness and usefulness. - For novelty, ReWoS–CM is better than ReWoS–WCM. It proved that the proposed component of context moding is useful in providing new information.

28 Future work Overcome the assumption about topic hierarchy tree Investigate better generation –Focus on local coherence and topic transition

29 Conclusion According to the best of our knowledge, automated related work summarization has not been studied before. This work took initial steps towards solving this problem, by dividing the task into general and specific summarization processes. Initial results showed an improvement over generic multi-document baselines in both automatic and human evaluation.

30 Thank you! Questions???