Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

1 Opinion Summarization Using Entity Features and Probabilistic Sentence Coherence Optimization (UIUC at TAC 2008 Opinion Summarization Pilot) Nov 19,
Query Chain Focused Summarization Tal Baumel, Rafi Cohen, Michael Elhadad Jan 2014.
SINAI-GIR A Multilingual Geographical IR System University of Jaén (Spain) José Manuel Perea Ortega CLEF 2008, 18 September, Aarhus (Denmark) Computer.
Overview of the KBP 2013 Slot Filler Validation Track Hoa Trang Dang National Institute of Standards and Technology.
COMP423 Intelligent Agents. Recommender systems Two approaches – Collaborative Filtering Based on feedback from other users who have rated a similar set.
NYU ANLP-00 1 Automatic Discovery of Scenario-Level Patterns for Information Extraction Roman Yangarber Ralph Grishman Pasi Tapanainen Silja Huttunen.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Web Document Clustering: A Feasibility Demonstration Hui Han CSE dept. PSU 10/15/01.
Dialogue – Driven Intranet Search Suma Adindla School of Computer Science & Electronic Engineering 8th LANGUAGE & COMPUTATION DAY 2009.
Explorations in Tag Suggestion and Query Expansion Jian Wang and Brian D. Davison Lehigh University, USA SSM 2008 (Workshop on Search in Social Media)
IR & Metadata. Metadata Didn’t we already talk about this? We discussed what metadata is and its types –Data about data –Descriptive metadata is external.
ITCS 6010 Natural Language Understanding. Natural Language Processing What is it? Studies the problems inherent in the processing and manipulation of.
Information Retrieval and Extraction 資訊檢索與擷取 Chia-Hui Chang National Central University
Employing Two Question Answering Systems in TREC 2005 Harabagiu, Moldovan, et al 2005 Language Computer Corporation.
Information Retrieval in Practice
Mining and Summarizing Customer Reviews
Text Summarisation based on Human Language Technologies and its Applications Elena Lloret Pastor Supervisor: Dr. Manuel Palomar Seminar - June 2011.
Overview of the Fourth Recognising Textual Entailment Challenge NIST-Nov. 17, 2008TAC Danilo Giampiccolo (coordinator, CELCT) Hoa Trang Dan (NIST)
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
RuleML-2007, Orlando, Florida1 Towards Knowledge Extraction from Weblogs and Rule-based Semantic Querying Xi Bai, Jigui Sun, Haiyan Che, Jin.
ONTOLOGY LEARNING AND POPULATION FROM FROM TEXT Ch8 Population.
Automatic Detection of Tags for Political Blogs Khairun-nisa Hassanali and Vasileios Hatzivassiloglou Human Language Technology Research Institute The.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
PAUL ALEXANDRU CHIRITA STEFANIA COSTACHE SIEGFRIED HANDSCHUH WOLFGANG NEJDL 1* L3S RESEARCH CENTER 2* NATIONAL UNIVERSITY OF IRELAND PROCEEDINGS OF THE.
Chapter 2 Architecture of a Search Engine. Search Engine Architecture n A software architecture consists of software components, the interfaces provided.
Question Answering.  Goal  Automatically answer questions submitted by humans in a natural language form  Approaches  Rely on techniques from diverse.
Automatic Detection of Tags for Political Blogs Khairun-nisa Hassanali Vasileios Hatzivassiloglou The University.
Search and Information Extraction Lab IIIT Hyderabad.
Web Document Clustering: A Feasibility Demonstration Oren Zamir and Oren Etzioni, SIGIR, 1998.
INTERESTING NUGGETS AND THEIR IMPACT ON DEFINITIONAL QUESTION ANSWERING Kian-Wei Kor, Tat-Seng Chua Department of Computer Science School of Computing.
Efficiently Computed Lexical Chains As an Intermediate Representation for Automatic Text Summarization H.G. Silber and K.F. McCoy University of Delaware.
Relevance Detection Approach to Gene Annotation Aid to automatic annotation of databases Annotation flow –Extraction of molecular function of a gene from.
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Mining Topic-Specific Concepts and Definitions on the Web Bing Liu, etc KDD03 CS591CXZ CS591CXZ Web mining: Lexical relationship mining.
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
A Novel Pattern Learning Method for Open Domain Question Answering IJCNLP 2004 Yongping Du, Xuanjing Huang, Xin Li, Lide Wu.
Summarization Focusing on Polarity or Opinion Fragments in Blogs Yohei Seki Toyohashi University of Technology Visiting Scholar at Columbia University.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
Summarizing Encyclopedic Term Descriptions on the Web from Coling 2004 Atsushi Fujii and Tetsuya Ishikawa Graduate School of Library, Information and Media.
Software Quality in Use Characteristic Mining from Customer Reviews Warit Leopairote, Athasit Surarerks, Nakornthip Prompoon Department of Computer Engineering,
Using a Named Entity Tagger to Generalise Surface Matching Text Patterns for Question Answering Mark A. Greenwood and Robert Gaizauskas Natural Language.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Finding frequent and interesting triples in text Janez Brank, Dunja Mladenić, Marko Grobelnik Jožef Stefan Institute, Ljubljana, Slovenia.
Information Retrieval
1 Generating Comparative Summaries of Contradictory Opinions in Text (CIKM09’)Hyun Duk Kim, ChengXiang Zhai 2010/05/24 Yu-wen,Hsu.
Advantages of Query Biased Summaries in Information Retrieval by A. Tombros and M. Sanderson Presenters: Omer Erdil Albayrak Bilge Koroglu.
Evaluating Answer Validation in multi- stream Question Answering Álvaro Rodrigo, Anselmo Peñas, Felisa Verdejo UNED NLP & IR group nlp.uned.es The Second.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
Answer Mining by Combining Extraction Techniques with Abductive Reasoning Sanda Harabagiu, Dan Moldovan, Christine Clark, Mitchell Bowden, Jown Williams.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Multilingual Information Retrieval using GHSOM Hsin-Chang Yang Associate Professor Department of Information Management National University of Kaohsiung.
哈工大信息检索研究室 HITIR ’ s Update Summary at TAC2008 Extractive Content Selection Using Evolutionary Manifold-ranking and Spectral Clustering Reporter: Ph.d.
Question Answering Passage Retrieval Using Dependency Relations (SIGIR 2005) (National University of Singapore) Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan,
Twitter as a Corpus for Sentiment Analysis and Opinion Mining
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
11 Thoughts on STS regarding Machine Reading Ralph Weischedel 12 March 2012.
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Summarizing Entities: A Survey Report
Presentation 王睿.
CSE 635 Multimedia Information Retrieval
Presentation transcript:

Department of Software and Computing Systems Research Group of Language Processing and Information Systems The DLSIUAES Team’s Participation in the TAC 2008 Tracks – Opinion Pilot Alexandra Balahur, Elena Lloret, Andrés Montoyo, Manuel Palomar

Overview Task definition Objectives of participation Question processing Answer retrieval Summary generation Evaluation & discussion Conclusions & future work

Opinion pilot task definition Input - (opinion) questions from the TAC QA Track and the text snippets output by QA systems. Goal - produce short coherent summaries of the answers to the questions from the text snippets themselves, or from the associated documents. Evaluation - readability and content (Nugget Pyramid Method )

Description of test data 25 topics 22 with two questions Usually asking positive/negative aspects on the topic Comparisons among 2 objects 3 with just one question Only the positive or negative aspects of an entity Answer snippets – variable number Correspondence between answer snippets and question not provided

Objectives of participation What is needed to build an MPQA system Difference to classical QA systems in question analysis & answer retrieval Test a general opinion mining system Test the relevance of different resources and techniques to these tasks Test importance of opinion strength to summarization

Question processing stage Question patterns interrogation formula opinion words. Examples of rules for the interrogation formula “What reasons” are: What reason(s) (.*?) for (not) (affect_verb + ing) (.*?)? What reason(s) (.*?) for (lack of) (affect_noun) (.*?)? What reason(s) (.*?) for (affect_adjective|positive|negativ e) opinions (.*?)?

Question processing stage Question polarity WordNet Affect (Strapparava and Valitutti, 2006) emotion lists the emotion triggers resource (fight, destroy, burn etc.) (Balahur and Montoyo, 2008) list of attitudes for the categories of criticism, support, admiration and rejection (em. triggers) two categories of value words (good and bad) - opinion mining system. Words that denote human needs and motivations, whose presence triggers emotion.

Question processing stage Question keywords filtering out stop words. Question focus determining the gist of the question. Output of the question processing stage: reformulation patterns (coherence to summaries), question focus, keywords and the question polarity (->define several rules to make a correspondence between the question and the answer snippets on the further processing stage).

Correspondence rules 1. One question on the topic ⇒ retrieved snippet has same polarity as the question. 2. Two questions on the topic with different polarity ⇒ the snippets retrieved are classified according to their polarity. 3. Two questions with different focus and polarity ⇒ the snippets retrieved are classified according to their focus and polarity. 4. Two questions with the same focus and polarity ⇒ the order of the entities in focus both in the question and in the answer snippets is taken into account, together with a polarity matching between the question and the snippet.

Answer retrieval 3 approaches, only 2 evaluated 1. Using the provided answer snippets – snippet-driven approach 2. Not using the provided snippets; including the blog answer candidate snippets – blog driven approach 3. Using the provided answer snippets and employing anaphora resolution on original blogs

Snippet-driven approach Blogs HTML tags removed; split into sentences Using answer snippets provided Snippets sought in the original blogs Those not literally contained -stemmed, stopwords removed Computed similarity to potential sentences in the blogs with Pedersen’s similarity package Extract the most similar blog sentences, and their focus

Snippet-driven approach

Eliminating “noise” Using Minipar and selecting only sentences with S and Pred Determining the polarity of the snippet/blog phrase With Pedersen’s Text Similarity Package, using the score with the terms in WN Affect, the ISEAR corpus and the emotion triggers Summing up positive scores Summing up negative scores Which is the greater (no machine learning possibility)

Snippet-driven approach 6 emotions: +shame+guilt

Snippet-driven approach Answering the questions By topic and polarity correspondance between the question and the retrieved snippets/blog phrases using the rules

Blog-phrase driven approach Not using the answer snippet provided Eliminated the stopwords of the questions Determined the question focus&keywords Using the keywords and focus, determine blog phrases that could be the answer using similarity

Blog-phrase driven approach Eliminating “noise” Using Minipar and selecting only sentences with S and Pred Determining the polarity of the snippet/blog phrase With Pedersen’s Text Similarity Package, using the score with the terms in WN Affect, the ISEAR corpus and the emotion triggers Answering the questions By topic and polarity correspondance between the question and the retrieved snippets/blog phrases using the rules

Summary generation Using the question reformulation patterns and the retrieved answers; Tree-Tagger POS-Tagging to find 3rd pers. sing. and change them to 3rd pers. pl.; use replacement patterns(I/it etc) Snippet-driven: final summary Blog-driven: sorting the retrieved snippets in descending order, with respect to their polarity scores;included in summary those with highest scores, until reaching the imposed limit

Evaluation 1. summarizerID 2. Run type “manual”/ “automatic” 3. Use of answer snippets provided by NIST – “yes”/ ”no” 4. Average pyramid F-score (Beta=1), *averaged over 22 summaries 5. Grammaticality* 6. Non-redundancy* 7. Structure/Coherence * 8. Overall fluency/readability* 9. Overall responsiveness* (0.123) (0.123) (0.123) 5.409

Evaluation 1. summarizerID 2. Run type “manual”/ “automatic” 3. Use of answer snippets provided by NIST – “yes”/ ”no” 4. Average pyramid F-score (Beta=1), *averaged over 22 summaries 5. Grammaticality* 6. Non-redundancy* 7. Structure/Coherence * 8. Overall fluency/readability* 9. Overall responsiveness*

Evaluation 1. summarizerID 2. Run type “manual”/ “automatic” 3. Use of answer snippets provided by NIST – “yes”/ ”no” 4. Average pyramid F-score (Beta=1), *averaged over 22 summaries 5. Grammaticality* 6. Non-redundancy* 7. Structure/Coherence * 8. Overall fluency/readability* 9. Overall responsiveness*

Discussion + System performed well regarding Precision and Recall, the first run begin classified 7th among the 36 as F-measure + Structure and coherence 4/36 –reform. patterns + Overall responsiveness 5/36 +Second approach was well as F-measure – similarity/polarity/polarity strength -- did not perform very well with respect of the non-redundancy criterion & grammaticality one

Conclusions With the participation in the TAC 2008 we could: 1. Test a general opinion mining system, working with different affect and opinion categories – worked well 2. Test the importance of the resources used and the relevance they have to this task – relevant resources 3. Test the relavance of polarity strength to the resultsand to computing the relevance of the retrieved text - positive 4. Test manners to generate coherence and grammaticality of text through patterns – evaluated well as coherence 5. Test a method of summarization based on polarity strength 6. Determine what is needed in order to build an MPQA system – a modified method from the classical QA systems

Future work 1. Employ a Textual Entailment system for redundancy detection 2. Check grammaticality 3. Develop alternative methods for retrieving the candidate answers, by query expansion, as for factual texts, but using affective and opinion vocabulary 4. Test how many of retrieved snippets were not included in summary due to polarity

Department of Software and Computing Systems Research Group of Language Processing and Information Systems Thank you! Alexandra Balahur, Elena Lloret, Andrés Montoyo, Manuel Palomar