The 2nd Workshop on Argumentation Mining

Slides:



Advertisements
Similar presentations
Progress update Lin Ziheng. System overview 2 Components – Connective classifier Features from Pitler and Nenkova (2009): – Connective: because – Self.
Advertisements

Made with OpenOffice.org 1 Sentiment Classification using Word Sub-Sequences and Dependency Sub-Trees Pacific-Asia Knowledge Discovery and Data Mining.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
A Framework for Named Entity Recognition in the Open Domain Richard Evans Research Group in Computational Linguistics University of Wolverhampton UK
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Natural Language Understanding
Top Ten Tips for teachers preparing students for the academic version of IELTS Sam McCarter Macmillan Online Conference 2013.
Mining the Peanut Gallery: Opinion Extraction and Semantic Classification of Product Reviews K. Dave et al, WWW 2003, citations Presented by Sarah.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
1 Academic Skills Tips for Essay Writing. 2 Outline of today’s lecture Academic skills Essay writing Paraphrasing Summarizing.
A Compositional Context Sensitive Multi-document Summarizer: Exploring the Factors That Influence Summarization Ani Nenkova, Stanford University Lucy Vanderwende,
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
Researcher affiliation extraction from homepages I. Nagy, R. Farkas, M. Jelasity University of Szeged, Hungary.
 Text Representation & Text Classification for Intelligent Information Retrieval Ning Yu School of Library and Information Science Indiana University.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
Transfer Learning Task. Problem Identification Dataset : A Year: 2000 Features: 48 Training Model ‘M’ Testing 98.6% Training Model ‘M’ Testing 97% Dataset.
1 Welcome to IST Department of Psychology
DISCRIMINATIVE TRAINING OF LANGUAGE MODELS FOR SPEECH RECOGNITION Hong-Kwang Jeff Kuo, Eric Fosler-Lussier, Hui Jiang, Chin-Hui Lee ICASSP 2002 Min-Hsuan.
A Systematic Exploration of the Feature Space for Relation Extraction Jing Jiang & ChengXiang Zhai Department of Computer Science University of Illinois,
Copyright © 2013 by Educational Testing Service. All rights reserved. 14-June-2013 Detecting Missing Hyphens in Learner Text Aoife Cahill *, Martin Chodorow.
Chapter 23: Probabilistic Language Models April 13, 2004.
National Taiwan University, Taiwan
Boundary Detection in Tokenizing Network Application Payload for Anomaly Detection Rachna Vargiya and Philip Chan Department of Computer Sciences Florida.
1 Masters Thesis Presentation By Debotosh Dey AUTOMATIC CONSTRUCTION OF HASHTAGS HIERARCHIES UNIVERSITAT ROVIRA I VIRGILI Tarragona, June 2015 Supervised.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Recognizing Stances in Online Debates Unsupervised opinion analysis method for debate-side classification. Mine the web to learn associations that are.
AN EFFECTIVE STATISTICAL APPROACH TO BLOG POST OPINION RETRIEVAL Ben He Craig Macdonald Iadh Ounis University of Glasgow Jiyin He University of Amsterdam.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Ontology-Based Argument Mining and Automatic Essay Scoring Nathan Ong, Diane Litman, Alexandra Brusilovsky University of Pittsburgh First Workshop on Argumentation.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
Word Sense and Subjectivity (Coling/ACL 2006) Janyce Wiebe Rada Mihalcea University of Pittsburgh University of North Texas Acknowledgements: This slide.
Modeling Student Benefits from Illustrations and Graphs Michael Lipschultz Diane Litman Computer Science Department University of Pittsburgh.
Selecting Relevant Documents Assume: –we already have a corpus of documents defined. –goal is to return a subset of those documents. –Individual documents.
Maximum Entropy techniques for exploiting syntactic, semantic and collocational dependencies in Language Modeling Sanjeev Khudanpur, Jun Wu Center for.
N-Gram Model Formulas Word sequences Chain rule of probability Bigram approximation N-gram approximation.
Recent Paper of Md. Akmal Haidar Meeting before ICASSP 2013 報告者:郝柏翰 2013/05/23.
Collaborative Deep Learning for Recommender Systems
Academic Writing Fatima AlShaikh. A duty that you are assigned to perform or a task that is assigned or undertaken. For example: Research papers (most.
The University of Illinois System in the CoNLL-2013 Shared Task Alla RozovskayaKai-Wei ChangMark SammonsDan Roth Cognitive Computation Group University.
Language Identification and Part-of-Speech Tagging
PSY 626: Bayesian Statistics for Psychological Science
Reading Skills for Academic Study
Sentiment analysis algorithms and applications: A survey
Essay.
Lindsay & Gordon’s Discovery Support Systems Model
Authorship Attribution Using Probabilistic Context-Free Grammars
Erasmus University Rotterdam
Aspect-based sentiment analysis
PSY 626: Bayesian Statistics for Psychological Science
Statistical NLP: Lecture 9
Topic 2 Computational Approaches to Argumentation Mining in Natural Language Text 11/23/2018 Huy V. Nguyen.
Jun Wu and Sanjeev Khudanpur Center for Language and Speech Processing
Eiji Aramaki* Sadao Kurohashi* * University of Tokyo
N-Gram Model Formulas Word sequences Chain rule of probability
Automatic Detection of Causal Relations for Question Answering
Michal Rosen-Zvi University of California, Irvine
Text Mining & Natural Language Processing
Topic Models in Text Processing
Ping LUO*, Fen LIN^, Yuhong XIONG*, Yong ZHAO*, Zhongzhi SHI^
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Statistical NLP : Lecture 9 Word Sense Disambiguation
Stance Classification of Ideological Debates
Stance Classification of Context-Dependent Claims
Writing to Discuss / Argue – Steps to Success
Presentation transcript:

The 2nd Workshop on Argumentation Mining NAACL 2015 Workshops June 4th 2015, Denver, CO Extracting argument and domain words for identifying argument components in texts Huy Nguyen1 Diane Litman1,2 1Computer Science Department 2Learning Research & Development Center University of Pittsburgh This research is supported by NSF Grant 1122504

Outline Argument mining in texts Our approach Evaluation Motivation of our approach Our approach Evaluation Data Baseline vs. proposed models Experiment results Conclusions and future work Outline

Argument mining in texts Automatically identify argument elements and argumentative relations between them Argument mining as classification tasks Argument detection, i.e., argumentative sentence vs. none (Moens et al. 2007) Sentence’s rhetorical status, e.g., aim, background, contrast (Teufel & Moens 2002) Argumentative discourse structures, i.e., major claim, claim, premise (Stab & Gurevych 2014ab) Argument mining in texts

Argument mining in texts Motivation Student essays E.g., persuasive essays, academic writings Optional title/topic but no section heading Lack of evidence, substantiated by personal experience rather than cited resources Specialized (useful) features, e.g., section heading, citation (Teufel & Moens 2002), are not available Ngrams and syntactic rules (e.g., VP → VBG NP) are commonly used (Burstein et al. 2003, Moens et al. 2007, Stab & Gurevych 2014b, Park & Cardie 2014) But have limitations Large and sparse feature space Feature selection helps with over-fitting but not efficiently Argument mining in texts

Argument mining in texts Example essays Persuasive essay Academic essay Argument mining in texts

Argument mining in texts A novel feature design Separate argument words from domain words Argument words: argument indicators and commonly used in different argument topics, e.g., reason, opinion, believe Domain words: specific terminologies commonly used within the topic’s domain, e.g., art, education, children Post-process topic model output using seeding Seeding requires human knowledge but minimal Unannotated data Derive novel features to replace ngrams and syntactic rules Feature reduction Topic-independence Argument mining in texts

Extraction of argument and domain words Topic model Latent Dirichlet Allocation (Blei et al. 2003) Extraction of argument and domain words

LDA topics vs. essay topics LDA topics approximate writing topics but not completely believe children music view parent art How can we identify the LDA topic of argument words and maximize its difference from the others (LDA topics) ? … opinion school creative however learn talent discuss teach idea Extraction of argument and domain words

Argument and domain word extraction Identification step Pre-defined argument keywords: most frequent argument words in essay titles, e.g., opinion, agree Domain seed words: title words but not argument keywords or stop words 3 weights of a LDA topic (LDA word lists) (1) Argument weight Count of argument keywords in the list (2) Domain weight Sum of domain seeds’ occurrence frequency (f) (3) Combined weight Argument weight – Domain weight AW=3: opinion believe think DW= f(art) + f(music) + f(creative) + … CW = AW – DW Extraction of argument and domain words

Argument and domain word extraction (2) Maximization step (aka the best number of LDA topics) Given number of LDA topics K Discrimination ratio: sK = [CW(T1) – CW(T2)] / CW(T2) Get T1 the topic with largest combined weight Get T2 the topic with second largest combined weight Vary K and select K* with the largest discrimination ratio Argument word list = LDA topic w. largest combined weight K=9 K=10 K=40 T1 TK … … … … … Extraction of argument and domain words

Data 90 annotated persuasive essays (Stab & Gurevych 2014a, collected from www.essayforum.com) Sentences were coded for possible argument components: major claim, claim, premise Evaluation

Development data to learn argument and domain words > 6000 essays from www.essayforum.com, not in the corpus 10 argument keywords agree, disagree, reason, support, advantage, disadvantage, think, conclusion, result, opinion Learned 263 argument words Keyword variants: think, believe, viewpoint, opinion, argument, claim… Connectives: therefore, however, despite… Stop words Evaluation

Prediction models Baseline (Stab & Gurevych 2014b) Proposed model 1-, 2-, 3-grams Verbs, adverbs Presence of model verb Discourse connectives First person pronouns 1-, 2-, 3-grams Argument words (unigrams) Lexical Same >5000 features Dependency pairs (subject-main verb) do not contain domain words 956 features Syntactic rules Tense of main verb #sub-clauses, depth of parse tree Syntactic rules Syntactic #tokens, #punctuation Sentence position First/last paragraph First/last sentence of paragraph Structural Same #tokens, #punctuation, #sub-clauses, modal verb of preceding and following sentences Contextual Evaluation

Our model outperforms the baseline 10-fold cross validation following (Stab & Gurevych 2014b) Top features in training folds, calculated by InfoGain algorithm LibLINEAR learning algorithm with default parameters #feat. Acc. Kappa F1 Prec. Recall Baseline 100 0.78 0.63 0.71 0.76 0.69 Proposed 0.79+ 0.65* 0.72 0.70 75-essay training set, 15-essay test set Estimate best #features with cross-validation in training set #feat. Acc. Kappa F1 Prec. Recall Baseline 130 0.80 0.64 0.71 0.76 0.68 Proposed 70 0.83* 0.69* 0.76+ 0.79 0.74 With less features! * Significant higher (p < 0.05) + Trending higher (p < 0.1) Evaluation

Argument words learned from different domain Can they be used? Alternative argument word list 254 academic essays from college Psychology classes 5 argument keywords taken from the writing assignment hypothesis, support, opposition, finding, study Extract 429 argument words Replace the 263 argument words of the persuasive set 10-fold cross validation #feat. Acc. Kappa F1 Baseline 100 0.78 0.63 0.71 Proposed 0.79+ 0.65* 0.72 Alternative 0.62 Quantitatively worse! Evaluation

… and qualitatively different Alternative model’s top 100 feature (argument words of academic set) Proposed model’s top 100 feature (argument words of persuasive set) 140 words ≡ Academic ∩ Persuasive 30 of common argument words 22 of common argument words Transferable part: mostly function words 5 content words: conclusion, topic, analyze, show, reason 3 content words: conclusion, topic, analyze 20 of unique argument words 15 of unique argument words Non-transferable part: genre-dependent 19 content words: believe, agree, discuss, view… 6 content words: university, value… Most of popular terms in academic writings were not selected: research, hypothesis, variable… Evaluation

Conclusions and future work Novel algorithm to post-process LDA output to extract argument and domain words Minimal seeding, unannotated data Features derived from extracted argument words Efficiently replace ngrams and syntactic rules Argument words extracted from different data domain Non-transferable part are genre-dependent and needed for the best performance Our next study is argumentative relation classification, i.e. support vs. attack Conclusions and future work

Thank you! Questions and Comments END

In comparison with prior studies Our argument words are subsets of generic unigrams We emphasize the topic-independence of features Argument and domain word notion is similar to argument shell and content in (Madnani et al. 2012) We have no requirement of physical boundaries between the two aspects Our idea of using seed words to guide the word separation is similar to (Louis and Nenkova, 2013) We need much less prior knowledge We identify the best number of topics that maximize topic discrimination Argument mining in texts

Feature analysis 31 argument words 34 unigrams 31 bigrams 13 trigrams Baseline (top 130 features) Proposed model (top 70 features) 34 unigrams 31 bigrams 13 trigrams 31 argument words 6 not in baseline analyze, controversial, could, debate, discuss, ordinal 5 dependency pairs: I.agree, I.believe, I.conclude, I.think, people.believe 21 syntactic rules Evaluation

Two argument word lists overlap Persuasive Academic result experi signific support idea topic oppos reason conclus mention analyz consider 72 content words 142 words in common 70 discourse connectives and stop-words whether however therefore despite instead although though regardless moreover should would still Evaluation

Academic set 270 not selected to top 100. Most are popular terms of academic writing: research, hypothesis, variable… Footer

Data 1673 sentences 1552 argument components 327 non-argumentative sentences 36 LDA topics 1804 domain words Footer

Prediction performance 10-fold cross validation #feat. Acc. Kappa F1 Prec. Recall F1: major Claim F1: Claim F1: Premise F1: None Baseline 100 0.78 0.63 0.71 0.76 0.69 0.54 0.47 0.84 1.00 Proposed 0.79+ 0.65* 0.72 0.70 0.51 0.53* Estimate best #features and train in 75-essay set, test in 15-essay set #feat. Acc. Kappa F1 Prec. Recall F1: major Claim F1: Claim F1: Premise F1: None Baseline 130 0.80 0.64 0.71 0.76 0.68 0.48 0.49 0.86 1.00 Proposed 70 0.83* 0.69* 0.76+ 0.79 0.74 0.59 0.56* 0.88* Top features in training folds, calculated by InfoGain algorithm (Stab & Gurevych 2014b) * Significant higher (p < 0.05) + Trending higher (p < 0.1) Evaluation

Alternative argument word list Can argument words be learned from different genre? College students’ essays in introductory Psychology classes 254 essays, 5 argument keywords taken from the writing assignment hypothesis, support, opposition, finding, study Return 14 LDA topics, 429 argument words, 1497 domain words Replace the 285 argument words of the persuasive set 10-fold cross validation #feat. Acc. Kappa F1: major Claim F1: Claim F1: Premise F1: None Proposed 100 0.79+ 0.65* 0.51 0.53* 0.84* 1.00 Alternative 0.78 0.62 0.56 0.47 0.83 Evaluation

Argument and domain words Lexical signals of argumentative content and argument topic Argument shell and content The argument states that based on the result of the recent research, there probably were grizzly bears in Labrador (cf. Madnani et al. 2012) result research probably My view is that the government should give priorities to invest more money on the basic social welfares such as education and housing instead of subsidizing arts relative programs (cf. persuasive essay corpus, Stab & Gurevych 2014a) should instead of Extraction of argument and domain words

Sample essay Evaluation

Refercence online review (Park & Cardie 2014) online debate (Boltužic & Šnajder 2014) Footer

Stab14 (Stab & Gurevych 2014b) wLDA (Nguyen & Litman 2015) wLDA+5 (this study) 1-, 2-, 3-grams Argument words as unigrams Same as wLDA Numbers of argument & domain word Numbers of common words with title and preceding sentence Comparative & superlative adverbs and POS Discourse relation labels Plural first personal pronouns Lexical Verbs, adverbs, presence of model verb Discourse connectives, Singular first person pronouns Same as Stab14 Production rules LDA-enabled subject-main verb pairs Parse Tense of main verb #sub-clauses, depth of parse tree Same as Stab14 #tokens, #punctuation, sentence position First/last paragraph First/last sentence of paragraph Same as Stab14 Structure #tokens, #punctuation, #sub-clauses, modal verb in preceding/following sentences Context