Www.amia.org S14: Interpretable Probabilistic Latent Variable Models for Automatic Annotation of Clinical Text Alexander Kotov 1, Mehedi Hasan 1, April.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Recognizing Human Actions by Attributes CVPR2011 Jingen Liu, Benjamin Kuipers, Silvio Savarese Dept. of Electrical Engineering and Computer Science University.
ICONIP 2005 Improve Naïve Bayesian Classifier by Discriminative Training Kaizhu Huang, Zhangbing Zhou, Irwin King, Michael R. Lyu Oct
Distant Supervision for Emotion Classification in Twitter posts 1/17.
Unsupervised Modeling of Twitter Conversations
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Simultaneous Image Classification and Annotation Chong Wang, David Blei, Li Fei-Fei Computer Science Department Princeton University Published in CVPR.
Rhee Dong Gun. Chapter The speaking process The differences between spoken and written language Speaking skills Speaking in the classroom Feedback.
Made with OpenOffice.org 1 Sentiment Classification using Word Sub-Sequences and Dependency Sub-Trees Pacific-Asia Knowledge Discovery and Data Mining.
Joint Sentiment/Topic Model for Sentiment Analysis Chenghua Lin & Yulan He CIKM09.
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Statistical Models for Networks and Text Jimmy Foulds UCI Computer Science PhD Student Advisor: Padhraic Smyth.
An Overview of Text Mining Rebecca Hwa 4/25/2002 References M. Hearst, “Untangling Text Data Mining,” in the Proceedings of the 37 th Annual Meeting of.
Sentence Classifier for Helpdesk s Anthony 6 June 2006 Supervisors: Dr. Yuval Marom Dr. David Albrecht.
1 Automated Feature Abstraction of the fMRI Signal using Neural Network Clustering Techniques Stefan Niculescu and Tom Mitchell Siemens Medical Solutions,
+ Doing More with Less : Student Modeling and Performance Prediction with Reduced Content Models Yun Huang, University of Pittsburgh Yanbo Xu, Carnegie.
Scalable Text Mining with Sparse Generative Models
Text Classification Using Stochastic Keyword Generation Cong Li, Ji-Rong Wen and Hang Li Microsoft Research Asia August 22nd, 2003.
Forecasting with Twitter data Presented by : Thusitha Chandrapala MARTA ARIAS, ARGIMIRO ARRATIA, and RAMON XURIGUERA.
CSC 478 Programming Data Mining Applications Course Summary Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
STRUCTURED PERCEPTRON Alice Lai and Shi Zhi. Presentation Outline Introduction to Structured Perceptron ILP-CRF Model Averaged Perceptron Latent Variable.
Dongyeop Kang1, Youngja Park2, Suresh Chari2
How To Do Multivariate Pattern Analysis
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
Exploiting Ontologies for Automatic Image Annotation M. Srikanth, J. Varner, M. Bowden, D. Moldovan Language Computer Corporation
2007. Software Engineering Laboratory, School of Computer Science S E Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying.
 Text Representation & Text Classification for Intelligent Information Retrieval Ning Yu School of Library and Information Science Indiana University.
Designing Ranking Systems for Consumer Reviews: The Economic Impact of Customer Sentiment in Electronic Markets Anindya Ghose Panagiotis Ipeirotis Stern.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
Unit 3 Lesson 1 Locate Resources to Answer Nutrition and Physical Activity Questions.
Transfer Learning Task. Problem Identification Dataset : A Year: 2000 Features: 48 Training Model ‘M’ Testing 98.6% Training Model ‘M’ Testing 97% Dataset.
Partially Supervised Classification of Text Documents by Bing Liu, Philip Yu, and Xiaoli Li Presented by: Rick Knowles 7 April 2005.
Problem of the Day  Why are manhole covers round?
Feature Detection in Ajax-enabled Web Applications Natalia Negara Nikolaos Tsantalis Eleni Stroulia 1 17th European Conference on Software Maintenance.
A Model for Learning the Semantics of Pictures V. Lavrenko, R. Manmatha, J. Jeon Center for Intelligent Information Retrieval Computer Science Department,
Experiments. The essential feature of the strategy of experimental research is that you… Compare two or more situations (e.g., schools) that are as similar.
BLOOM’S TAXONOMY OF THE COGNITIVE DOMAIN. BLOOM’S TAXONOMY Benjamin Bloom (et al.) created this taxonomy for categorizing levels of abstraction of questions.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Latent Dirichlet Allocation
Automatic recognition of discourse relations Lecture 3.
Hello, Who is Calling? Can Words Reveal the Social Nature of Conversations?
Created by Denisova Elena Uranievna, School No.5, Strezhevoy, Tomsk Region (Conversational practice for study group work)
Towards Total Scene Understanding: Classification, Annotation and Segmentation in an Automatic Framework N 工科所 錢雅馨 2011/01/16 Li-Jia Li, Richard.
Automatic Labeling of Multinomial Topic Models
Exploiting Named Entity Taggers in a Second Language Thamar Solorio Computer Science Department National Institute of Astrophysics, Optics and Electronics.
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Efficient Text Categorization with a Large Number of Categories Rayid Ghani KDD Project Proposal.
Using Latent Dirichlet Allocation for Child Narrative Analysis Khairun-nisa Hassanali 1, Yang Liu 1 and Thamar Solorio 2
Text-classification using Latent Dirichlet Allocation - intro graphical model Lei Li
Unit 1 How often do you exercise? What does he /she do on weekends? usually often hardly ever sometimes always never.
Collaborative Deep Learning for Recommender Systems
Predicting and Adapting to Poor Speech Recognition in a Spoken Dialogue System Diane J. Litman AT&T Labs -- Research
Understanding unstructured texts via Latent Dirichlet Allocation Raphael Cohen DSaaS, EMC IT June 2015.
David Amar, Tom Hait, and Ron Shamir
Queensland University of Technology
Sentiment analysis algorithms and applications: A survey
The topic discovery models
Erasmus University Rotterdam
The topic discovery models
Predicting the Outcome of Patient-Provider Communication Sequences using Recurrent Neural Networks and Probabilistic Models S38: Predictive Modeling.
Code Switching Shaw.
The topic discovery models
Michal Rosen-Zvi University of California, Irvine
S04: Machine Learning in Clinical Predictive Modeling I
Presentation transcript:

S14: Interpretable Probabilistic Latent Variable Models for Automatic Annotation of Clinical Text Alexander Kotov 1, Mehedi Hasan 1, April Carcone 1, Ming Dong 1, Sylvie Naar-King 1, Kathryn Brogan Hartlieb 2 1 Wayne State University 2 Florida International University

Disclosure I have nothing to disclose 2

Motivation Annotation = assignment of codes from a codebook to fragments of clinical text Integral part of clinical practice or qualitative data analysis Codes (or labels) can viewed as summaries abstractions Analyzing sequences of codes allows to discover patterns and associations 3

Study context We focus on clinical interview transcripts: – motivational interviews with obese adolescents conducted at a Pediatric Prevention Research Center at Wayne State University Codes designate the types of patient’s utterances Distinguish the subtle nuances of patient’s behavior Analysis of coded successful interviews allows clinicians to identify communication strategies that trigger patient’s motivational statements (i.e. “change talk”) Change talk has been shown to predict actual behavior change, as long as 34 months later 4

Problem Annotation is traditionally done by trained coders – time-consuming, tedious and expensive process We study the effectiveness of machine learning methods for automatic annotation of clinical text Such methods can have tremendous impact: – decrease the time for designing interventions from months to weeks – increase the pace of discoveries in motivational interviewing and other qualitative research 5

Challenges Annotation in case of MI = inferring psychological state of patients from text Important indicators of emotions (e.g. gestures, facial expressions and intonations) are lost during transcription Children and adolescents often use incomplete sentences and frequently change subjects Annotation methods need to be interpretable 6

Coded interview fragments 7 CodeExample CL-I eat a lot of junk food. Like, cake and cookies, stuff like that. CL+Well, I've been trying to lose weight, but it really never goes anywhere. CT-It can be anytime; I just don't feel like I want to eat (before) I'm just not hungry at all. CT+ Hmm. I guess I need to lose some weight, but you know, it's not easy. AMBFried foods are good. But it's not good for your health.

Methods Proposed methods: – Latent Class Allocation (LCA) – Discriminative Labeled Latent Dirichlet Allocation (DL-LDA) Baselines: – Multinomial Naïve Bayes – Labeled Latent Dirichlet Allocation (Ramage et al., EMNLP’09) 8

Latent Class Allocation c 9

Discriminative Labeled LDA c 10

Classification 11

Experiments 2966 manually annotated fragments of motivational interviews conducted at the Pediatric Prevention Research Center of Wayne State University’s School of Medicine Only unigram lexical features were used Preprocessing: – RAW: no stemming or stop-words removal – STEM: stemming but no stop-words removal – STOP: stop-words removal, but no stemming – STOP-STEM: stemming and stop-words removal Randomized 5-fold cross-validation – results are based on weighted macro-averaging 12

Task 1: classifying 5 original classes 5 classes: CL-, CL+, CT-, CT+, AMB Class distribution: class# samples% CL CL CT CT AMB

Task 1: performance 14 RecallPrecisionF1-measure RAW STEM STOP STOP-STEM LCA: DL-LDA: RecallPrecisionF1-measure RAW STEM STOP STOP-STEM

Naïve Bayes: L-LDA: 15 RecallPrecisionF1-measure RAW STEM STOP STOP-STEM RecallPrecisionF1-measure RAW STEM STOP STOP-STEM Task 1: performance

Task 1: summary of performance LCA shows the best performance in terms of precision and F1- measure LCA and DL-LDA outperform NB in L-LDA in terms of all metrics DL-LDA has higher recall than LCA and comparable precision and F1-measure – probabilistic separation of words by specificity + dividing class specific multinomials translates into better classification results RecallPrecisionF1-measure NB LCA L-LDA DL-LDA

Most characteristic terms CodeTerms CL-drink sugar gatorade lot hungry splenda beef tired watch tv steroids sleep home nervous confused starving appetite asleep craving pop fries computer CL+stop run love tackle vegetables efforts juice swim play walk salad fruit CT-got laughs sleep wait answer never tired fault phone joke weird hard don’t CT+time go mom brother want happy clock boy can move library need adopted reduce sorry solve overcoming lose AMBwhat taco mmm know say plus snow pain weather 17

Task 2: classifying CL, CT and AMB 3 classes: CL (CL+ and CL-), CT (CT+ and CT-) and AMB Class distribution: Performance: RecallPrecisionF1-measure NB LCA L-LDA DL-LDA classsamples% CL CT AMB

Task 3: classifying -, + and AMB 3 classes: + (CL+ and CT+), - (CL- and CT-) and AMB Class distribution: Performance: RecallPrecisionF1-measure NB LCA L-LDA DL-LDA class# samples% AMB

Summary We proposed two novel interpretable latent variable models for probabilistic classification of textual fragments Latent Class Allocation probabilistically separates discriminative from common terms Discriminative Labeled LDA is an extension of Labeled LDA that differentiates between class specific topics and background LM Experimental results indicated that LCA and DL-LDA outperform state-of-the-art interpretable probabilistic classifiers (Naïve Bayes and Labeled LDA) for the task of automatic annotation of interview transcripts 20

Thank you! Questions? 21