Download presentation
Presentation is loading. Please wait.
Published byFrederick Bell Modified over 9 years ago
1
Automating the Inpatient Chronic Heart Failure Quality Measures in VA Jennifer Garvin PhD, MBA, RHIA, CPHQ, CCS, CTR, FAHIMA Salt Lake City VA Healthcare System IDEAS Center University of Utah 5/16/14 This study is supported by the VA HSR&D IBE- 09-069-1 grant www.wordle.net
2
Overview Describe development of natural language processing (NLP) tools Describe inpatient chronic heart failure (CHF) quality measures Demonstrate the development of natural language process development tools using the case study of CHF
3
Clinical Guidelines Formulary Clinical Processes Appropriateness Measures Decision Support Performance Measures Clinical Reminders Applied Use Case Purpose & Background Evidence http://www.healthquality.va.gov/chf/
4
VA Informatics and Computing Infrastructure- VINCI Research approvals Assigned a research folder on VINCI In VINCI all approved investigators and staff can access data http://www.hsrd.research.va.gov/for_researchers/vinci / http://www.hsrd.research.va.gov/for_researchers/vinci / 4
5
Preparation for Research Workflow analysis- visitation and discussion with 2 echo laboratories Understanding how the documents are developed- variety of approaches Initial review of document structure to inform our sampling strategy- we planned to oversampled free text and semi-structured text – Paragraph with no outline structure (free text) – Outline with some free text (semi-structured text) – Outline (highly structured text) 5
6
Document Format (Fake Texts using XXX in Place of Letters) 6
7
EF Sampling Strategy We had a total of 765 documents available, we needed 367 minimum documents for the test set, and used the remaining 398 documents were available for training. However, if during system training, the performance of the system reaches the pre-specified level of accuracy without using all available training documents, the remaining unused documents in the training set could be added to the test set. The 765 documents were randomly assigned to a training and test set in preparation for annotation. 7
8
NLP Development Methods- Reference (Gold) Standard Development All documents in the training and test sets must have an accompanying reference (gold) standard so that the accuracy of the system can be measured during training and testing Software program called Knowtator was used for annotation Two independent reviewers with a third adjudicator when disagreement occurred 8
9
Ejection Fraction Annotation Schema 1. Ejection Fraction – annotate all mentions of left ventricular ejection fraction. 2.Value – annotate all mentions of the quantitative value associated with left ventricular ejection fraction. 3.Qualitative assessment – annotate all mentions of qualitative assessment of LV ejection fraction and LV systolic function. 4.LV systolic function – annotate all mentions of left ventricular systolic function. 5.Document level a)EF Range ( =40%, undetermined) b)Informativeness: The format of the document was sufficiently predictable that I could skim it rapidly to find what I wanted c)Consistency: In order to verify that the information in the document was internally consistent, I found that I had to go “back and forth” Classes
10
NLP Development Methods- Training and Testing Training – The system we used was developed but was not “trained” for this specific use case – Separated the training documents into batches of documents and ran the system of a set of the batches – Evaluated false positives and false negatives based on comparison to the reference (gold) standard – Reprogram the system and run against the next batch – When pre-specified level of accuracy reached, measure accuracy at the last iteration Testing – The system is run on the sequestered documents and the output received by the statistician 10
11
Diagram of Overall Classification and Sub-classifications Automated Data Acquisition for Heart Failure (ADAHF)
12
Summary of Development Steps Determine a use case Investigate if there are existing tools that have been used for a given use case If none, an existing tool may need to be may be generalized or a new one developed Determine data elements Determine development environment Train and Test Assess accuracy via sensitivity,(recall) specificity, positive predictive value (precision)
13
Stakeholder Engagement Theoretic Framework and Model The Promoting Action on Research on Implementation in Health Services (PARIHS) framework 1-2 – Evidence – Context – Facilitation Socio-Technical Model (STM) 3 Eight Dimensions of which we are using four: – hardware and software – clinical content – workflow and communication – internal organizational features 1 Stetler, 2011 http://www.implementationscience.com/content/6/1/99 1 Stetler, 2011 http://www.implementationscience.com/content/6/1/99 2 Kitson, 2008 http://www.implementationscience.com/content/3/1/1 Kitson, 2008 http://www.implementationscience.com/content/3/1/1 3 Sittig and Singh, A new sociotechnical model for studying health information technology in complex adaptive healthcare systems, Qual Saf Health Care 2010;19
14
Stakeholder Engagement: Semi- structured Interview and Thematic Analysis Approach is “Applied”- to solve a problem 6 using a theoretical thematic analysis 7 Two independent reviewers each create summary and organize identify themes to answer research questions Research group met to develop consensus codes on master themes. Three documents resulted – two summaries, group- consensus codes, consensus codes with highlighted text 6 Guest et al, Applied Thematic Analysis, 2012 7 Braun et al, Using Thematic Analysis in Psychology, 2006
15
Stakeholder Interview Results: Respondent Characteristics We interviewed 13 stakeholders. The interviewees included among others: clinical quality specialists; directors of quality management, clinical analysis and reporting; epidemiology; clinicians and pharmacists; and program analysts The range for the number of years in VA of respondents is 2- 35 And similarly, the range for the number of years in quality/patient safety is 2-33
16
Stakeholder Engagement Preliminary Results - Internal Factors Internal factors that facilitate implementation of an automated system include: – Use of evidence-based care – A culture of continuous quality improvement coupled with measurement and accountability processes – Quality control reporting both within and external to the VA.
17
Stakeholder Engagement (cont.) Hardware and Software-Preliminary Findings Informatics is used for: – Communication between Providers and Patients Secure messaging Blue Button download Kiosks Mobile technology MyHealtheVet Informatics is used for (cont.): – Clinical Care CPRS Clinical Decision Support Templates designed to facilitate clinically relevant content Smart forms CART-CL Primary Care Almanac We have an Informatics-Rich Environment in VA
18
Stakeholder Engagement- Preliminary Results- Informatics Applications Used with Quality Metrics Quality Improvement Functionality (current) Performance integrated tracking application (PITA) Measure Master Report System extraction and output specifics: Capture the concepts and values as well as the words around the concepts Have the ability to adjust the EF value captured In Development Growing number of informatics tools Process chart notes using NLP Surveillance tools Analytic tools Meaningful Use – Determine how we could provide data for meaningful use.
19
Formative Evaluation Process Initial stakeholder engagement- went through a couple cycles of: – Develop a prototype of the tool with report – Revise based on feedback Developed a final prototype of the table Develop an initial functional HMP Module User-centered Design Analysis
20
Thank you! Questions or Comments? Please contact me at Jennifer.garvin@va.gov Jennifer.garvin@va.gov This study is undertaken as part of the VA HSR&D IBE- 09-069-1 grant. The views expressed in this article are those of the authors and do not necessarily represent the views of the Department of Veterans Affairs or the University of Utah School of Medicine. I thank the Department of Veterans Affairs for my fellowship and Gail Graham RHIA and Mark Weiner MD for being my mentors ADAHF Team: – Julia Heavirland/ Jenifer Williams (Annotators) – Youngjun Kim (Application Specialist) – Stephane Meystre MD, PhD (Faculty System Developer) – Drs. Bruce Bray, Paul Heidenreich, Mary Goldstein, Wendy Chapman, Michael Matheny, Gobbel (Co-investigators) – Andrew Redd PhD and Dan Bolton MS (Statisticians) – Megha Kalsy MS and Natalie Kelly MBA (Stakeholder Engagement) – Jennifer Garvin PhD, MBA (Principal Investigator)
21
Extra Slides
22
EF Sampling Strategy To account for clustering, the sample size was increased by the design effect. For ICC = 0.005, Deff = 1 + 25(0.005) = 1.125 so we require 179(1.125) = 201.375 202 positive cases or 29 positive cases per facility; Dividing by prevalence estimate of EF in documents, we have 29/0.80 = 36.25 37 documents required per facility. We multiply that result by the number of sites and the product is 37(7) = 259 Doubling the sample size for the three facilities with free- or semi-structured text resulted in a minimum of 367 documents in the test set. 22
23
Definitions Sensitivity is the proportion of patients with disease who test positive. In probability notation: P(T + |D + ) = TP / (TP+FN). Specificity is the proportion of patients without disease who test negative. In probability notation: P(T - |D - ) = TN / (TN + FP). Sensitivity and specificity describe how well the test discriminates between patients with and without disease. They address a different question than we want answered when evaluating a patient, however. What we usually want to know is: given a certain test result, what is the probability of disease? This is the predictive value of the test. Predictive value of a positive test (PPV) is the proportion of patients with positive tests who have disease. In probability notation: (D + |T + ) = TP / (TP+FP). 23
24
Definitions The weighted harmonic mean of precision and recall is the F-measure F= 2*Precision*Recall/(Precision + Recall) Kappa- more accurate than percent agreement as it accounts for chance agreement K= Observed agreement + Hypothetical probability of chance agreement/1- hypothetical probability of chance agreement 24
25
Definitions Regular expressions: – A regular expression (regex or regexp for short) is a special text string for describing a search pattern. – www.regular-expressions.info 25
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.