Strategic Health IT Advanced Research Projects (SHARP) Area 4: Secondary Use Dr. Friedman on-site visit, Mayo Clinic 3 September 2010.

Slides:



Advertisements
Similar presentations
Connected Health Framework
Advertisements

Data Governance and Normalization with the Mayo Clinic Enterprise
FEBRUARY 25, 2014 STANLEY M. HUFF, MD CHIEF MEDICAL INFORMATICS OFFICER INTERMOUNTAIN HEALTHCARE Healthcare Services Platform: Goals and Vision.
Consistent and standardized common model to support large-scale vocabulary use and adoption Robust, scalable, and common API to reduce variation in clinical.
Test Automation Success: Choosing the Right People & Process
Strategic Health IT Advanced Research Projects (SHARP) Area 4: Secondary Use of EHR Data Project 3: High-Throughput Phenotyping Project Lead: Jyotishman.
Catherine Hoang Ioana Singureanu Greg Staudenmaier Detailed Clinical Models for Medical Device Domain Analysis Model 1.
Data Normalization Milestones. Data Normalization  Goals –To conduct the science for realizing semantic interoperability and integration of diverse data.
Maine SIM Evaluation: Presentation to Steering Committee December 10, 2014.
Area 4 SHARP Face-to-Face Conference Phenotyping Team – Centerphase Project Assessing the Value of Phenotyping Algorithms June 30, 2011.
Guoqian Jiang, MD, PhD Mayo Clinic
MAY 1, 2014 STANLEY M. HUFF, MD CHIEF MEDICAL INFORMATICS OFFICER INTERMOUNTAIN HEALTHCARE Healthcare Services Platform: Goals and Vision.
FEBRUARY 5, 2015 STANLEY M. HUFF, MD CHIEF MEDICAL INFORMATICS OFFICER INTERMOUNTAIN HEALTHCARE HSPC Meeting Introduction.
Massachusetts: Transforming the Healthcare Economy John D. Halamka MD CIO, Harvard Medical School and Beth Israel Deaconess Medical Center.
University of Pittsburgh Department of Biomedical Informatics Healthcare institutions have established local clinical data research repositories to enable.
Business Intelligence Dr. Mahdi Esmaeili 1. Technical Infrastructure Evaluation Hardware Network Middleware Database Management Systems Tools and Standards.
S&I Data Provenance Initiative Presentation to the HITSC on Data Provenance September 10, 2014.
Meaningful Use, Standards and Certification Under HITECH—Implications for Public Health InfoLinks Community of Practice January 14, 2010 Bill Brand, MPH,
Enterprise Architecture
AUGUST 21, 2014 STANLEY M. HUFF, MD CHIEF MEDICAL INFORMATICS OFFICER INTERMOUNTAIN HEALTHCARE HSPC Meeting Introduction.
WGClimate Work Plan for John Bates, Chair WGClimate 4th Working Group on Climate Meeting.
Harmonization of SHARPn Clinical Element Models with CDISC SHARE Clinical Study Data Standards Guoqian Jiang, MD, PhD Mayo Clinic On behalf of CDISC CEMs.
Annual SERC Research Review - Student Presentation, October 5-6, Extending Model Based System Engineering to Utilize 3D Virtual Environments Peter.
What is Business Analysis Planning & Monitoring?
SHARPn Data Normalization November 18, Data-driven Healthcare Big Data Knowledge Research Practice Analytics Domain Pragmatics Experts.
Initial Prototype for Clinical Data Normalization and High Throughput Phenotyping SHARPn F2F June 30,2011.
Initial slides for Layered Service Architecture
Chapter 2 The process Process, Methods, and Tools
Strategic Health IT Advanced Research Projects (SHARP) Area 4: Secondary Use of EHR Data SHARPfest June 2-3, 2010 PI: Christopher G Chute, MD DrPH.
Using the Open Metadata Registry (openMDR) to create Data Sharing Interfaces October 14 th, 2010 David Ervin & Rakesh Dhaval, Center for IT Innovations.
SE Minnesota Beacon Program Working Together to Improve Health Care Program Partners: Agilex Technologies; Dodge County & Public Health Department; Fillmore.
SHARPn High-Throughput Phenotyping (HTP) November 18, 2013.
Open Health Natural Language Processing Consortium (OHNLP)
EHR System (EHR-S) Functional Requirements Implementation Guide: Laboratory Results Interface (LRI) Kickoff March 3 rd,
Interoperability Showcase In collaboration with IHE Use Case 3 Care Theme: Leveraging National Healthcare Registries in Care Delivery Biosurveillance Monitoring.
Public Health Data Standards Consortium
State HIE Program Chris Muir Program Manager for Western/Mid-western States.
OOI CI LCA REVIEW August 2010 Ocean Observatories Initiative OOI Cyberinfrastructure Architecture Overview Michael Meisinger Life Cycle Architecture Review.
Methodology - Conceptual Database Design. 2 Design Methodology u Structured approach that uses procedures, techniques, tools, and documentation aids to.
1/26/2004TCSS545A Isabelle Bichindaritz1 Database Management Systems Design Methodology.
Value Set Resolution: Build generalizable data normalization pipeline using LexEVS infrastructure resources Explore UIMA framework for implementing semantic.
UK Repository Search Project Phase II Project Overview Phil Cross Vic Lyte September 2006.
CTAKES The clinical Text Analysis and Knowledge Extraction System.
Methodology - Conceptual Database Design
Information Systems Engineering. Lecture Outline Information Systems Architecture Information System Architecture components Information Engineering Phases.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Cracow Grid Workshop ‘06 17 October 2006 Execution Management and SLA Enforcement in Akogrimo Antonios Litke Antonios Litke, Kleopatra Konstanteli, Vassiliki.
Clinical Collaboration Platform Overview ST Electronics (Training & Simulation Systems) 8 September 2009 Research Enablers  Consulting  Open Standards.
Strategic Health IT Advanced Research Projects (SHARP) Area 4: Secondary Use if EHR Data Project Initiation Thursday April 29th PI: Christopher G Chute,
Clinical Data Normalization Dr. Chute Aims: Build generalizable data normalization pipeline Semantic normalization annotators involving LexEVS Establish.
Exploring ‘Workspaces’ Tom Visser, SARA compute and networking services, Amsterdam Garching Workshop 21 st September 2010.
Health Management Information Systems Unit 3 Electronic Health Records Component 6/Unit31 Health IT Workforce Curriculum Version 1.0/Fall 2010.
Open Source & Interoperability Profit Proprietary Closed Free Collaborative Open.
Process Asad Ur Rehman Chief Technology Officer Feditec Enterprise.
Part I: Introduction to SHARPn Normalization Hongfang Liu, PhD, Mayo Clinic Tom Oniki, PhD, Intermountain Healthcare.
S&I FRAMEWORK PROPOSED INITIATIVE SUMMARIES Dr. Douglas Fridsma Office of Interoperability and Standards December 10, 2010.
Open Health Natural Language Processing Consortium
Helping the Cause of Medical Device Interoperability Through Standards- based Test Tools DoC/NIST John J. Garguilo January 25,
SAGE Nick Beard Vice President, IDX Systems Corp..
Rule Engine for executing and deploying the SAGE-based Guidelines Jeong Ah Kim', Sun Tae Kim 2 ' Computer Education Department, Kwandong University, KOREA.
Data Infrastructure Building Blocks (DIBBS) NSF Solicitation Webinar -- March 3, 2016 Amy Walton, Program Director Advanced Cyberinfrastructure.
IoT R&I on IoT integration and platforms INTERNET OF THINGS
CIM Modeling for E&U - (Short Version)
DataNet Collaboration
SHARP: Secondary Use Project 1: Data Normalization
Strategic Health IT Advanced Research Projects (SHARP)
AI Discovery Template IBM Cloud Architecture Center
Implementing Sláintecare
Presentation transcript:

Strategic Health IT Advanced Research Projects (SHARP) Area 4: Secondary Use Dr. Friedman on-site visit, Mayo Clinic 3 September 2010

SHARP: Area 4: Secondary Use of EHR Data 14 academic and industry partners Develop tools and resources that influence and extend secondary uses of clinical data Cross-integrated suite of project and products Clinical Data Normalization Natural Language Processing (NLP) Phenotyping (cohorts and eligibility) Common pipeline tooling (UIMA) and scaling Data Quality (metrics, missing value management) Evaluation Framework (population networks) © 2009 Mayo Clinic2

Collaborations Agilex Technologies CDISC (Clinical Data Interchange Standards Consortium) Centerphase Solutions Deloitte Group Health, Seattle IBM Watson Research Labs University of Utah Harvard Univ. & i2b2 Intermountain Healthcare Mayo Clinic Minnesota HIE (MNHIE) MIT and i2b2 SUNY and i2b2 University of Pittsburgh University of Colorado

Themes & Projects

Major Achievements Foster social connections across projects Recognition by team members that not all problems must be solved within their team NLP and phenotypes Phenotypes and CEM normalization Shared responsibility for overlapping dependencies

The bookends - Projects 1&6 Data Normalization & Evaluation Christopher G. Chute Stan Huff (Peter Haug)

Overview Build generalizable data normalization pipeline Establish a globally available resource for health terminologies and value sets Establish and expand modular library of normalization algorithms Iteratively test normalization pipelines, including NLP where appropriate, against normalized forms, and tabulate discordance. Use cohort identification algorithms in both EMR data and EDW data. (normalize against CEMs)

Progress Designation of Clinical Element Models (CEMs) as canonical form Utilizing use case scenario’s (PAD, CPNA, etc) for CEM normalization. Exploration into generalizable CEM models – diagnosis, medications, labs. Development of processes/tools to identify relevant existing CEM models within CEM libraries Development of processes to identify missing CEMs for data (and classes of data) in use-cases Preliminary population of phenotype use-cases

Planned Adopt eMERGE EleMap tooling for CEMs to population canonical model Formalize Meaningful Use vocabularies into LexGrid server Design other components of Data Normalization framework (Terminology Services - NHIN connections) Model end-to-end flow needed to produce normalized data from structured data and unstructured (natural language) data: High level description of process for taking “wild-type” data instances to canonical CEM instances Applicability to use-case data as well as to general classes of data Adopt UMIA data flows for normalization services Examine Regenstreif and SHARP 3 modules

Project 2 Clinical Natural Language Processing (cNLP) Dr. Guergana Savova

Overview Overarching goal High-throughput phenotype extraction from clinical free text based on standards and the principle of interoperability Focus Information extraction (IE): transformation of unstructured text into structured representations (CEMs) Merging clinical data extracted from free text with structured data

Progress Detailed 4-year project plan Tasks in execution: Investigative tasks: (1) defining CEMs and attributes as normalization targets for NLP, (2) defining set of clinical named entities and their attributes, (3) methods for cNE Engineering tasks: (1) defining users, (2) incorporating site NLP tools into cTAKES and UIMA, (3) common conventions and requirements, (4) de-identification flow and data sharing Forging cross-SHARP collaborations (SHARP 3, PI Kohane and Mandl)

Planned Y1 Gold standard for cNEs, relations and CEMs Focus on methods for cNE discovery and populating relevant CEMs (many subtasks) Projected module releases: Medication extraction (Nov’10) CEM OrderMedAmb population (Mar’11) Deep parser for cTAKES (Nov’10) Dependency parser for cTAKES (Jan’11) Collaboration with SHARP 3 by providing medication extraction capabilities for the medication SMaRT app

Project 3 High throughput Phenotyping (HTP) Dr. Jyoti Pathak

Overview Overarching goal To develop techniques and algorithms that operate on normalized EMR data to identify cohorts of potentially eligible subjects on the basis of disease, symptoms, or related findings Focus Portability of phenotyping algorithms Representation of phenotyping logic Measure goodness of EMR data 06/21/10© 2010 Mayo Clinic15

Progress Explored use case phenotypes from eMERGE network for HTP process validation Representation of phenotype descriptions and data elements using Clinical Element Models Preliminary execution of phenotyping algorithms (Peripheral Arterial Disease) to compare aggregate data

Planned Interaction and collaboration with Data Normalization and NLP teams to develop “data collection widgets” Representation of phenotyping execution logic in a machine processable format/language Development of machine learning methods for semi-automatic cohort identification

Project 4 Infrastructure & Scalability Jeff Ferraro Marshal Schor Calvin Beebe

UIMA exploitation Some initial discussions on UIMA were held in a meeting at MIT attended by Peter Szolovits (MIT) and Guergana Savova (Harvard) and some of their team members. A plan is underway for a UIMA "deep dive" for other members from Intermountain Health and Mayo. A discussion is pending to understand the how UIMA might fit with RPE (in particular, BPEL) RPE = Retrieve Process for Execution: an IHE (Integrating the Health Enterprise) profile to automate collaborative workflow between healthcare and secondary use domains)

Infrastructure Progress Code repository – Reviewed requirements (e.g. SVN), need pre-release work areas for project teams, bulk of materials will all be in public repository. Licensing compatibility discussion. Initial discussions on Open Source licensing which is consistent with UIMA and other project teams tooling. Will need to survey teams. Initial platform discussions Still working on Sandbox (“Shared”) environment, need to consider Cloud in later phases of project.

Planned Review repository options with: ONC, Source Forge, Open Health Tools Need to establish straw man proposal for Sandbox configuration. Conduct cross-project discussions Inventory tools that can be shared. Inventory data that can be shared. Identify shared environment site location. Initiate high-level requirements gathering.

Project 5 Data Quality Dr. Kent Bailey (Kim Lemmerman)

Overview Support data quality and ascertain data quality issues across projects Deploy and enhance methods for missing or conflicting data resolution Integrate methods into UIMA pipelines

Progress & Planned Integrate across projects and gather requirements and standards to establish data quality plan and metrics Compare expected quality of data to actual data quality Provide recommendation and methods to improve data quality and/or possible outcomes

Cross-Area 4 Program Efforts Lacey Hart

Progress Started with early with face-to-face collaboration; cross-knowledge pollination Individual project efforts synergized with timelines in synch; use cases vetted and determined for the first six months of focus. IRB & Data Sharing issues have been raised with best practice sharing and inventory of existing agreements between institutions reviewed.

Planned Best practices for IRB submissions and template protocol material will be made available w/ applicable state implications Data use agreements will be completed across sites where needed in short term; effort for ‘consortium’ agreement will commence for long-term data sharing needs

Cross-ONC Efforts Dr. Christopher Chute

SHARP Area Synergies 1. Security: ensure piplined data does not have compromisable integrity 2. Cognitive: explore how normalized data and phenotypes can contribute to decisions 3. Applications: Potential for shared architectural strategies © 2009 Mayo Clinic29

Beacon Synergies High-throughput data normalization and phenotyping (SHARP) Applied to population laboratory (Beacon) Validate on consented sub-samples Potential to include ALL patients in population area – regardless of provider © 2009 Mayo Clinic30

SHARP Area 4: More information…

© 2009 Mayo Clinic32 SE MN Beacon: More information…