August 20, 2009 NEMO Year 1: From Theory to Application — Ontology-based analysis of ERP data

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

Early studies of the development of recognition memory in infants demonstrated a negative component over central leads (Nc) with greater amplitude in event-related.
Office of SA to CNS GeoIntelligence Introduction Data Mining vs Image Mining Image Mining - Issues and Challenges CBIR Image Mining Process Ontology.
Measure Projection Analysis
NEMO ERP Analysis Toolkit ERP Metric Extraction An Overview.
All slides © S. J. Luck, except as indicated in the notes sections of individual slides Slides may be used for nonprofit educational purposes if this copyright.
Electroencephalogram (EEG) and Event Related Potentials (ERP) Lucy J. Troup 28 th January 2008 CSU Symposium on Imaging.
Haishan Liu.  to identify correspondences (“mappings”) between ERP patterns derived from:  different data decomposition methods (temporal PCA and spatial.
New Technologies Supporting Technical Intelligence Anthony Trippe, 221 st ACS National Meeting.
Contrasts & Inference - EEG & MEG Outi Tuomainen & Rimona Weil mfd.
Pre-processing for EEG and MEG
Artifact (artefact) reduction in EEG – and a bit of ERP basics CNC, 19 November 2014 Jakob Heinzle Translational Neuromodeling Unit.
February 11, 2011 Overview of All-Hands Meeting Agenda Gwen Frishkoff
IVITA Workshop Summary Session 1: interactive text analytics (Session chair: Professor Huamin Qu) a) HARVEST: An Intelligent Visual Analytic Tool for the.
MAPPING RESULTS Experiments were performed on two simulated datasets, each using both metric sets. Cross spatial join: calculate the Euclidean distance.
OHBM Morning Workshop – June 20, 2009 Neurocognitive ontologies: Methods for sharing and integration of human brain data Neural ElectroMagnetic Ontologies.
Haishan Liu 1, Gwen Frishkoff 2, Robert Frank 1, Dejing Dou 1 1 University of Oregon 2 Georgia State University.
February 11, 2011 Overview of All-Hands Meeting Agenda Gwen Frishkoff
NEMO ERP Analysis Toolkit ERP Pattern Segmentation An Overview.
Haishan Liu 1, Gwen Frishkoff 2, Robert Frank 1, Dejing Dou 1 1 University of Oregon 2 Georgia State University.
September 16, 2009 NEMO OWL ontologies: Viewing & editing OWL/RDF files, part II
NEMO Data Analysis Workflow and MATLAB Tools
February 26, 2010 NEMO All-Hands Meeting: Overview of Day 1
Feb 28, 2010 NEMO data meta-analysis: Application of NEMO analysis workflow to consortium datasets (redux)
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
The Neural ElectroMagnetic Ontology (NEMO) System: Design & Implementation of a Sharable EEG/MEG Database with ERP ontologies G. A. Frishkoff 1,3 D. Dou.
Handwritten Character Recognition using Hidden Markov Models Quantifying the marginal benefit of exploiting correlations between adjacent characters and.
General Linear Model & Classical Inference
LÊ QU Ố C HUY ID: QLU OUTLINE  What is data mining ?  Major issues in data mining 2.
IE 594 : Research Methodology – Discrete Event Simulation David S. Kim Spring 2009.
Data Mining Chun-Hung Chou
ERP DATA ACQUISITION & PREPROCESSING EEG Acquisition: 256 scalp sites; vertex recording reference (Geodesic Sensor Net)..01 Hz to 100 Hz analogue filter;
July 17, 2009 NEMO Year 1: Overview & Planning
1 Dejing Dou Computer and Information Science University of Oregon, Eugene, Oregon September, Kent State University.
SVCL Automatic detection of object based Region-of-Interest for image compression Sunhyoung Han.
2 nd International Conference on Biomedical Ontology (ICBO’11) Ontology-Based Analysis of Event-Related Potentials Gwen Frishkoff 12, Robert Frank 2, Paea.
FMRI Methods Lecture7 – Review: analyses & statistics.
Studying Memory Encoding with fMRI Event-related vs. Blocked Designs Aneta Kielar.
All slides © S. J. Luck, except as indicated in the notes sections of individual slides Slides may be used for nonprofit educational purposes if this copyright.
Contrasts & Inference - EEG & MEG Himn Sabir 1. Topics 1 st level analysis 2 nd level analysis Space-Time SPMs Time-frequency analysis Conclusion 2.
EXPERIMENT DESIGN  Variations in Channel Density  The original 256-channel data were downsampled:  127 channel datasets  69 channels datasets  34.
TUH EEG Corpus Data Analysis 38,437 files from the Corpus were analyzed. 3,738 of these EEGs do not contain the proper channel assignments specified in.
Process Quality in ONS Rachel Skentelbery, Rachael Viles & Sarah Green
fMRI Task Design Robert M. Roth, Ph.D.
WERST – Methodology Group
Automatic Discovery and Processing of EEG Cohorts from Clinical Records Mission: Enable comparative research by automatically uncovering clinical knowledge.
Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav.
EEG DATA EEG Acquisition: 256 scalp sites; vertex recording reference (Geodesic Sensor Net)..01 Hz to 100 Hz analogue filter; 250 samples/sec. EEG Preprocessing:
Principal components analysis (PCA) as a tool for identifying EEG frequency bands: I. Methodological considerations and preliminary findings Jürgen Kayser,
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
CognitiveViews of Learning Chapter 7. Overview n n The Cognitive Perspective n n Information Processing n n Metacognition n n Becoming Knowledgeable.
A direct comparison of Geodesic Sensor Net (128-channel) and conventional (30-channel) ERPs in tonal and phonetic oddball tasks Jürgen Kayser, Craig E.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Title Authors/Collaborators Affiliations When was the project discussed in a research group meeting? Which group? Note (see also
Topographic mapping on memory test of verbal/spatial information assessed by event related potentials (ERPs) Petrini L¹², De Pascalis V², Arendt-Nielsen.
Development of NeuroElectroMagnetic Ontologies (NEMO): A Framework for Mining Brainwave Ontologies Dejing Dou 1, Gwen Frishkoff 2, Jiawei Rong 1, Robert.
[Ran Manor and Amir B.Geva] Yehu Sapir Outlines Review
A High-Density EEG investigation of the Misinformation Effect: Differentiating between True and False Memories John E. Kiat & Robert F. Belli Department.
Theoretical and empiral rationale for using unrestricted PCA solutions
Brain Initiative Informational Conference Call
Meng Lu and Edzer Pebesma
Optimizing Channel Selection for Seizure Detection
A Similarity Retrieval System for Multimodal Functional Brain Images
Dynamic Causal Modelling for M/EEG
CHAPTER 7: Information Visualization
Machine Learning for Visual Scene Classification with EEG Data
Parietal and Frontal Cortex Encode Stimulus-Specific Mnemonic Representations during Visual Working Memory  Edward F. Ester, Thomas C. Sprague, John T.
A Dissertation Proposal by: Vinit Shah
Theoretical rationale and empirical evaluation
Presentation transcript:

August 20, 2009 NEMO Year 1: From Theory to Application — Ontology-based analysis of ERP data

Overview Agenda ICBO highlights (5 mins) Logistics (5 mins) ERP pattern analysis methods (20 mins) ERP measure generation (10 mins) Linking measures to ontology (10 mins) Data annotation (deep, ontology-based) (10 mins) Action items highlighted in lime green!

Overview Agenda ICBO highlights (5 mins) Logistics (5 mins) ERP pattern analysis methods (20 mins) ERP measure generation (10 mins) Linking measures to ontology (10 mins) Data annotation (deep, ontology-based) (10 mins) Action items highlighted in lime green!

First International Conference on Biomedical Ontologies (ICBO’09)

First International Conference on Biomedical Ontologies (ICBO’09) High-level issues and "best practices" for onto dev't Tools that may be of use for NEMO Potential collaborations Practical Questions/Issues to resolve

Overview Agenda ICBO highlights (5 mins) Logistics (5 mins) ERP pattern analysis methods (20 mins) ERP measure generation (10 mins) Linking measures to ontology (10 mins) Data annotation (deep, ontology-based) (10 mins) Action items highlighted in lime green!

NEMO “to do” items Identify "point person" at each site who will be responsible for contributing feedback on NEMO wiki and ontologies and for uploading data and testing matlab-based tools for data markup – Please provide name & contact info for this person in an Bookmark NEMO website & explore links under “Collaboration” (more to come next time on how specifically you can contribute)

Overview Agenda ICBO highlights (5 mins) Logistics (5 mins) ERP pattern analysis methods (20 mins) ERP measure generation (10 mins) Linking measures to ontology (10 mins) Data annotation (deep, ontology-based) (10 mins) Action items highlighted in lime green!

ERP Pattern Analysis An embarrassment of riches – A wealth of data – A plethora of methods A lack of integration – How to compare patterns across studies, labs? – How to do valid meta-analyses in ERP research? A need for robust pattern classification – Bottom-up (data-driven) methods – Top-down (science-driven) methods

Ontologies for high- level, explicit representation of domain knowledge  theoretical integration Ontologies to support principled mark-up of data (inc. ERP patterns)  practical integration

NEMO principles that inform our pattern analysis strategies Current Challenges (motivations) – Tracking what we know Ontologies – Integrating knowledge to achieve high-level understanding of brain–functional mappings Meta-analyses Important Considerations (disiderata) – Stay true to data bottom-up (data-driven methods) – Achieve high-level understanding top-down (hypothesis-driven methods)

Top-down vs. Bottom-up Top-DownBottom-Up PROS Familiar Science-driven (integrative) Formalized Data-driven (robust) CONS Informal Paradigm- affirming? Unfamiliar Study-specific results?

Combining Top-Down & Bottom-Up

TOP-DOWN Traditional approach to bio-ontology dev’t Encode knowledge of concepts (=> classes, relations, & axioms that involve classes & relations) in a formal ontology (e.g., owl/rdf) NEMO owl ontologies being developed & version-tracked on Sourceforge (the main topic of our last meeting)

TOP-DOWN NEMO top-down approach NEMO emphasis on pattern rules/descriptions — way to enforce rigorous definitions Of complex concepts (patterns or “components”) that are central to ERP research

Superposition of ERP Patterns

What do we know about ERP patterns? Observed Pattern = “P100” iff  Event type is visual stimulus AND  Peak latency is between 70 and 160 ms AND  Scalp region of interest (ROI) is occipital AND  Polarity over ROI is positive (>0) FUNCTION TIME SPACE ?

Why does it matter? Robust pattern rules a good foundation for–  Development of ERP ontologies  Labeling of ERP data based on pattern rules  Cross-experiment, cross-lab meta-analyses

BOTTOM-UP

Two classes of methods for ERP pattern analysis Pattern decomposition – Temporal factor analysis (tPCA, tICA) – Spatial factor analysis (sPCA, sICA Windowing/segmentation – Microstate analysis (use global field “maps”; compute “global field dissimilarity” between adjacent maps to determine where there are significant shifts in topography Focus today (already implemented & almost ready for YOU to test )

Decomposition approach PCA, ICA, dipoles etc. multiple methods for principled separation of patterns using factor-analytic approach P100 N100 fP2 P1r/ N3 P1r/ MFN P ms 170ms 200ms 280ms 400ms 600ms

Windowing/segmentation approach P100 N100 fP2 P1r/ N3 P1r/ MFN P ms 170ms 200ms 280ms 400ms 600ms Michel, et al., 2004; Koenig, 1995; Lehmann & Skrandies, 1985 Advantages over factor-analytic/ decomposition methods: Familiarity — Closer to what most ERP researchers do (manually) Less (or at least different!) concerns regarding misallocation of variance Robustness to latency diffs across subjects, conditions

What we’ve done (to date…) Implemented sPCA, tPCA, sICA, & microstate analysis Tested & evaluated sPCA, tPCA & sICA (following Dien, Khoe, & Mangun, 2008) using simulated ERP data Explored two different approaches to pattern classification & labeling (the step AFTER decomposition)

1. Classification of Datasets 13+ EEG datasets collected so far (from Tucker/GF, Perfetti/GF, Curran). Waiting on datasets from Molfese, Kilborn, & maybe Dien. [using EGI data only for pilot study] Experimental paradigms/contrasts fall into 3 general categories (some expt yield multiple contrasts):  Semantic priming: prime–target relatedness (6)  Word vs. Nonword recognition (6)  Word/Nonword repetition: episodic memory,  “old/new” comparisons(7) Need studies for a reasonable meta-analysis (GF collecting more data from partners)

1. Data preprocessing 1. filter & segment data 2. detect & reject artifacts 3. interpolate bad channels 4. average across trials w/in subjects 5. manual detection of bad channels 6. interpolate bad channels 7. re-reference montage (PARE) 8. baseline-correct (200ms)

2. Component Analysis Our current practice (NOT set in stone!) - Step 1. Apply eigenvalue decomposition method (eg., tPCA) - Step 2: Rotate ALL latent factors (unrestricted PCA) - Step 3: Retain fairly large number of factors based on log of scree - Step 4: Let ontology-based labeling (next slide) help determine which factors to keep and analyze!

3. Component Labeling NEXT MAJOR CHALLENGE: How to tune pattern rules (particularly TI-max begin and end) to fit each individual dataset. Data mining on results from different component analyses? (Note mining of tPCA data won’t help to refine temporal criteria.)

4. Meta-analysis (next milestone!!) Apply pattern decomposition & labeling to NEMO consortium datasets Identify one experimental contrast for each analysis Compute Effect Size (ES) estimates for each study Run mixed effects analysis: test homogeneity of variance across studies if rejected, then test effects of variables that differ across studies, laboratories (e.g., nature of stimuli, task, subjects)

ERP Meta-analysis goals 1.Demonstrate working NEMO consortium 2.Demonstrate application of BrainMap-like taxonomy for classification of functional (experimental) contrasts. 3.Show that ERP component analysis, measure generation, and component labeling tools can be used on a large scale 4.** Show that combination of bottom-up and top-down methods for refining pattern rules can be used to tune rules for detecting target ERP patterns across different datasets 5.** Show that we can (semi-)automatically indentify analogous patterns across datasets (follows from 4), enabling us to carry out statistical meta-analyses ** harder problems to discuss…

A Case Study with real data (CIN’07 paper) 1.Real 128-channel ERP data 2.Temporal PCA used for pattern analysis 3.Spatial & temporal metrics for labeling of discrete patterns 4.Revision of pattern rules based on mining of labeled data

Example: Rule for “P100” For any n, FA n = PT 1 iff – temp criterion #1: 70ms > TI-max (FA n ) < 170ms AND – spat criterion #1 : SP-r (FA n, SP(PT 1 )) >.7 AND – func criterion #1: EVENT (FA n ) = stimon AND – func criterion #2: MODAL (EV) = visual AND

Example of output [1] values for summary measures (for one subject, one/six expt conditions)

Example of output [2] Matches to spatial, temporal & functional criteria for one subject & one/six experimental conditions

Summary results for Rule #1

A Case Study with simulated ERPs (HBM’08 tak) 1.Simulated ERP datasets 2.PCA & ICA methods for spatial & temporal pattern analysis 3.Spatial & temporal metrics for labeling of discrete patterns 4.Revision of pattern rules based on mining of labeled data

Simulated ERPs (n=80) P100 N100 N3 MFN P300 + NOISE

Simulated ERP Datasets (in DipSim) Dipole Simulator (P. Berg)

Patrick Berg’s Dipole Simulator Simulated ERP data: Creating individual ERPs Random jitter in intensity NO temporal jitter NO spatial jitter

BOTTOM-UP

Pattern Analysis with PCA & ICA (Decomposition approach)

ERP pattern analysis Temporal PCA (tPCA) – Gives invariant temporal patterns (new bases) – Spatial variability as input to data mining Spatial ICA (sICA) – Gives invariant spatial patterns (new bases) – Temporal variability as input to data mining Spatial PCA (sPCA) ✔ ✔ Multiple measures used for evaluation (correlation + L1/L2 norms) X

New inputs to NEMO PATTERN DEFINITIONS (Revised) “P100”1.70 ms < TI-max ≤ 140 ms 2. ROI = Occipital 3. IN-mean (ROI) > 0 “N100”1.141 ms < TI-max ≤ 220 ms 2. ROI = Occipital 3. IN-mean (ROI) < 0 “N3c”1.221 ms < TI-max ≤ 260 ms 2. ROI = Anterior Temporal 3. IN-mean (ROI) < 0 “MFN”1.261 ms < TI-max ≤ 400 ms 2. ROI = Mid Frontal 3. IN-mean (ROI) < 0 “P300”1.401 ms < TI-max ≤ 600 ms 2. ROI = Parietal 3. IN-mean (ROI) > 0 SPATIALTEMPORAL

What we’ve learned (so far…) Bottom-up methods result in validation & refinement of top-down pattern rules  Validation of expert selection of temporal concepts (peak latency)  Refinement of expert specification of spatial concepts (± centroids) Alternative pattern analysis methods (e.g., tPCA & sICA) provide complementary input to bottom- up (data mining) procedures

BOTTOM-UP

Measure Generation A B T1 T2 S1 S2 Vector attributes = Input to Data mining (clustering & classification) CoP CoN ROI± Centroids Input to data mining: 32 attribute vectors, defined over 80 “individual” ERPs (observations)

BOTTOM-UP

Data mining Vectors of spatial & temporal attributes as input Clustering observations  patterns (E-M accuracy >97%) Attribute selection (“Information gain”) Figure 3. Info gain results for spatial ICA. CoP CoN ✔ ± Centroids Peak Latency

Revised Rule for the “P100” Pattern = P100v iff  Event type is visual stimulus AND  Peak latency is between 76 and 155 ms AND  Positive centroid is right occipital AND  Negative centroid is left frontal SPACE TIME FUNCTION

Simulated ERP Patterns “P100”“N100”“N3”“MFN”“P300”

Alternative Spatial Metrics Scalp (ROI) “regions-of-intrest” Positive and negative “centroids” (topographic source & sink) CPOS CNEG

TPCA “Grand Average” Results Misallocation when there is high temporal, but low spatial correlation

SICA “Grand Average” Results Misallocation when there is high temporal and high spatial correlation

Labeling data Matches to pattern rules for one subject & one experiment condition

NeuroElectroMagnetic Ontologies (NEMO) 1. Statistical measure generation (basic concepts)  Temporal: peak latency, duration, onset of pattern  Spatial: electrode location, scalp “region,” ±centroid  Functional: subject, stimulus, task & measurement attributes 2. Labeling data (ERP patterns) w/ basic concepts 3. Analysis & mining of labeled data to derive pattern rules (complex concepts) data-driven 4. Coding of concepts in data-driven ontology

Overview Agenda ICBO highlights (5 mins) Logistics (5 mins) ERP pattern analysis methods (20 mins) ERP measure generation (10 mins) Linking measures to ontology (10 mins) Data annotation (deep, ontology-based) (10 mins) Action items highlighted in lime green!

BOTTOM-UP

Statistical Measure Generation Temporal – Peak latency – Duration (cf. spectral measures) Spatial (topographic) – Scalp regions-of-Interest (ROI) – Positive & negative centroids Functional (experimental) – Concepts borrowed from BrainMap (Laird et al.) where possible

Measure Generation A B T1 T2 S1 S2 Vector attributes = Input to Data mining (clustering & classification) CoP CoN ROI± Centroids Input to data mining: 32 attribute vectors, defined over 80 “individual” ERPs (observations)

Overview Agenda ICBO highlights (5 mins) Logistics (5 mins) ERP pattern analysis methods (20 mins) ERP measure generation (10 mins) Linking measures to ontology (10 mins) Data annotation (deep, ontology-based) (10 mins) Action items highlighted in lime green!

Automated ontology-based labeling of ERP data Pattern Labels Functional attributes Temporal attributes Spatial attributes =++ Robert M. Frank Concepts encoded in NEMO_Data.owl

NEMO Data Ontology: Where ontology meets epistemology Ontology for Biological Investigations (OBI) & Information Artifact Ontology (IAO) Ontology for Biological Investigations (OBI) & Information Artifact Ontology (IAO)

Overview Agenda ICBO highlights (5 mins) Logistics (5 mins) ERP pattern analysis methods (20 mins) ERP measure generation (10 mins) Linking measures to ontology (10 mins) Data annotation (deep, ontology-based) (10 mins) Action items highlighted in lime green!