1 Modeling the fMRI signal via Hierarchical Clustered Hidden Process Models Stefan Niculescu, Tom Mitchell, R. Bharat Rao Siemens Medical Solutions Carnegie.

Slides:



Advertisements
Similar presentations
1 Challenge the future Multi-scale mining of fMRI data with hierarchical structured sparsity – R. Jenatton et al, SIAM Journal of Imaging Sciences, 2012.
Advertisements

ECG Signal processing (2)
Fast Algorithms For Hierarchical Range Histogram Constructions
Slide 1 Bayesian Model Fusion: Large-Scale Performance Modeling of Analog and Mixed- Signal Circuits by Reusing Early-Stage Data Fa Wang*, Wangyang Zhang*,
Christoph F. Eick Questions and Topics Review Nov. 22, Assume you have to do feature selection for a classification task. What are the characteristics.
The General Linear Model Or, What the Hell’s Going on During Estimation?
A (very) brief introduction to multivoxel analysis “stuff” Jo Etzel, Social Brain Lab
The use of unlabeled data to improve supervised learning for text summarization MR Amini, P Gallinari (SIGIR 2002) Slides prepared by Jon Elsas for the.
1 Exploiting Parameter Domain Knowledge for Learning in Bayesian Networks Thesis Committee: Tom Mitchell (Chair) John Lafferty Andrew Moore Bharat Rao.
Hidden Process Models with applications to fMRI data Rebecca Hutchinson Oregon State University Joint work with Tom M. Mitchell Carnegie Mellon University.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
1 Hidden Process Models Rebecca Hutchinson Joint work with Tom Mitchell and Indra Rustandi.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference (Sec. )
Evaluating Hypotheses
Hidden Process Models: Decoding Overlapping Cognitive States with Unknown Timing Rebecca A. Hutchinson Tom M. Mitchell Carnegie Mellon University NIPS.
1 Learning fMRI-Based Classifiers for Cognitive States Stefan Niculescu Carnegie Mellon University April, 2003 Our Group: Tom Mitchell, Luis Barrios, Rebecca.
1 Classifying Instantaneous Cognitive States from fMRI Data Tom Mitchell, Rebecca Hutchinson, Marcel Just, Stefan Niculescu, Francisco Pereira, Xuerui.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Hidden Process Models Rebecca Hutchinson Tom M. Mitchell Indrayana Rustandi October 4, 2006 Women in Machine Learning Workshop Carnegie Mellon University.
Presented by Zeehasham Rasheed
1 Automated Feature Abstraction of the fMRI Signal using Neural Network Clustering Techniques Stefan Niculescu and Tom Mitchell Siemens Medical Solutions,
Hidden Process Models Rebecca Hutchinson May 26, 2006 Thesis Proposal Carnegie Mellon University Computer Science Department.
Hidden Process Models for Analyzing fMRI Data Rebecca Hutchinson Joint work with Tom Mitchell May 11, 2007 Student Seminar Series In partial fulfillment.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Clustering with Bregman Divergences Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon, Joydeep Ghosh Presented by Rohit Gupta CSci 8980: Machine Learning.
Dimensionality Reduction for fMRI Brain Imaging Data Leman Akoglu Carnegie Mellon University, Computer Science Department Abstract Functional Magnetic.
Modeling fMRI data generated by overlapping cognitive processes with unknown onsets using Hidden Process Models Rebecca A. Hutchinson (1) Tom M. Mitchell.
Learning to Identify Overlapping and Hidden Cognitive Processes from fMRI Data Rebecca Hutchinson, Tom Mitchell, Indra Rustandi Carnegie Mellon University.
Chapter 5 Data mining : A Closer Look.
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
This week: overview on pattern recognition (related to machine learning)
Comparison of Boosting and Partial Least Squares Techniques for Real-time Pattern Recognition of Brain Activation in Functional Magnetic Resonance Imaging.
Cut-And-Stitch: Efficient Parallel Learning of Linear Dynamical Systems on SMPs Lei Li Computer Science Department School of Computer Science Carnegie.
1 Blockwise Coordinate Descent Procedures for the Multi-task Lasso with Applications to Neural Semantic Basis Discovery ICML 2009 Han Liu, Mark Palatucci,
Segmental Hidden Markov Models with Random Effects for Waveform Modeling Author: Seyoung Kim & Padhraic Smyth Presentor: Lu Ren.
Today’s Topics Chapter 2 in One Slide Chapter 18: Machine Learning (ML) Creating an ML Dataset –“Fixed-length feature vectors” –Relational/graph-based.
Improved Gene Expression Programming to Solve the Inverse Problem for Ordinary Differential Equations Kangshun Li Professor, Ph.D Professor, Ph.D College.
Annealing Paths for the Evaluation of Topic Models James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine* *James.
Using Neural Networks to Predict Claim Duration in the Presence of Right Censoring and Covariates David Speights Senior Research Statistician HNC Insurance.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
OLAP : Blitzkreig Introduction 3 characteristics of OLAP cubes: Large data sets ~ Gb, Tb Expected Query : Aggregation Infrequent updates Star Schema :
1 Preliminary Experiments: Learning Virtual Sensors Machine learning approach: train classifiers –fMRI(t, t+  )  CognitiveState Fixed set of possible.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
Contrasts & Inference - EEG & MEG Himn Sabir 1. Topics 1 st level analysis 2 nd level analysis Space-Time SPMs Time-frequency analysis Conclusion 2.
OLAP Recap 3 characteristics of OLAP cubes: Large data sets ~ Gb, Tb Expected Query : Aggregation Infrequent updates Star Schema : Hierarchical Dimensions.
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
MLE’s, Bayesian Classifiers and Naïve Bayes Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 30,
INTRODUCTION TO Machine Learning 3rd Edition
MVPD – Multivariate pattern decoding Christian Kaul MATLAB for Cognitive Neuroscience.
Effective Automatic Image Annotation Via A Coherent Language Model and Active Learning Rong Jin, Joyce Y. Chai Michigan State University Luo Si Carnegie.
Learning to distinguish cognitive subprocesses based on fMRI Tom M. Mitchell Center for Automated Learning and Discovery Carnegie Mellon University Collaborators:
Simulation Study for Longitudinal Data with Nonignorable Missing Data Rong Liu, PhD Candidate Dr. Ramakrishnan, Advisor Department of Biostatistics Virginia.
Statistical Analysis An Introduction to MRI Physics and Analysis Michael Jay Schillaci, PhD Monday, April 7 th, 2007.
ANSWERS!. Completing the Square Level 1 Answers Completing the Square Level 2 Answers.
ICCV 2007 National Laboratory of Pattern Recognition Institute of Automation Chinese Academy of Sciences Half Quadratic Analysis for Mean Shift: with Extension.
Computational Intelligence: Methods and Applications Lecture 26 Density estimation, Expectation Maximization. Włodzisław Duch Dept. of Informatics, UMK.
Learning to Decode Cognitive States from Brain Images Tom Mitchell et al. Carnegie Mellon University Presented by Bob Stark.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Glenn Fung, Murat Dundar, Bharat Rao and Jinbo Bi
Ch3: Model Building through Regression
Hidden Process Models with applications to fMRI data
'Linear Hierarchical Models'
School of Computer Science, Carnegie Mellon University
Hierarchical Models and
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Outlines Introduction & Objectives Methodology & Workflow
Presentation transcript:

1 Modeling the fMRI signal via Hierarchical Clustered Hidden Process Models Stefan Niculescu, Tom Mitchell, R. Bharat Rao Siemens Medical Solutions Carnegie Mellon University November, 2007

2 Outline  Introduction Hidden Process Models (HPMs) Hierarchical Clustered HPMs Experiments Summary / Future Work

3 …

4

5 Motivation / Related Work Challenges: –High dimensionality (thousands) –Few examples per task (tens) –Variability in subjects’ brains Discriminative Models: –SVMs, kNN, ANNs –Decent accuracy –Do not necessarily model the fMRI signal properly Generative Models: –Hidden Process Models (HPMs) Do not take into account similarities among voxels –Hierarchical Clustered HPMs Extensions of standard HPMs Automatically identify clusters of “similar” voxels to improve estimation

6 Outline Introduction  Hidden Process Models (HPMs) Hierarchical Clustered HPMs Experiments Summary / Future Work

7 Starplus Dataset Collected by Just et al. Trial: –read sentence –view picture –answer whether sentence describes picture 40 trials – 32 time slices (2/sec) –picture presented first in half of trials –sentence first in the other half Three possible objects: star, dollar, plus

8 It is true that the star is above the plus?

9

*

11

12 Hidden Process Models One observation (trial): N different trials: All trials and all Processes have equal length T

13 Outline Introduction Hidden Process Models (HPMs)  Hierarchical Clustered HPMs Experiments Summary / Future Work

14 Parameter Sharing in HPMs similar shape activity different amplitudes XvXv

15 Parameter Sharing in HPMs ~ Maximum Likelihood Estimation ~ l’(P,C) quadratic in (P,C): each square linear in P and C a maximizer for l is also a maximizer for l’

16 Parameter Sharing in HPMs ~ Maximum Likelihood Estimation ~

17 Parameter Sharing in HPMs ~ Hierarchical Partitioning Algorithm ~

18 Outline Introduction Hidden Process Models (HPMs) Hierarchical Clustered HPMs  Experiments Summary / Future Work

19 Parameter Sharing in HPMs ~ Experiments ~ We compare three models: –Based on Average (per trial) Likelihood –StHPM – Standard, per voxel HPM –ShHPM – One HPM for all voxels in an ROI (24 total) –HieHPM – Hierarchical HPM Effect of training set size (6 to 40) in CALC: –ShHPM biased here Better than StHPM at small sample size Worse at 40 examples – HieHPM – the best It can represent both models 106 times better data likelihood than StHPM at 40 examples StHPM needs 2.9 times more examples to catch up

20 Parameter Sharing in HPMs ~ Experiments ~ Performance over whole brain (40 examples): –HieHPM – the best 1792 times better data likelihood than StHPM Better than StHPM in 23/24 ROIs Better than ShHPM in 12/24 ROIs, equal in 11/24 –ShHPM – second best 464 times better data likelihood than StHPM Better than StHPM in 18/24 ROIs It has bias, but makes sense to share whole ROIs not involved in the cognitive task

21 Learned Voxel Clusters In the whole brain:  ~ 300 clusters  ~ 15 voxels / cluster In CALC:  ~ 60 clusters  ~ 5 voxels / cluster

22 Sentence Process in CALC

23 Outline Introduction Hidden Process Models (HPMs) Hierarchical Clustered HPMs Experiments  Summary / Future Work

24 Summary Extended on previous non-temporal models for describing the fMRI signal Introduced a new method of clustering voxels with similar activities based on Hidden Process Models Showed how taking advantage of domain knowledge improved over standard temporal models

25 Research Opportunities Evaluate this algorithm on multiple studies/subjects Take advantage of data from multiple subjects to learn better clusters Accommodate uncertainty in the onset times of the hidden processes Explore soft clustering

26 Questions ?

27 THE END