Deep Residual Learning for Automatic Seizure Detection

Slides:



Advertisements
Similar presentations
Corpus Development EEG signal files and reports had to be manually paired, de-identified and annotated: Corpus Development EEG signal files and reports.
Advertisements

Abstract EEGs, which record electrical activity on the scalp using an array of electrodes, are routinely used in clinical settings to.
Automatic Labeling of EEGs Using Deep Learning M. Golmohammadi, A. Harati, S. Lopez I. Obeid and J. Picone Neural Engineering Data Consortium College of.
Abstract The emergence of big data and deep learning is enabling the ability to automatically learn how to interpret EEGs from a big data archive. The.
1 RECENT DEVELOPMENTS IN MULTILAYER PERCEPTRON NEURAL NETWORKS Walter H. Delashmit Lockheed Martin Missiles and Fire Control Dallas, TX 75265
The Goal Use computers to aid physicians in diagnosis of neurological diseases, particularly epilepsy Detect pathological events in real time Currently,
TUH EEG Corpus Data Analysis 38,437 files from the Corpus were analyzed. 3,738 of these EEGs do not contain the proper channel assignments specified in.
Big Mechanism for Processing EEG Clinical Information on Big Data Aim 1: Automatically Recognize and Time-Align Events in EEG Signals Aim 2: Automatically.
Automatic Discovery and Processing of EEG Cohorts from Clinical Records Mission: Enable comparative research by automatically uncovering clinical knowledge.
Abstract Automatic detection of sleep state is important to enhance the quick diagnostic of sleep conditions. The analysis of EEGs is a difficult time-consuming.
Demonstration A Python-based user interface: Waveform and spectrogram views are supported. User-configurable montages and filtering. Scrolling by time.
Generating and Using a Qualified Medical Knowledge Graph for Patient Cohort Retrieval from Big Clinical Electroencephalography (EEG) Data Sanda Harabagiu,
Abstract Automatic detection of sleep state is an important queue in accurate detection of sleep conditions. The analysis of EEGs is a difficult time-consuming.
Deep Residual Learning for Image Recognition
Deep Learning Overview Sources: workshop-tutorial-final.pdf
The Neural Engineering Data Consortium Mission: To focus the research community on a progression of research questions and to generate massive data sets.
Constructing Multiple Views The architecture of the window was altered to accommodate switching between waveform, spectrogram, energy, and all combinations.
Descriptive Statistics The means for all but the C 3 features exhibit a significant difference between both classes. On the other hand, the variances for.
Big data classification using neural network
Deep Residual Learning for Image Recognition
Convolutional Sequence to Sequence Learning
Deep Residual Networks
Scalable EEG interpretation using Deep Learning and Schema Descriptors
Deep Feedforward Networks
Environment Generation with GANs
Summary of “Efficient Deep Learning for Stereo Matching”
Computer Science and Engineering, Seoul National University
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
CLASSIFICATION OF SLEEP EVENTS IN EEG’S USING HIDDEN MARKOV MODELS
G. Suarez, J. Soares, S. Lopez, I. Obeid and J. Picone
THE TUH EEG SEIZURE CORPUS
Robust Lung Nodule Classification using 2
Inception and Residual Architecture in Deep Convolutional Networks
Intro to NLP and Deep Learning
Mini Presentations - part 2
Convolutional Networks
CS 698 | Current Topics in Data Science
Deep Residual Learning for Image Recognition
Neural Networks and Backpropagation
N. Capp, E. Krome, I. Obeid and J. Picone
Goodfellow: Chap 6 Deep Feedforward Networks
Grid Long Short-Term Memory
Presented by Xinxin Zuo 10/20/2017
SBNet: Sparse Blocks Network for Fast Inference
Optimizing Channel Selection for Seizure Detection
Deep Learning Hierarchical Representations for Image Steganalysis
EEG Recognition Using The Kaldi Speech Recognition Toolkit
GATED RECURRENT NETWORKS FOR SEIZURE DETECTION
Deep learning Introduction Classes of Deep Learning Networks
Big Data Resources for EEGs: Enabling Deep Learning Research
To learn more, visit The Neural Engineering Data Consortium Mission: To focus the research community on a progression of research questions.
AN ANALYSIS OF TWO COMMON REFERENCE POINTS FOR EEGS
E. von Weltin, T. Ahsan, V. Shah, D. Jamshed, M. Golmohammadi, I
Smart Robots, Drones, IoT
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Neural Networks Geoff Hulten.
Vinit Shah, Joseph Picone and Iyad Obeid
Automatic Interpretation of EEGs for Clinical Decision Support
feature extraction methods for EEG EVENT DETECTION
LECTURE 42: AUTOMATIC INTERPRETATION OF EEGS
LECTURE 41: AUTOMATIC INTERPRETATION OF EEGS
EEG Event Classification Using Deep Learning
Inception-v4, Inception-ResNet and the Impact of
EEG Event Classification Using Deep Learning
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
This material is based upon work supported by the National Science Foundation under Grant #XXXXXX. Any opinions, findings, and conclusions or recommendations.
Natalie Lang Tomer Malach
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
Learning and Memorization
Presentation transcript:

Deep Residual Learning for Automatic Seizure Detection M. Golmohammadi, I. Obeid and J. Picone The Neural Engineering Data Consortium, Temple University College of Engineering Temple University www.nedcdata.org Abstract Automated seizure detection using clinical electroencephalograms (EEGs) is a challenging machine learning problem owing to low signal to noise ratios, signal artifacts and benign variants. Commercially available seizure detection systems suffer from unacceptably high false alarm rates. Deep learning algorithms, like Convolutional Neural Networks (CNNs), have not previously been effective due to the lack of big data resources. A significant big data resource, known as TUH EEG Corpus, has recently become available for EEG interpretation creating a unique opportunity to advance technology using CNNs. It is also proved that the CNNs depth is of crucial importance, and the state of the art results can be achieved by exploiting very deep models. On the other hand, very deep models are prone to degradation and convergence problems. In this study, a deep residual learning framework for automatic seizure detection task is introduced that mitigates these problem by reformulating the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. This architecture delivers 30% sensitivity at 13 false alarms per 24 hours. Our work enables designing deeper architectures that are easier to optimize and can achieve better performance from considerably increased depth. In a previous study by our team, two state of the art systems were designed for automatic seizure detection: CNN/MLP: Two-dimensional decoding of EEG signals using a CNN/MLP hybrid architecture is presented at below that consists of six convolutional layers, three max pooling layers and two fully-connected layers. CNN/LSTM: Deep recurrent convolutional architecture for two-dimensional decoding of EEG signals is shown at below that integrates 2D CNNs, 1-D CNNs and LSTM networks. Performance on Clinical Data Performance on TUSZ: The results are reported in Any-Overlap Method (OVLP). TPs are counted when the hypothesis overlaps with reference annotation. FPs correspond to situations in which the hypothesis does not overlap with the reference. A DET curve comparing performance on TUSZ: In conclusion, while ResNet can significantly improve the results of CNN/MLP, it cannot deliver better results than CNN/LSTM. System Sensitivity Specificity FA/24 Hrs. CNN/MLP 39.09% 76.84% 77 CNN/LSTM 30.83% 96.86% 7 ResNet 30.50% 94.24% 13.78 CNN/LSTM CNN/MLP Formally, denoting the desired underlying mapping as H(x), we let the stacked nonlinear layers fit another mapping of F(x) = H(x) - x. The original mapping is recast into F(x) + x. It is easier to optimize the residual mapping than to optimize the original, unreferenced mapping. To the extreme if an identity mapping were optimal, it would be easier to push the residual to zero than to fit an identity mapping by a stack of nonlinear layers. From implementation perspective, there are two different approaches for defining deep learning models in tools like Keras and TensorFlow: Sequential: applied in design of CNN/MLP and CNN/LSTM architectures. Functional API: applied in design of deep residual learning architecture (ResNet). The identity shortcuts (x) can be directly used when the input and output are of the same dimensions. A deep residual learning structure is designed for automatic seizure detection. We arrived at an architecture which is 14 layers of convolution followed by a fully connected layer and a sigmoid. The network consists of 6 residual blocks with two 2D convolutional layers per block. The 2D convolutional layers all have a filter length of (3, 3). The first 7 layers of 2D-CNN have 32 and the last layers have 64 filters. Except for the first and last layers of the network, before each convolutional layer we apply a rectified linear activation. We apply Dropout between the convolutional layers and after ReLU. We use the Adam optimizer with parameters of lr=0.00005, beta_1=0.9, beta_2=0.999, epsilon=1e-08, decay=0.0001. There are two obstacles for increasing the depth of CNN: Convergence Problem: created by vanishing/exploding gradients; Solution is normalized initialization and intermediate normalization layers. Degradation Problem: By increasing the number of convolutional layers, the accuracy gets saturated and starts degrading rapidly. Solution is residual learning. The core idea of ResNet is introducing an “identity shortcut connection” that skips one or more layers, as shown in the residual learning block. Training the stacked layers to fit a residual mapping is easier than training them directly to fit the desired underlying mapping. Weight Layer Activation + x F(x) F(x) + x TUH EEG Seizure Detection Subset Subset of the publicly available TUH EEG Corpus (www.isip.piconepress.com/projects/tuh_eeg). Evaluation Data: 50 patients, 239 sessions, 1015 files 171 hours of data including 16 hours of seizures. Training Data: 264 patients, 584 sessions, 1989 files 330 hours of data including 21 hours of seizures. Seizure event annotations include: start and stop times; localization of a seizure (e.g., focal, generalized) with the appropriate channels marked; type of seizure (e.g., simple partial, complex partial, tonic-clonic, gelastic, absence, atonic); nature of the seizure (e.g., convulsive) Residual learning block Summary For the first time, a deep residual learning structure was developed for automatic seizure detection. The result of ResNet was compared against two other structures: CNN/MLP and CNN/LSTM. The ResNet structure outperforms CNN/MLP. However CNN/LSTM delivers better results than ResNet. The key to the better performance of ResNet in comparison with CNN/MLP is very deep convolutional network. Acknowledgements Research reported in this publication was most recently supported by the National Human Genome Research Institute of the National Institutes of Health under award number U01HG008468. This material is also based in part upon work supported by the National Science Foundation under Grant No. IIP-1622765. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation. Deep Residual Learning Structure