The Team Adriano Claro Monteiro Alain de Cheveign Anahita Mehta

Slides:



Advertisements
Similar presentations
Time-dependent cortical asymmetries induced by emotional arousal
Advertisements

Time-Frequency Analysis Analyzing sounds as a sequence of frames
Combined classification and channel/basis selection with L1-L2 regularization with application to P300 speller system Ryota Tomioka & Stefan Haufe Tokyo.
Hearing relative phases for two harmonic components D. Timothy Ives 1, H. Martin Reimann 2, Ralph van Dinther 1 and Roy D. Patterson 1 1. Introduction.
Signal processing techniques for fNIRS and application to Brain Computer Interfaces Gautier Durantin, ISAE/CERCO French community for functional NIRS.
Brain-computer interfaces: classifying imaginary movements and effects of tDCS Iulia Comşa MRes Computational Neuroscience and Cognitive Robotics Supervisors:
SPEECH PERCEPTION 2 DAY 17 – OCT 4, 2013 Brain & Language LING NSCI Harry Howard Tulane University.
The frequency spectrum
AS14.1 – The Role of Neural Oscillations in Auditory Resolution and Speech Perception – Implications for CAPD, Phonological Awareness and Reading Disorders.
Chapter 1: Information and Computation. Cognitive Science  José Luis Bermúdez / Cambridge University Press 2010 Overview Review key ideas from last few.
Induced Brain Waves by Binaural Beats: A Study on Numerosity.
A.Diederich– International University Bremen – Sensation and Perception – Fall Frequency Analysis in the Cochlea and Auditory Nerve cont'd The Perception.
The Science of Sound Chapter 8
Discussion Section: Review, Viirre Lecture Adrienne Moore
Human Auditory Cognition 2014 Alain de Cheveigné, Andre van Schaik, Chetan Singh Thakur, David Karpul, Dorothee Arzounian, Edmund Lalor, Ernst Niebur,
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Sound source segregation (determination)
The Science of Sound Chapter 8
Human Psychoacoustics shows ‘tuning’ for frequencies of speech If a tree falls in the forest and no one is there to hear it, will it make a sound?
Applications of Independent Component Analysis Terrence Sejnowski Computational Neurobiology Laboratory The Salk Institute.
Measuring the brain’s response to temporally modulated sound stimuli Chloe Rose Institute of Digital Healthcare, WMG, University of Warwick, INTRODUCTION.
ECSE 6610 Pattern Recognition Professor Qiang Ji Spring, 2011.
COMBINATION TONES The Science of Sound Chapter 8 MUSICAL ACOUSTICS.
AUDIO COMPRESSION msccomputerscience.com. The process of digitizing audio signals is called PCM PCM involves sampling audio signal at minimum rate which.
Human Auditory Cognition Adriano Claro Monteiro, Alain de Cheveign, Anahita Mehta, Byron Galbraith, Dimitra Emmanouilidou, Deniz Erdogmus, James OSullivan,
1. 2 Abstract - Two experimental paradigms : - EEG-based system that is able to detect high mental workload in drivers operating under real traffic condition.
Graz-Brain-Computer Interface: State of Research By Hyun Sang Suh.
Basics of Neural Networks Neural Network Topologies.
Authors: Sriram Ganapathy, Samuel Thomas, and Hynek Hermansky Temporal envelope compensation for robust phoneme recognition using modulation spectrum.
Image Processing Architecture, © 2001, 2002, 2003 Oleh TretiakPage 1 ECE-C490 Image Processing Architecture MP-3 Compression Course Review Oleh Tretiak.
Classifying Event-Related Desynchronization in EEG, ECoG, and MEG Signals Kim Sang-Hyuk.
Distributed Representative Reading Group. Research Highlights 1Support vector machines can robustly decode semantic information from EEG and MEG 2Multivariate.
Analysis of Movement Related EEG Signal by Time Dependent Fractal Dimension and Neural Network for Brain Computer Interface NI NI SOE (D3) Fractal and.
Using Feed Forward NN for EEG Signal Classification Amin Fazel April 2006 Department of Computer Science and Electrical Engineering University of Missouri.
Analysis Techniques for Weak Brain Activity Alain de Cheveigné CNRS / Université Paris Descartes / École normale supérieure / UCL Ear Institute AIM Isolate.
EEG – BASED EMOTION RECOGNITION in MUSIC LEARNING.
EE 4BD4 Lecture 11 The Brain and EEG 1. Brain Wave Recordings Recorded extra-cellularly from scalp (EEG) Recorded from extra-cellularly from surface of.
BY KALP SHAH Sentence Recognizer. Sphinx4 Sphinx4 is the best and versatile recognition system. Sphinx4 is a speech recognition system which is written.
Cultural Differences and Similarities in Emotion Recognition Vladimir Kurbalija, Mirjana Ivanović, Miloš Radovanović, Zoltan Geler, Dejan Mitrović, Weihui.
Efficient Encoding of Vocalizations in the Auditory Midbrain Lars Holmstrom Systems Science PhD Program Portland State University.
BCI2000: 2D Control. Getting Started Follow the Passive Stimulus Presentation Data Collection Tutorial on the wiki – However, when the tutorial tells.
Chapter 15: Classification of Time- Embedded EEG Using Short-Time Principal Component Analysis by Nguyen Duc Thang 5/2009.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
All slides © S. J. Luck, except as indicated in the notes sections of individual slides Slides may be used for nonprofit educational purposes if this copyright.
What is Sound?? n “If a tree falls in the forest and no one is there to hear it, does it make a sound?” n There are 2 ways to answer this: n Objectively.
Sundeep Teki Wellcome Trust Centre for Neuroimaging University College London, UK Auditory figure-ground segregation in complex acoustic scenes.
COMBINATION TONES The Science of Sound Chapter 8 MUSICAL ACOUSTICS.
Detection Of Anger In Telephone Speech Using Support Vector Machine and Gaussian Mixture Model Prepared By : Siti Marahaini Binti Mahamood.
[Ran Manor and Amir B.Geva] Yehu Sapir Outlines Review
* the sampling rate and filter bandwidth were set to 500Hz and 1-10 Hz, respectively. * an additional 60Hz notch filter was employed to avoid the power-line.
When to engage in interaction – and how
Advanced applications of the GLM: Cross-frequency coupling
Information-Theoretic Listening
Advanced applications of the GLM: Cross-frequency coupling
Temporal Processing and Adaptation in the Songbird Auditory Forebrain
Volume 77, Issue 5, Pages (March 2013)
Perceptual Echoes at 10 Hz in the Human Brain
Sam Norman-Haignere, Nancy G. Kanwisher, Josh H. McDermott  Neuron 
Giovanni M. Di Liberto, James A. O’Sullivan, Edmund C. Lalor 
Cycle 10: Brain-state dependence
Dynamic Causal Modelling for M/EEG
Jason Samaha, Bradley R. Postle  Current Biology 
Progress Seminar 권순빈.
Volume 54, Issue 6, Pages (June 2007)
Machine Learning for Visual Scene Classification with EEG Data
Uses of filters To remove unwanted components in a signal
Temporal Processing and Adaptation in the Songbird Auditory Forebrain
Stefano Panzeri, Jakob H. Macke, Joachim Gross, Christoph Kayser 
Advanced applications of the GLM: Cross-frequency coupling
EEG and MEG dynamics give clues as to the role of macrodynamics (producing EEG and MEG records) in brain function. Here, a cluster of lateral occipital.
Presentation transcript:

Human Cognition: Decoding Perceived, Attended, Imagined Acoustic Events and Human-Robot Interfaces

The Team Adriano Claro Monteiro Alain de Cheveign Anahita Mehta Byron Galbraith Dimitra Emmanouilidou Edmund Lalor Deniz Erdogmus Jim O’Sullivan Mehmet Ozdas Lakshmi Krishnan Malcolm Slaney Mike Crosse Nima Mesgarani Jose L Pepe Contreras-Vidal Shihab Shamma Thusitha Chandrapala

The Goal To determine a reliable measure of imagined audition using electroencephalography (EEG). To use this measure to communicate.

What types of imagined audition? Speech: Short (~3-4s) sentences “The whole maritime population of Europe and America.” “Twinkle-twinkle little star.” “London bridge is falling down, falling down, falling down.” Music Short (~3-4s) phrases Imperial March from Star Wars. Simple sequence of tones. Steady-State Auditory Stimulation 20 s trials Broadband signal amplitude modulated at 4 or 6 Hz

The Experiment 64 – channel EEG system (Brain Vision LLC – thanks!) 500 samples/s Each “trial” consisted of the presentation of the actual auditory stimulus (“perceived” condition) followed (2 s later) by the subject imagining hearing that stimulus again (“imagined” condition).

+ The Experiment 4, 3, 2, 1, Careful control of experimental timing. Perceived...2s... Imagined...2 s x 5 ... Break... next stimulus 4, 3, 2, 1, +

Data Analysis - Preprocessing Filtering Independent Component Analysis (ICA) Time-Shift Denoising Source Separation (DSS) Looks for reproducibility over stimulus repetitions

Data Analysis: Hypothesis-driven. The hypothesis: EEG recorded while people listen to (actual) speech varies in a way that relates to the amplitude envelope of the presented (actual) speech. EEG recorded while people IMAGINE speech will vary in a way that relates to the amplitude envelope of the IMAGINED speech.

Data Analysis: Hypothesis-driven. Phase consistency over trials... EEG from same sentence imagined over several trials should show consistent phase variations. EEG from different imagined sentences should not show consistent phase variations.

Data Analysis: Hypothesis-driven. Actual speech Imagined speech Consistency in theta (4-8Hz) band Consistency in alpha (8-14Hz) band

Data Analysis: Hypothesis-driven.

Data Analysis: Hypothesis-driven. Red line – perceived music Green line – imagined music

Data Analysis - Decoding

Data Analysis - Decoding Original Reconstruction London’s Bridge Twinkle Twinkle r = 0.30, p = 3e-5 r = 0.19, p = 0.01

Data Analysis - SSAEP

Data Analysis - SSAEP Perceived Imagined 4Hz 6Hz

Data Analysis Data Mining/Machine Learning Approaches:

Data Analysis Data Mining/Machine Learning Approaches:

SVM Classifier Input ⋮ ⋮ Class Labels EEG data (channels × time) : 1 ⋮ EEG data (channels × time) : 𝐸𝐸𝐺= 𝑒 𝑡 1 ⋮ 𝑒(𝑡) 64 Concatenate channels: 𝐸𝐸𝐺= 𝑒 𝑡 1 … 𝑒(𝑡) 64 Group N trials: Predicted Labels 𝑋= 𝐸𝐸𝐺 1 ⋮ 𝐸𝐸𝐺 𝑁 1 ⋮ Input covariance matrix: 𝐶 𝑋 =𝑋 𝑋 𝑇

SVM Classifier Results Decoding imagined speech and music: Mean DA = 87% Mean DA = 90% Mean DA = 90%

DCT Processing Chain Raw EEG Signal (500Hz data) DSS Output (Look for repeatability) DCT Output (Reduce dimensionality)

DCT Classification Performance Percentage accuracy

Data Analysis Data Mining/Machine Learning Approaches: Linear Discriminant Analysis on Different Frequency Bands Music vs Speech Speech 1 vs Speech 2 Music 1 vs Music 2 Speech vs Rest Music vs Rest - results ~ 50 – 66%

Summary Both hypothesis drive and machine-learning approaches indicate that it is possible to decode/classify imagined audition Many very encouraging results that align with our original hypothesis More data needed!! In a controlled environment!! To be continued...