Challenge in neuroscience  Neuroscience is a very broad field. It covers everything from gene expression, to a single neuron firing, to activity across.

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

Basis Functions. What’s a basis ? Can be used to describe any point in space. e.g. the common Euclidian basis (x, y, z) forms a basis according to which.
Slides from: Doug Gray, David Poole
Neural networks Introduction Fitting neural networks
SPM 2002 C1C2C3 X =  C1 C2 Xb L C1 L C2  C1 C2 Xb L C1  L C2 Y Xb e Space of X C1 C2 Xb Space X C1 C2 C1  C3 P C1C2  Xb Xb Space of X C1 C2 C1 
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Purpose The aim of this project was to investigate receptive fields on a neural network to compare a computational model to the actual cortical-level auditory.
More MR Fingerprinting
Artificial Neural Networks - Introduction -
Machine Learning Neural Networks
Introduction to Data-driven Animation Jinxiang Chai Computer Science and Engineering Texas A&M University.
Lecture 14 – Neural Networks
Pattern Recognition and Machine Learning
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Spike-triggering stimulus features stimulus X(t) multidimensional decision function spike output Y(t) x1x1 x2x2 x3x3 f1f1 f2f2 f3f3 Functional models of.
Neural Networks Marco Loog.
Introduction to Neural Networks Simon Durrant Quantitative Methods December 15th.
Face Recognition Jeremy Wyatt.
What is Cognitive Science? … is the interdisciplinary study of mind and intelligence, embracing philosophy, psychology, artificial intelligence, neuroscience,
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Neural Networks (NN) Ahmad Rawashdieh Sa’ad Haddad.
Data mining and statistical learning - lecture 11 Neural networks - a model class providing a joint framework for prediction and classification  Relationship.
Neural Decoding: Model and Algorithm for Evidence Accumulator Inference Thomas Desautels University College London Gatsby Computational Neuroscience Group.
Classification with Hyperplanes Defines a boundary between various points of data which represent examples plotted in multidimensional space according.
N. Laskaris. [ IEEE SP Magazine, May 2004 ] N. Laskaris, S. Fotopoulos, A. Ioannides ENTER-2001.
Genetic Regulatory Network Inference Russell Schwartz Department of Biological Sciences Carnegie Mellon University.
Artificial Neural Network Theory and Application Ashish Venugopal Sriram Gollapalli Ulas Bardak.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
Introduction to Artificial Neural Network Models Angshuman Saha Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg.
Outline 1-D regression Least-squares Regression Non-iterative Least-squares Regression Basis Functions Overfitting Validation 2.
NEURAL NETWORKS FOR DATA MINING
2 2  Background  Vision in Human Brain  Efficient Coding Theory  Motivation  Natural Pictures  Methodology  Statistical Characteristics  Models.
Methods Neural network Neural networks mimic biological processing by joining layers of artificial neurons in a meaningful way. The neural network employed.
Modelling Language Evolution Lecture 1: Introduction to Learning Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
Ch 5b: Discriminative Training (temporal model) Ilkka Aho.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 23: Linear Support Vector Machines Geoffrey Hinton.
Reverse engineering the brain Prof. Jan Lauwereyns Advanced Engineering A.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CSC321: Neural Networks Lecture 1: What are neural networks? Geoffrey Hinton
1 Azhari, Dr Computer Science UGM. Human brain is a densely interconnected network of approximately neurons, each connected to, on average, 10 4.
RiskTeam/ Zürich, 6 July 1998 Andreas S. Weigend, Data Mining Group, Information Systems Department, Stern School of Business, NYU 2: 1 Nonlinear Models.
Bump Hunting The objective PRIM algorithm Beam search References: Feelders, A.J. (2002). Rule induction by bump hunting. In J. Meij (Ed.), Dealing with.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Does the brain compute confidence estimates about decisions?
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Today’s Lecture Neural networks Training
Classification of models
Deep Feedforward Networks
Intelligent Information System Lab
Machine Learning Basics
Convolutional Networks
Machine Learning Today: Reading: Maria Florina Balcan
Contrasts & Statistical Inference
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
Contrasts & Statistical Inference
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Deep learning enhanced Markov State Models (MSMs)
Contrasts & Statistical Inference
Presentation transcript:

Challenge in neuroscience  Neuroscience is a very broad field. It covers everything from gene expression, to a single neuron firing, to activity across the whole brain in humans.  As such, one must have a wide range of knowledge and a diverse set of techniques.  Often makes it hard to have the best domain-specific knowledge.

Mapping the world onto the brain  The trick is to fit some function that links brain activity with the outside world.  However, we also want it to be a function that is scientifically meaningful.

Neuroscience/Psychology and computation  Historically, there has been a focus on tightly- controlled experiments and simple questions.  Advances in imaging and electrophysiological methods have increased the quality and quantity of data.

Electrocorticography – a blend of temporal and spatial resolution  ECoG involves the application of electrodes directly to the surface of the brain.  This avoid many problems with EEG, while retaining the rich temporal precision of the signal.

Complex and noisy data requires careful methods  ECoG is only possible in those with some sort of pathology. Moreover, recording time is short.  Data driven methods – bad data in = bad models out.

Merging ECoG and Computational Methods  Might be possible to leverage the spatial precision of ECoG to decode the nature of this processing.

Challenge 1: GLMs in Neuroscience

Computational Challenge #1 How to fit a model that is both interpretable and a good fit for the electrode’s response.  The parameter space is increasingly complex for more hypotheses.  Oftentimes, this is paired with a limited dataset. Especially in ECoG.  Regularization and Feature Selection become very important

Want it simple? Use a GLM! Linear models allow us to predict some output with a model that is both interpretable and (relatively) easy to fit.

One problem with this…  However, the brain assuredly does not vary linearly in response to inputs from the outside world.

Basis functions  Instead, we can decompose an input stimulus as a combination of “basis functions”  Basically, this entails a non-linear transformation of your stimulus, so that fitting linear models to brain activity makes more sense.

Exploring the brain through basis functions “dog” “hat” “car” “man”

Fitting weights with gradient descent  We can find the values for these weights by following the typical least-squares regression approach.  Early stopping must be tuned carefully in order to regularize.  Full gradient descent  Coordinate gradient descent  Threshold gradient descent

An application of the GLM for neural decoding

Neural Decoding  If you can map stimuli onto brain activity, then you could also map brain activity onto stimuli.  Same approach, but now our inputs are values from the electrodes, and the output is sound.  Implications in Neural Prostheses and Brain Computer Interfaces Speech Decoding

Decoding with a linear model Original Spectrogram Reconstructed Spectrogram High Gamma Neural Signal Decoding Model = X

Pasley et al. Plos Biology, 2012 Decoding Listened Speech High Gamma ( Hz)

Speech Reconstruction from ECoG

Challenge 2: From model output to language

Challenge #2 Turn a noisy, variable spectrogram reconstruction into linguistic output.  Simpler methods are often not powerful enough to account for these small variations  How to take advantage of temporal correlations between words / phonemes?  How to accomplish this without a ton of data?

How to classify this output? TownDoubtPropertyPencil

From model output to language  Borrow ideas from the speech recognition literature.  Currently using Dynamic Time Warping to match output spectrograms to words.

Dynamic Time Warping  Compute dissimilarity matrix between every pair of elements  Find the optimal path in order to minimize the overall accumulated distance  Effectively warps and realigns the two signals

Current output workflow Reconstructed Spectrogram DTW

Where to go from here?

Improving the decoder fit  Clever methods of dealing with finite and noisy datasets  Finding better features (basis functions)  Interactions between features  Fitting more complicated models  Interactions between features  Nonlinear models are useful for engineering, but require much more data

Turning output into reconstructed language  Leverage the spectro-temporal statistics of language  Focus on a classification rather than arbitrary decoding /ch//ks/ /w/ /g/

The “Big Data” Angle  Right now, the field of ECoG is in a bit of a transition period  Excitement around using computational methods, but many labs (including my own) don’t have the infrastructure and culture to tackle “big data” problems.  That said, we do have the potential to collect increasingly large datasets, once we know what to do with them.

The Long-Term Goal Create a modeling framework that allows us to use ECoG to decode linguistic information.

Fellow Decoders Special thanks Frederic Theunissen and co. Jack Gallant and co. STRFLab Team Stéphanie Brian GervEddie Peter

Linguistic features for model output  Hidden Markov Models allow us to model spectrogram output as a function of “hidden states”  Capture the probabilistic nature of spectrogram output for a given word, as well as the temporal correlations between components of that word. /ch//ks/ /w/ /g/

Designing stimulus sets  Data collection is very rare – we’re lucky if we get 2 subjects per month.  Need to be clever about how we design our behavioral tasks.  Stimuli must be rich, and ideally could be used to answer many different questions.  Need to think about what kind of stimuli we need in order to achieve the goal we prioritize.  E.g., classification vs. regression

The Big Data Angle  We have access to some ECoG recordings of a patient simply sitting in a room with a microphone placed nearby.  These are often >24 hours long, and include a wide range of sounds and speech.  Being able to parse through this data might allow us to fit increasingly complicated models, and vastly improve the speech recognition approach.