A Bifurcation Theoretical Approach to the Solving the Neural Coding Problem June 28 Albert E. Parker Complex Biological Systems Department of Mathematical.

Slides:



Advertisements
Similar presentations
Feature Selection as Relevant Information Encoding Naftali Tishby School of Computer Science and Engineering The Hebrew University, Jerusalem, Israel NIPS.
Advertisements

Statistical Techniques I EXST7005 Sample Size Calculation.
Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov Tomas Gedeon John P. Miller.
Sorted list matching & Experimental run-Time COP 3502.
Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data John Lafferty Andrew McCallum Fernando Pereira.
Continuation and Symmetry Breaking Bifurcation of the Information Distortion Function September 19, 2002 Albert E. Parker Complex Biological Systems Department.
Gizem ALAGÖZ. Simulation optimization has received considerable attention from both simulation researchers and practitioners. Both continuous and discrete.
On Constrained Optimization Approach To Object Segmentation Chia Han, Xun Wang, Feng Gao, Zhigang Peng, Xiaokun Li, Lei He, William Wee Artificial Intelligence.
Hidden Markov Models Ellen Walker Bioinformatics Hiram College, 2008.
Planning under Uncertainty
Visual Recognition Tutorial
COMPUTER MODELS IN BIOLOGY Bernie Roitberg and Greg Baker.
Brian Merrick CS498 Seminar.  Introduction to Neural Networks  Types of Neural Networks  Neural Networks with Pattern Recognition  Applications.
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Modelling and Control Issues Arising in the Quest for a Neural Decoder Computation, Control, and Biological Systems Conference VIII, July 30, 2003 Albert.
Symmetry Breaking Bifurcations of the Information Distortion Dissertation Defense April 8, 2003 Albert E. Parker III Complex Biological Systems Department.
Today: Entropy Information Theory. Claude Shannon Ph.D
Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov John P. Miller Zane Aldworth.
Symmetry Breaking Bifurcation of the Distortion Problem Albert E. Parker Complex Biological Systems Department of Mathematical Sciences Center for Computational.
Economics 214 Lecture 37 Constrained Optimization.
We use Numerical continuation Bifurcation theory with symmetries to analyze a class of optimization problems of the form max F(q,  )=max (G(q)+  D(q)).
Symmetry breaking clusters when deciphering the neural code September 12, 2005 Albert E. Parker Department of Mathematical Sciences Center for Computational.
Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov John P. Miller Zane Aldworth.
© 2005, it - instituto de telecomunicações. Todos os direitos reservados. Gerhard Maierbacher Scalable Coding Solutions for Wireless Sensor Networks IT.
CS 104 Introduction to Computer Science and Graphics Problems Data Structure & Algorithms (3) Recurrence Relation 11/11 ~ 11/14/2008 Yang Song.
Code and Decoder Design of LDPC Codes for Gbps Systems Jeremy Thorpe Presented to: Microsoft Research
Collaborators: Tomas Gedeon Alexander Dimitrov John P. Miller Zane Aldworth Information Theory and Neural Coding PhD Oral Examination November 29, 2001.
We use Numerical continuation Bifurcation theory with symmetries to analyze a class of optimization problems of the form max F(q,  )=max (G(q)+  D(q)).
Neural Coding Through The Ages February 1, 2002 Albert E. Parker Complex Biological Systems Department of Mathematical Sciences Center for Computational.
Sufficient Dimensionality Reduction with Irrelevance Statistics Amir Globerson 1 Gal Chechik 2 Naftali Tishby 1 1 Center for Neural Computation and School.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
NIPS 2003 Workshop on Information Theory and Learning: The Bottleneck and Distortion Approach Organizers: Thomas Gedeon Naftali Tishby
Phase Transitions in the Information Distortion NIPS 2003 workshop on Information Theory and Learning: The Bottleneck and Distortion Approach December.
Professor Walter W. Olson Department of Mechanical, Industrial and Manufacturing Engineering University of Toledo Solving ODE.
Frame by Frame Bit Allocation for Motion-Compensated Video Michael Ringenburg May 9, 2003.
Problem Solving Strategies EDU 412/413 Special Thanks to: Matthieu Petit.
Managerial Decision Making and Problem Solving
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
Loop Application: Numerical Methods, Part 1 The power of Matlab Mathematics + Coding.
Physics Fluctuomatics / Applied Stochastic Process (Tohoku University) 1 Physical Fluctuomatics Applied Stochastic Process 9th Belief propagation Kazuyuki.
1 LING 696B: Midterm review: parametric and non-parametric inductive inference.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
Exact and heuristics algorithms
The Information Bottleneck Method clusters the response space, Y, into a much smaller space, T. In order to informatively cluster the response space, the.
Name Iterative Source- and Channel Decoding Speaker: Inga Trusova Advisor: Joachim Hagenauer.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
BCS547 Neural Decoding.
Prototype Classification Methods Fu Chang Institute of Information Science Academia Sinica ext. 1819
Synaptic Dynamics: Unsupervised Learning
Arithmetic Test Pattern Generation: A Bit Level Formulation of the Optimization Problem S. Manich, L. García and J. Figueras.
John Lafferty Andrew McCallum Fernando Pereira
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Bayesian Brain: Probabilistic Approaches to Neural Coding Chapter 12: Optimal Control Theory Kenju Doya, Shin Ishii, Alexandre Pouget, and Rajesh P.N.Rao.
Deterministic Algorithms for Submodular Maximization Problems Moran Feldman The Open University of Israel Joint work with Niv Buchbinder.
By: Jesse Ehlert Dustin Wells Li Zhang Iterative Aggregation/Disaggregation(IAD)
Economics 2301 Lecture 37 Constrained Optimization.
DEPARTMENT/SEMESTER ME VII Sem COURSE NAME Operation Research Manav Rachna College of Engg.
Hongjie Zhu,Chao Zhang,Jianhua Lu Designing of Fountain Codes with Short Code-Length International Workshop on Signal Design and Its Applications in Communications,
Intro. ANN & Fuzzy Systems Lecture 38 Mixture of Experts Neural Network.
Ch. Eick: Num. Optimization with GAs Numerical Optimization General Framework: objective function f(x 1,...,x n ) to be minimized or maximized constraints:
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Information Bottleneck Method & Double Clustering + α Summarized by Byoung Hee, Kim.
Center for Computational Biology Department of Mathematical Sciences Montana State University Collaborators: Alexander Dimitrov John P. Miller Zane Aldworth.
Compressive Coded Aperture Video Reconstruction
LECTURE 11: Advanced Discriminant Analysis
Introduction to Information theory
2018/9/16 Distributed Source Coding Using Syndromes (DISCUS): Design and Construction S.Sandeep Pradhan, Kannan Ramchandran IEEE Transactions on Information.
Simple Kmeans Examples
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Presentation transcript:

A Bifurcation Theoretical Approach to the Solving the Neural Coding Problem June 28 Albert E. Parker Complex Biological Systems Department of Mathematical Sciences Center for Computational Biology Montana State University Collaborators: Tomas Gedeon, Alex Dimitrov, John Miller, and Zane Aldworth

 The Neural Coding Problem  A Clustering Problem  The Role of Bifurcation Theory  A new algorithm to solve the Neural Coding Problem Outline

The Neural Coding Problem GOAL: To understand the neural code. EASIER GOAL: We seek an answer to the question, How does neural activity represent information about environmental stimuli? “The little fly sitting in the fly’s brain trying to fly the fly”

stimulus X response Y Looking for the dictionary to the neural code … decoding encoding

… but the dictionary is not deterministic! Given a stimulus, an experimenter observes many different neural responses: X Y i | X i = 1, 2, 3, 4

… but the dictionary is not deterministic! Given a stimulus, an experimenter observes many different neural responses: Neural coding is stochastic!! X Y i | X i = 1, 2, 3, 4

Similarly, neural decoding is stochastic: Y X i |Y i = 1, 2, …, 9

Probability Framework X Y environmental stimuli neural responses decoder: P(X|Y) encoder: P(Y|X)

The Neural Coding Problem: How to determine the encoder P(Y|X) or the decoder P(X|Y)? Common Approaches: parametric estimations, linear methods Difficulty: There is never enough data. As we attempt search for an answer to the neural coding problem, we proceed in the spirit of John Tukey: It is better to be approximately right than exactly wrong.

One Approach: Cluster the responses X Y StimuliResponses YNYN q(Y N |Y) Clusters K objects {y i } N objects {y Ni }L objects {x i } p(X,Y)

One Approach: Cluster the responses To address the insufficient data problem, one clusters the outputs Y into clusters Y N so that the information that one can learn about X by observing Y N, I(X;Y N ), is as close as possible to the mutual information I(X;Y) X Y StimuliResponses YNYN q(Y N |Y) Clusters K objects {y i } N objects {y Ni }L objects {x i } p(X,Y)

Information Bottleneck Method (Tishby, Pereira, Bialek 1999) min I(Y,Y N ) constrained by I(X;Y N )  I 0 max –I(Y,Y N ) +  I(X;Y N ) Information Distortion Method (Dimitrov and Miller 2001) max H(Y N |Y) constrained by I(X;Y N )  I 0 max H(Y N |Y) +  I(X;Y N ) Two optimization problems which use this approach

An annealing algorithm to solve max q  (G(q)+  D(q)) Let q 0 be the maximizer of max q G(q), and let  0 =0. For k  0, let (q k,  k ) be a solution to max q G(q) +  D(q ). Iterate the following steps until  K =  max for some K. 1.Perform  -step: Let  k+1 =  k + d k where d k >0 2.The initial guess for q k+1 at  k+1 is q k+1 (0) = q k +  for some small perturbation . 3.Optimization: solve max q (G(q) +  k+1 D(q)) to get the maximizer q k+1, using initial guess q k+1 (0).

Application of the annealing method to the Information Distortion problem max q  (H(Y N |Y) +  I(X;Y N )) when p(X,Y) is defined by four gaussian blobs Stimuli Responses X Y 52 responses 52 stimuli p(X,Y) YYNYN q(Y N |Y) 52 responses4 clusters

Evolution of the optimal clustering: Observed Bifurcations for the Four Blob problem: We just saw the optimal clusterings q * at some  * =  max. What do the clusterings look like for  <  max ??

Application to cricket sensory data E(X|Y N ): stimulus means conditioned on each of the classes typical spike patterns optimal quantizer

We have used bifurcation theory in the presence of symmetries to totally describe how the the optimal clusterings of the responses must evolve…

Symmetries?? Y YNYN q(Y N |Y) : a clustering K objects {y i } N objects {y Ni } class 1 class 3

Y YNYN q(Y N |Y) : a clustering K objects {y i } N objects {y Ni } class 3 class 1 Symmetries??

Observed Bifurcation Structure

Observed Bifurcation Structure Group Structure

 q* Observed Bifurcation Structure

Continuation techniques provide numerical confirmation of the theory

Additional structure!!

Conclusions … We have a complete theoretical picture of how the clusterings of the responses evolve for any problem of the form max q  (G(q)+  D(q)) oWhen clustering to N classes, there are N-1 bifurcations. oIn general, there are only pitchfork and saddle-node bifurcations. oWe can determine whether pitchfork bifurcations are either subcritical or supercritical (1 st or 2 nd order phase transitions) oWe know the explicit bifurcating directions SO WHAT?? This yields a new and improved algorithm for solving the neural coding problem …

A numerical algorithm to solve max(G(q)+  D(q)) Let q 0 be the maximizer of max q G(q),  0 =1 and  s > 0. For k  0, let (q k,  k ) be a solution to max q G(q) +  D(q ). Iterate the following steps until  K =  max for some K. 1.Perform  -step: solve for and select  k+1 =  k + d k where d k = (  s sgn(cos  )) /(||   q k || 2 + ||   k || 2 +1) 1/2. 2.The initial guess for (q k+1, k+1 ) at  k+1 is (q k+1 (0), k+1 (0) ) = (q k, k ) + d k (   q k,   k ). 3.Optimization: solve max q (G(q) +  k+1 D(q)) using pseudoarclength continuation to get the maximizer q k+1, and the vector of Lagrange multipliers k+1 using initial guess (q k+1 (0), k+1 (0) ). 4.Check for bifurcation: compare the sign of the determinant of an identical block of each of  q [G(q k ) +  k D(q k )] and  q [G(q k+1 ) +  k+1 D(q k+1 )]. If a bifurcation is detected, then set q k+1 (0) = q k + d_k u where u is bifurcating direction and repeat step 3.