Pattern Separation and Completion in the Hippocampus

Slides:



Advertisements
Similar presentations
Multiuser Detection for CDMA Systems
Advertisements

Topic 12 – Further Topics in ANOVA
Lecture 13: Associative Memory References: D Amit, N Brunel, Cerebral Cortex 7, (1997) N Brunel, Network 11, (2000) N Brunel, Cerebral.
Neural Network of the Cerebellum: Temporal Discrimination and the Timing of Responses Michael D. Mauk Dean V. Buonomano.
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Mode shifting between storage and recall based on novelty detection in oscillating hippocampal circuits M. Meeter J. M. J. Murre L. M. Talamini Date of.
September 14, 2010Neural Networks Lecture 3: Models of Neurons and Neural Networks 1 Visual Illusions demonstrate how we perceive an “interpreted version”
Introduction to Statistical Inferences
Chapter 5 Sampling Distributions
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
RMTD 404 Lecture 8. 2 Power Recall what you learned about statistical errors in Chapter 4: Type I Error: Finding a difference when there is no true difference.
Neural Networks Architecture Baktash Babadi IPM, SCS Fall 2004.
Lecture 10: Mean Field theory with fluctuations and correlations Reference: A Lerchner et al, Response Variability in Balanced Cortical Networks, q-bio.NC/ ,
Artificial Intelligence Lecture No. 29 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators.
CSC321: Introduction to Neural Networks and machine Learning Lecture 16: Hopfield nets and simulated annealing Geoffrey Hinton.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
1 Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
Date of download: 6/3/2016 Copyright © 2016 American Medical Association. All rights reserved. From: Reduced Parahippocampal Connectivity Produces Schizophrenia-like.
A Role for Hilar Cells in Pattern Separation in the Dentate Gyrus: A Computational Approach Journal Club 5/16/12.
INTRODUCTION TO ECONOMIC STATISTICS Topic 5 Discrete Random Variables These slides are copyright © 2010 by Tavis Barr. This work is licensed under a Creative.
Cholinergic Modulation of the Hippocampus Computational Models of Neural Systems Lecture 2.5 David S. Touretzky September, 2007.
Complementary Learning Systems
Neural Network Architecture Session 2
Deep Feedforward Networks
What is the role of the Dentate Gyrus? The role of DG CA1 Sb CA3 DG
CSC321 Lecture 18: Hopfield nets and simulated annealing
Theta, Gamma, and Working Memory
Neural Networks.
Real Neurons Cell structures Cell body Dendrites Axon
Sampling Distributions
Modeling and Simulation CS 313
Chapter 5 Sampling Distributions
Chapter 5 Sampling Distributions
Elementary Statistics
Lecture Slides Elementary Statistics Thirteenth Edition
Statistics Review ChE 477 Winter 2018 Dr. Harding.
Hidden Markov Models Part 2: Algorithms
Models of Network Formation
Lesson Comparing Two Means.
Chapter 5 Sampling Distributions
Models of Network Formation
Models of Network Formation
Optimal Degrees of Synaptic Connectivity
Artificial Intelligence Lecture No. 28
Volume 74, Issue 3, Pages (May 2012)
The Naïve Bayes (NB) Classifier
Attentional Modulations Related to Spatial Gating but Not to Allocation of Limited Resources in Primate V1  Yuzhi Chen, Eyal Seidemann  Neuron  Volume.
Linking Memories across Time via Neuronal and Dendritic Overlaps in Model Neurons with Active Dendrites  George Kastellakis, Alcino J. Silva, Panayiota.
Inferential Statistics
Psych 231: Research Methods in Psychology
Psych 231: Research Methods in Psychology
Associative memory models: from the cell-assembly theory to biophysically detailed cortex simulations  Anders Lansner  Trends in Neurosciences  Volume.
Psych 231: Research Methods in Psychology
Patrick Kaifosh, Attila Losonczy  Neuron 
Cian O’Donnell, Terrence J. Sejnowski  Neuron 
Adult Neurogenesis in the Hippocampus: From Stem Cells to Behavior
Volume 96, Issue 1, Pages e7 (September 2017)
Relating Hippocampal Circuitry to Function
Volume 25, Issue 23, Pages R1116-R1121 (December 2015)
Introduction to Neural Network
David Kauchak CS158 – Spring 2019
Simulation Berlin Chen
SISSA limbo CÉREBRO, VIDA e CULTURA We can easily describe the brain;
Patrick Kaifosh, Attila Losonczy  Neuron 
Volume 27, Issue 2, Pages (January 2017)
Presentation transcript:

Pattern Separation and Completion in the Hippocampus Computational Models of Neural Systems Lecture 3.5 David S. Touretzky October, 2013

Computational Models of Neural Systems Overview Pattern separation Pulling similar patterns apart reduces memory interference. Pattern Completion Noisy or incomplete patterns should be mapped to more complete or correct versions. How can both functions be accomplished in the same architecture? Use conjunction (codon units; DG) for pattern separation. Learned weights plus thresholding gives pattern completion. Recurrent connections (CA3) can help with completion, but aren't used in the model described here. 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Information Flow perforant path Cortex mossy fibers Cortical projections from many areas form an EC representation of an event. EC layer II projects to CA3 (both directly and via DG), forming a new representation better suited to storage and retrieval. EC layer III projects to CA1, forming an invertible representation that can reconstitute the EC pattern. Learning occurs in all these connections. 5/23/2018 Computational Models of Neural Systems

Features of Hippocampal Organization Local inhibitory interneurons in each region. May regulate overall activity levels, as in a kWTA network. CA3 and CA1 have less activity than EC and subiculum. DG has less activity than CA3/CA1. Less activity means representation is more sparse, hence can be more highly orthogonal. 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Connections EC layer II (perf. path) projects diffusely to DG and CA3. Each DG granule cell receives 5,000 inputs from EC. Each CA3 pyramidal cell receives 3750-4500 inputs from EC. This is about 2% of the rat's 200,000 EC layer II neurons. DG has roughly 1 million granule cells. CA3 has 160,000 pyramidal cells; CA1 has 250,000. DG to CA3 projection (mossy fibers) is sparse and topographic. CA3 cells receive 52-87 mossy fiber synapses. NMDA-dependent LTP has been demonstrated in perforant path and Schafer collaterals. LTP also demonstrated in mossy fiber pathway (non-NMDA). LTD may also be present in these pathways. 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Model Parameters O'Reilly & McClelland investigated several models, starting with a simple two-layer k-WTA model (like Marr). Ni,No = # units in the layer ki,ko = # active inputs in one pattern ai,ao = fractional activity in the layer; ao = ko/No F = fan-in of units in the output layer (must be < Ni) Ha = # of hits for pattern A EC(i) CA3(o) Fan-in F = 9 Hits Ha = 4 5/23/2018 Computational Models of Neural Systems

Measuring the Hits a Unit Receives How many input patterns? What is the expected number of hits Ha for an output unit? What is the distribution of hits, P(Ha) ? Hypergeomtric (not binomial; we're drawing without replacement)  𝑁 𝑖 𝑘 𝑖  𝐻 𝑎 = 𝑘 𝑖 𝑁 𝑖 𝐹= α 𝑖 𝐹 5/23/2018 Computational Models of Neural Systems

Hypergeometric Distribution What is the probability of getting exactly Ha hits from an input pattern with ki active units, given that the fan-in is F and the total input size is Ni? C(ki, Ha) ways of choosing active units to be hits C(Ni-ki, F-Ha) ways of choosing inactive units for the remaining ones sampled by the fan-in C(Ni, F) ways of sampling F inputs from a population of size Ni 𝑃 𝐻 𝑎 ∣ 𝑘 𝑖 , 𝑁 𝑖 ,𝐹 =  𝑘 𝑖 𝐻 𝑎   𝑁 𝑖 − 𝑘 𝑖 𝐹− 𝐻 𝑎   𝑁 𝑖 𝐹  # of ways to wire an output cell with Ha hits # of ways to wire an output cell 5/23/2018 Computational Models of Neural Systems

Determining the kWTA Threshold Assume we want the output layer to have an expected activity level of ao. Must set the threshold for output units to select the tail of the hit distribution. Call this Hat. Use the summation to choose Hat to produce the desired value of ao. α 𝑜 = 𝐻 𝑎 = 𝐻 𝑎 𝑡 min 𝑘 𝑖 ,𝐹 𝑃 𝐻 𝑎 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Pattern Overlap In order to measure pattern separation properties of the two-layer model, consider two patterns A and B. Measure the input overlap Wi = number of units in common. Compute the expected output overlap Wo as a function of Wi. If Wo < Wi the model is doing pattern separation. To calculate output overlap we need to know Hab, the number of hits an output unit receives for pattern B given that it is already known to be part of the representation for pattern A. 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Distribution of Hab For small input overlap, the patterns are virtually independent, and Hab is distributed like Ha. As input overlap increases, Hab moves rightward (more hits expected), and narrows: output overlap increases. But the relationship is nonlinear. 5/23/2018 Computational Models of Neural Systems

Visualizing the Overlap a) Hits from pattern A. b) Hab = overlap of A&B hits 5/23/2018 Computational Models of Neural Systems

Prob. of b Hits And Specific Values for Ha, Hab, Hab Given Overlap Wi 𝑃 𝑏 𝐻 𝑎 , Ω 𝑖 , 𝐻 𝑎𝑏 , 𝐻 𝑎 ˉ 𝑏 =  𝐻 𝑎 𝐻 𝑎𝑏   𝐹− 𝐻 𝑎 𝐻 𝑎 ˉ 𝑏   𝑘 𝑖 − 𝐻 𝑎 Ω 𝑖 − 𝐻 𝑎𝑏   𝑁 𝑖 − 𝑘 𝑖 −𝐹+ 𝐻 𝑎 𝑘 𝑖 − Ω 𝑖 − 𝐻 𝑎 ˉ 𝑏   𝑘 𝑖 Ω 𝑖   𝑁 𝑖 − 𝑘 𝑖 𝑘 𝑖 − Ω 𝑖  1 2 3 4 # of ways of achieving Hb hits given overlap Wi # of ways of achieving ki – Hb non-hits given overlap Wi 3 # of ways of achieving overlap Wi 4 1,3 2,4 2 Note: 𝐻 𝑏 = 𝐻 𝑎𝑏 + 𝐻 𝑎 ˉ 𝑏 1 To calculate𝑃 𝐻 𝑏 we must sum 𝑃 𝑏 over all combinations of 𝐻 𝑎 , 𝐻 𝑎𝑏 , 𝐻 𝑎 ˉ 𝑏 5/23/2018 Computational Models of Neural Systems

Estimating Overlap for Rat Hippocampus We can use the formula for Pb to calculate expected output overlap as a function of input overlap. To do this for rodent hippocampus, O'Reilly & McClelland chose numbers close to the biology but tailored to avoid round-off problems in the overlap formula. 5/23/2018 Computational Models of Neural Systems

Estimated Pattern Separation in CA3 5/23/2018 Computational Models of Neural Systems

Sparsity Increases Pattern Separation Pattern separation performance of a generic network with activity levels comparable to EC, CA3, or DG. Sparse patterns yield greater separation. 5/23/2018 Computational Models of Neural Systems

Fan-In Size Has Little Effect 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Adding Input from DG DG makes far fewer connections (64 vs. 4003), but they may have higher strength. Let M = mossy fiber strength. Separation in DG better than in CA3 w/o DG. DG connections help for M ≥ 15. With M=50, DG projection alone is as good as DG+EC. 5/23/2018 Computational Models of Neural Systems

Combining Two Distributions CA3 has far fewer inputs from DG than from EC. But the DG input has greater variance in hit distribution. When combining two equally-weighted distributions, the one with the greater variance has the most effect on the tail. For 0.25 input overlap: DG hit distribution has std. dev. of 0.76 EC hit distribution has std. dev. of 15. Setting M=20 would balance the effects of the two projections. In the preceding plot, the M=20 line appears in between the M=0 line (EC only) and the “M only” line. 5/23/2018 Computational Models of Neural Systems

Without Learning, Partial Inputs Are Separated, Not Completed Less separation between A and subset(A) than between patterns A and B, because there are no noise inputs. But Wo is still less than Wi 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Pattern Completion Without learning, completion cannot happen. Two learning rules were tried: WI: Weight Increase (like Marr) WID: Weight Increase/Decrease WI learning multiplies weights in Hab by (1+Lrate). WID learning increases weights as per WI, but also exponentially decreases weights to units in F-Ha by multiplying by (1-Lrate). Result: WID learning improves both separation and completion. 5/23/2018 Computational Models of Neural Systems

WI Learning and Pattern Completion 5/23/2018 Computational Models of Neural Systems

WI Learning Reduces Pattern Separation 5/23/2018 Computational Models of Neural Systems

WI Learning Hurts Completion Percent of possible improvement No learning (learning rate = 0) Learning rate = 0.1 5/23/2018 Computational Models of Neural Systems

WID Learning Has A Good Tradeoff 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems WI vs. WID Learning Sweet spot 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Hybrid Systems Multiple completion stages don't help (cf. Willshaw & Buckingham's comparison of Marr models.) With noisy cues, completion produces a somewhat noisy result which would lead to further separation at the next stage. MSEPO — mossy fibers only for separation (learning). Perhaps partial EC inputs aren't strong enough to drive DG. FM —fixed mossy system: no learning on these fibers. Learning reduces pattern separation. Real mossy fibers undergo LTP, but it's not NMDA-dependent (so non-Hebbian). FMSEPO — combination of FM + SEPO. Optimal tradeoff between separation and completion. 5/23/2018 Computational Models of Neural Systems

Performance of Hybrid Models 5/23/2018 Computational Models of Neural Systems

What Is the Mossy Fiber Pathway Doing? Adds a high variance signal to the CA3 input, which... Selects a random subset of CA3 cells that are already highly activated by EC input. This enhances separation when recruiting the representations of stored patterns. But it hurts retrieval with partial or noisy cues. So don't use it. Use MSEPO or FMSEPO. 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Conclusions The main contribution of this work is to show how separation and completion can be accomplished in the same architecture. The model uses realistic figures for numbers of units and connections. Fan-in size doesn't seem to matter. WID learning is necessary for a satisfactory tradeoff between separation and completion. DG contributes to separation but perhaps not to completion. 5/23/2018 Computational Models of Neural Systems

Limitations of the Model Simplified anatomy: the model only included ECCA3 and ECDGCA3 connections. No CA3 recurrent connections. No CA1. Only a single pattern stored at a time: Store A, measure overlap with B. No attempt to measure memory capacity. A more realistic model would be too hard to analyze. 5/23/2018 Computational Models of Neural Systems

Possible Different Functions of CA3 and CA1 Measured by IEG (Immediate Early Genes) 5/23/2018 Computational Models of Neural Systems

Hasselmo's Model: Novelty Detection CA1 CA3 Medial Septum ACh fimbria/fornix pp Sch Acetycholine reduces synaptic efficacy (prevent CA3 from altering CA1 pattern) and enhances synaptic plasticity. 5/23/2018 Computational Models of Neural Systems

Pattern Separation in Human Hippocampus Bakker et al., Science, March 2008: fMRI study Subjects were shown 144 pairs of images that differed slightly, plus additional foils. Asked for an unrelated judgment about each image (indoor vs. outdoor object). Three types of trials: (i) new object, (ii) repetition of a previously seen object, (iii) slightly different version of a previously seen object: a lure. 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Eight ROIs Found Couldn't resolve DG vs. CA3 so treated as one region. Regions outlined above: CA3/DG CA1 Subiculum Areas of significant activity within MTL shown in white. New objects, repetitions, and lures were reliably discriminable. Generally, repetitions  lower activity. 5/23/2018 Computational Models of Neural Systems

Computational Models of Neural Systems Bias Scores for ROIs bias = (first – lure) / (first – repetition) Scores close to 1  completion; 0  separation. CA3/DG shows more pattern separation than other areas. 5/23/2018 Computational Models of Neural Systems