Connectionist Units, Probabilistic and Biologically Inspired

Slides:



Advertisements
Similar presentations
Representations for KBS: Uncertainty & Decision Support
Advertisements

PDP: Motivation, basic approach. Cognitive psychology or “How the Mind Works”
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Brain Mechanisms of Unconscious Inference J. McClelland Symsys 100 April 22, 2010.
COGNITIVE NEUROSCIENCE
How does the mind process all the information it receives?
Perceptual Inference and Information Integration in Brain and Behavior PDP Class Jan 11, 2010.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Connectionism. ASSOCIATIONISM Associationism David Hume ( ) was one of the first philosophers to develop a detailed theory of mental processes.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Naive Bayes Classifier
The Interactive Activation Model. Ubiquity of the Constraint Satisfaction Problem In sentence processing –I saw the grand canyon flying to New York –I.
Brain Mechanisms of Unconscious Inference J. McClelland Symsys 100 April 7, 2011.
1 Chapter 1: Introduction to Design of Experiments 1.1 Review of Basic Statistical Concepts (Optional) 1.2 Introduction to Experimental Design 1.3 Completely.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
Principled Probabilistic Inference and Interactive Activation Psych209 January 25, 2013.
Computer vision: models, learning and inference Chapter 2 Introduction to probability.
1 Probability and Statistics Confidence Intervals.
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
REASONING UNDER UNCERTAINTY: CERTAINTY THEORY
Some Terminology experiment vs. correlational study IV vs. DV descriptive vs. inferential statistics sample vs. population statistic vs. parameter H 0.
MLPR - Questions. Can you go through integration, differentiation etc. Why do we need priors? Difference between prior and posterior. What does Bayesian.
Recuperação de Informação B Modern Information Retrieval Cap. 2: Modeling Section 2.8 : Alternative Probabilistic Models September 20, 1999.
Sampling and Sampling Distributions
Virtual University of Pakistan
Tests About a Population Proportion
Bayesian inference in neural networks
DEEP LEARNING BOOK CHAPTER to CHAPTER 6
Artificial Neural Networks
Network States as Perceptual Inferences
Perception, interaction, and optimality
Inference for Proportions
Naive Bayes Classifier
Chapter 4: Studying Behavior
Warm-up (10 min.) I. Factor the following expressions completely over the real numbers. 3x3 – 15x2 + 18x x4 + x2 – 20 II. Solve algebraically and graphically.
V5 Stochastic Processes
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
A Simple Artificial Neuron
Chapter 9: Testing a Claim
Chapter 8: Inference for Proportions
Neural Networks A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Bayesian inference in neural networks
Network States as Perceptual Inferences
Simple learning in connectionist networks
Emergence of Semantics from Experience
Chapter 9 Hypothesis Testing.
Hidden Markov Models Part 2: Algorithms
Chapter 9 Hypothesis Testing.
More about Posterior Distributions
King Fahd University of Petroleum and Minerals
Emergent Functions of Simple Systems
OVERVIEW OF BIOLOGICAL NEURONS
Professor Marie desJardins,
Graph Stories 2 Teacher Notes
Neuro-RAM Unit in Spiking Neural Networks with Applications
Adaptive Resonance Theory
Artificial Intelligence Lecture No. 28
Some Basic Aspects of Perceptual Inference Under Uncertainty
Class #21 – Monday, November 10
Simple Linear Regression
The Naïve Bayes (NB) Classifier
Graded Constraint Satisfaction, the IA Model, and Network States as Perceptual Inferences Psychology 209 January 15, 2019.
Parametric Methods Berlin Chen, 2005 References:
Adaptive Resonance Theory
Bayesian vision Nisheeth 14th February 2019.
28th September 2005 Dr Bogdan L. Vrusias
Simple learning in connectionist networks
Recuperação de Informação B
Unit 5: Hypothesis Testing
Presentation transcript:

Connectionist Units, Probabilistic and Biologically Inspired Psych 209 January 11, 2013

Neurons and Units At rest, neurons are negatively polarized compared to the surrounding medium. Excitatory inputs cause them to become depolarized Inhibitory inputs cause them to become hyperpolarized. A neuron’s tendency to emit action potentials increases as it becomes less and less polarized, and correspondingly decreases as it becomes more an more polarized. Firing rate levels off at 0 at the bottom and ~100 spikes per second at the top. In PDP models our units correspond to notional populations of neurons (maybe 10,000 per unit). The continuous valued output of a unit, which usually ranges from 0 to 1, can be thought of as approximating the proportion of neurons in the population emitting an action potential in a small time interval, divided by the maximum value this proportion could take. For example, if we choose the time interval to be one msec, the maximum proportion of neurons firing per msec would be .1. So an output of .2 would correspond to .02 of the 10,000 neurons firing (or 200 of the 10,000 neurons) firing per millisecond.

Neuro-similitude or implementation of a computational theory? Early models (Grossberg’s, the iac model you will explore in the first homework) embodied attempts to capture characteristics of real neurons in a simplified way. Simulations were used to show how a process could be modeled. Later models are more grounded in theory Some of the key features of the physiology are still captured And the formulation allows us to relate what neurons are doing to a theory of what they should be doing. For many purposes, the two formulations work very similarly, even if one is not exactly valid from a probabilistic inference perspective.

Probabilistic Formulation: Review and Discussion of Probabilistic Concepts Two different sources of evidence e1 and e2 are conditionally independent given hi, iff p(e1&e2|hi) = p(e1|hi)p(e2|hi) If this holds for all i we can say the different elements of evidence are ‘conditionally independent’ In case of N sources of evidence, all conditionally independent under h, then we get: p(e|hi) = Pj p(ej|hi) Combining this with the prior we get a quantity I call the Support for hypothesis i: Si = p(hi) Pj p(ej|hi) Taking logs we get: log(Si) = log(p(hi)) + Sj log(p(ej|hi)) In the case of a single h and its alternative ~h we get: log(S/(1-S)) = log(p(h)/(1-p(h)) + Sj log(p(ej|h)|p(ej|~h))

How this relates to connectionist units (or populations of neurons) The baseline activation of the unit is thought to depend on a constant background input called its ‘bias’. When other units are active, their influences are combined with the bias to yield a quantity called the ‘net input’. The influence of a unit j on another unit i depends on the activation of j and the weight or strength of the connection to i from j. Connection weights can be positive (excitatory) or negative (inhibitory). These influences are summed to determine the net input to unit i: neti = biasi + Sjajwij where aj is the activation of unit j, and wij is the strength of the connection to unit i from unit j. Input from unit j wij unit i

A Unit’s Activation can Reflect P(h|e) The activation of unit i given its net input neti is assumed to be given by: ai = exp(neti) 1 + exp(neti) This function is called the ‘logistic function’. It is usually written in the numerically identical form: ai = 1/[1 + exp(-neti)] In the reading we showed that ai = p(hi|e) iff aj = 1 when ej is present, or 0 when ej is absent wij = log(p(ej|hi)/p(ej|~hi)) biasi = log(p(hi)/p(~hi)) This assumes the evidence is conditionally independent given the state of h. ai neti

Posterior probability when there are N alternatives In this case, the probability of a particular hypothesis given the evidence becomes: P(hi|e) = p(e|hi)p(hi) Pi’p(e|hi’)p(hi’) The normalization implied here can be performed by using a ‘net input’ as before but now setting each unit’s activation according to: ai = exp(neti) Si’exp(neti’) In this case, ai = p(hi|e) iff aj = 1 when ej is present, or 0 when ej is absent wij = log(p(ej|hi)) biasi = log(p(hi)) h e

The Grossberg Unit (Biologically Inspired) max = 1 min = -.2 rest a Unit i Output from unit j wij ei = Sj+outputj wij ; ii = Sj-outputj wij outputi = [ai]+ Dai = (max –ai)ei – (ai-min)ii – decay(ai-rest)

The IAC Unit Unit i Output from unit j wij max = 1 min = -.2 rest a

Suppose the net input to a unit is constant (and positive) Suppose the net input to a unit is constant (and positive). What is its equilibrium activation value?

How Competition Works

Jets & Sharks Model (IAC) Allows you to explore a simple localist / hard-wired PDP type model that has been applied to many problems in perception, cognition and social psychology. In the Jets and Sharks case, we will explore: Information retrieval by name or attributes Assignment of plausible default values Spontaneous generalization