NEURAL NETWORK THEORY NEURONAL DYNAMICS Ⅰ : ACTIVATIONS AND SIGNALS 2004.10.13.

Slides:



Advertisements
Similar presentations
Neural Network Toolbox COMM2M Harry R. Erwin, PhD University of Sunderland.
Advertisements

The linear/nonlinear model s*f 1. The spike-triggered average.
Slides from: Doug Gray, David Poole
Introduction to Neural Networks Computing
Chapter 3 Image Enhancement in the Spatial Domain.
Self Organization: Competitive Learning
Artificial Neural Networks - Introduction -
Artificial Neural Networks - Introduction -
Neural Networks.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Introduction to Systems What are signals and what are systems The system description Classification of systems Deriving the system model – Continuous systems.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
How does the mind process all the information it receives?
Artificial Neurons, Neural Networks and Architectures
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Review of Probability and Random Processes
System State and Differential Equations Linear Systems and Signals Lecture 3 Spring 2008.
Artificial Neural Network
Neural Networks Lecture 8: Two simple learning algorithms
Autumn 2008 EEE8013 Revision lecture 1 Ordinary Differential Equations.
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Jochen Triesch, UC San Diego, 1 Short-term and Long-term Memory Motivation: very simple circuits can store patterns of.
Artificial Neural Networks
Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University EE459 Neural Networks The Structure.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
CS344 : Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 29 Introducing Neural Nets.
NEURONAL DYNAMICS 2: ACTIVATION MODELS
Feedback Control Systems (FCS) Dr. Imtiaz Hussain URL :
CS-424 Gregory Dudek Today’s Lecture Neural networks –Training Backpropagation of error (backprop) –Example –Radial basis functions.
CS 478 – Tools for Machine Learning and Data Mining Perceptron.
Artificial Neural Networks Students: Albu Alexandru Deaconescu Ionu.
Synaptic Dynamics: Unsupervised Learning
Lecture 5 Neural Control
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
CSCE 441: Keyframe Animation/Smooth Curves (Cont.) Jinxiang Chai.
Architecture and Equilibria 结构和平衡 Chapter 6 神经网络与模糊系统 学生: 李 琦 导师:高新波.
Back-propagation network (BPN) Student : Dah-Sheng Lee Professor: Hahn-Ming Lee Date:20 September 2003.
Signals and Systems Analysis NET 351 Instructor: Dr. Amer El-Khairy د. عامر الخيري.
6.4 Random Fields on Graphs 6.5 Random Fields Models In “Adaptive Cooperative Systems” Summarized by Ho-Sik Seok.
NEURAL NETWORK THEORY NEURONAL DYNAMICS Ⅰ : ACTIVATIONS AND SIGNALS 欢迎大家提出意见建议!
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
From LIF to HH Equivalent circuit for passive membrane The Hodgkin-Huxley model for active membrane Analysis of excitability and refractoriness using the.
1 Review of Probability and Random Processes. 2 Importance of Random Processes Random variables and processes talk about quantities and signals which.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
CSC321: Neural Networks Lecture 1: What are neural networks? Geoffrey Hinton
1 Azhari, Dr Computer Science UGM. Human brain is a densely interconnected network of approximately neurons, each connected to, on average, 10 4.
Intro. ANN & Fuzzy Systems Lecture 3 Basic Definitions of ANN.
CS621: Artificial Intelligence Lecture 10: Perceptrons introduction Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Today’s Lecture Neural networks Training
Chapter 7. Classification and Prediction
Artificial Intelligence (CS 370D)
第 3 章 神经网络.
Lecture 22 Definite Integrals.
CS344: Introduction to Artificial Intelligence (associated lab: CS386)
NEURONAL DYNAMICS 2: ACTIVATION MODELS
Synaptic Dynamics: Unsupervised Learning
Artificial Intelligence Chapter 3 Neural Networks
Perceptron as one Type of Linear Discriminants
1.3 Vector Equations.
Artificial Intelligence Lecture No. 28
Artificial Intelligence Chapter 3 Neural Networks
Recursively Adapted Radial Basis Function Networks and its Relationship to Resource Allocating Networks and Online Kernel Learning Weifeng Liu, Puskal.
Artificial Intelligence Chapter 3 Neural Networks
NEURAL DYNAMIC1: ACTIVATIONHS AND SIGNALS
Artificial Intelligence Chapter 3 Neural Networks
Introduction to Neural Network
Artificial Intelligence Chapter 3 Neural Networks
Presentation transcript:

NEURAL NETWORK THEORY NEURONAL DYNAMICS Ⅰ : ACTIVATIONS AND SIGNALS

ACTIVATIONS AND SIGNALS NEURONS AS FUNCTIONS SIGNAL MONOTONICITY BIOLOGICAL ACTIVATIONS AND SIGNALS NEURON FIELDS NEURONAL DYNAMICAL SYSTEMS COMMON SIGNAL FUNCTION PULSE-CODED SIGNAL FUNCTION

NEURONS AS FUNCTIONS Neurons behave as functions. Neurons transduce an unbounded input activation x(t) at time t into a bounded output signal S(x(t)).

NEURONS AS FUNCTIONS S(x) x 0 -∞-++∞ Fig.1 s(x) is a bounded monotone-nondecreasing function of x If c → +∞ , we get threshold signal function (dash line), Which is piecewise differentiable

NEURONS AS FUNCTIONS The transduction description: a sigmoidal or S-shaped curve the logistic signal function: The logistic signal function is sigmoidal and strictly increases for positive scaling constant c >0.

NEURONS AS FUNCTIONS S(x) x 0 -∞-++∞ Fig.1 s(x) is a bounded monotone-nondecreasing function of x If c → +∞ , we get threshold signal function (dash line), Which is piecewise differentiable

NEURONS AS FUNCTIONS S would transduce the four-neuron vector of activations ( – 689) to the four- dimensional bit vector of signal ( ) Zero activations to unity,zero,or the previous signal

SIGNAL MONOTONICITY In general, signal functions are monotone nondecreasing S’>=0. This means signal functions have an upper bound or saturation value. S(x) x 0 -∞-++∞

SIGNAL MONOTONICITY An important exception: bell-shaped signal function or Gaussian signal functions The sign of the signal-activation derivation s’ is opposite the sign of the activation x. We shall assume signal functions are monotone nondecreasing unless stated otherwise.

SIGNAL MONOTONICITY Generalized Gaussian signal function define potential or radial basis function : input activation vector: variance: mean vector: we shall consider only scalar-input signal functions:

SIGNAL MONOTONICITY A property of signal monotonicity: semi-linearity Comparation: a. Linear signal functions: computation and analysis is comparatively easy; do not suppress noise. b. Nonlinear signal functions: increases a network ’ s computational richness and facilitates noise suppression; risks computational and analytical intractability;

SIGNAL MONOTONICITY Signal and activation velocities the signal velocity: =dS/dt Signal velocities depend explicitly on action velocities

BIOLOGICAL ACTIVATIONS AND SIGNALS Fig.2 Neuron anatomy 神经元 (Neuron) 是由细胞核 (cell nucleus) ,细胞体 (soma) ,轴 突 (axon) ,树突 (dendrites) 和突触 (synapse) 所构成的

BIOLOGICAL ACTIVATIONS AND SIGNALS Dendrites Synapse Axon

BIOLOGICAL ACTIVATIONS AND SIGNALS X= ( x1 , x2 , … , xn ) W= ( w1 , w2 , … , wn ) net=∑xiwi net=XW x 2 w 2 ∑f o=f ( net ) x n w n … net=XW x 1 w 1

BIOLOGICAL ACTIVATIONS AND SIGNALS Competitive Neuronal Signal logical signal function ( Binary  Bipolar ) The neuron “ wins ” at time t if, “ loses ” if and otherwise possesses a fuzzy win-loss status between 0 an 1. a. Binary signal functions : [0,1] b. Bipolar signal functions : [-1,1] McCulloch—Pitts (M—P) neurons

NEURON FIELDS Neurons within a field are topologically ordered, often by proximity. In the simplest case, neuron are not topologically. (zeroth-order topology)

NEURON FIELDS Denotation:,, neural system samples the function m times to generate the associated pairs,..., The overall neural network behaves as an adaptive filter and sample data changed network parameters.

NEURON FIELDS

NEURONAL DYNAMICAL SYSTEMS Description:a system of first-order differential or difference equations that govern the time evolution of the neuronal activations or membrane potentials Activation differential equations: in vector notation: (1) (2) (3) (4)

NEURONAL DYNAMICAL SYSTEMS Autonomous Nonautonomous automomous systems are usually easier to analyze than nonautomous systems.

NEURONAL DYNAMICAL SYSTEMS Neuronal State spaces So the state space of the entire neuronal dynamical system is: Augmentation:

NEURONAL DYNAMICAL SYSTEMS Signal state spaces as hypercubes The signal state of field at time t: The signal state space: an n-dimensional hypercube The unit hypercube : or, The relationship between hyper-cubes and the fuzzy set :, subsets of correspond to the vertices of

NEURONAL DYNAMICAL SYSTEMS Neuronal activations as short-term memory Short-term memory(STM) : activation Long-term memory(LTM) : synapse

COMMON SIGNAL FUNCTION 1 、 Liner Function S(x) = cx + k, c>0 x S o k

COMMON SIGNAL FUNCTION 2. Ramp Function r if x≥θ S(x)= cx if |x|<θ -rif x≤-θ r>0, r is a constant. r -r-r θ -θ -θ x S

COMMON SIGNAL FUNCTION 3 、 threshold linear signal function: a special Ramp Function Another form:

COMMON SIGNAL FUNCTION 4 、 logistic signal function: Where c>0. So the logistic signal function is monotone increasing.

COMMON SIGNAL FUNCTION 5 、 threshold signal function: Where T is an arbitrary real-valued threshold,and k indicates the discrete time step.

COMMON SIGNAL FUNCTION 6 、 hyperbolic-tangent signal function:

COMMON SIGNAL FUNCTION 7 、 threshold exponential signal function: When,

COMMON SIGNAL FUNCTION 8 、 exponential-distribution signal function: When,

COMMON SIGNAL FUNCTION 9 、 the family of ratio-polynomial signal function: An example For,

PULSE-CODED SIGNAL FUNCTION Definition: where

PULSE-CODED SIGNAL FUNCTION Pulse-coded signals take values in the unit interval [0,1]. Proof: when

PULSE-CODED SIGNAL FUNCTION Velocity-difference property of pulse-coded signals The first-order linear inhomogenous differential equation: The solution to this differential equation: A simple form for the signal velocity:

PULSE-CODED SIGNAL FUNCTION The central result of pulse-coded signal functions: The instantaneous signal-velocity equals the current pulse minus the current expected pulse frequency the velocity-difference property of pulse- coded signal functions

End Thank you ! Li Wei