The McCullough-Pitts Neuron

Slides:



Advertisements
Similar presentations
Slides from: Doug Gray, David Poole
Advertisements

Introduction to Neural Networks Computing
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Artificial Neural Networks
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Classification Neural Networks 1
Simple Neural Nets For Pattern Classification
Connectionist Models: Backprop Jerome Feldman CS182/CogSci110/Ling109 Spring 2008.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
Artificial Neural Networks
Data Mining with Neural Networks (HK: Chapter 7.5)
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
DIGITAL IMAGE PROCESSING Dr J. Shanbehzadeh M. Hosseinajad ( J.Shanbehzadeh M. Hosseinajad)
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Non-Bayes classifiers. Linear discriminants, neural networks.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Chapter 2 Single Layer Feedforward Networks
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Machine Learning Supervised Learning Classification and Regression
Neural networks.
Fall 2004 Backpropagation CS478 - Machine Learning.
Introduction to Neural Networks
Neural Networks.
Artificial Neural Networks
Learning with Perceptrons and Neural Networks
Chapter 2 Single Layer Feedforward Networks
Artificial neural networks:
第 3 章 神经网络.
Real Neurons Cell structures Cell body Dendrites Axon
DEPARTMENT: COMPUTER SC. & ENGG. SEMESTER : VII
slides created by Leon Barrett
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Dr. Kenneth Stanley September 6, 2006
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Artificial Neural Networks
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
CSE P573 Applications of Artificial Intelligence Neural Networks
Simple learning in connectionist networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
CS621: Artificial Intelligence
CSC 578 Neural Networks and Deep Learning
Data Mining with Neural Networks (HK: Chapter 7.5)
ECE 471/571 - Lecture 17 Back Propagation.
Simple Learning: Hebbian Learning and the Delta Rule
Classification Neural Networks 1
Chapter 3. Artificial Neural Networks - Introduction -
Neuro-Computing Lecture 4 Radial Basis Function Network
Perceptron as one Type of Linear Discriminants
CSE 573 Introduction to Artificial Intelligence Neural Networks
Neural Networks Chapter 5
Neural Network - 2 Mayank Vatsa
Lecture Notes for Chapter 4 Artificial Neural Networks
Backpropagation.
McCulloch–Pitts Neuronal Model :
Artificial Intelligence 12. Two Layer ANNs
Chapter - 3 Single Layer Percetron
Simple learning in connectionist networks
Introduction to Neural Network
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
Active, dynamic, interactive, system
David Kauchak CS158 – Spring 2019
PYTHON Deep Learning Prof. Muhammad Saeed.
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

The McCullough-Pitts Neuron yj wij xi f yi ti : target xi = ∑j wij yj yi = f(xi – qi) Threshold yj: output from unit j wij: weight on connection from j to i (can be positive (excitatory or negative (inhibitory) xi: weighted sum of input to unit i

Simple Threshold Linear Unit (TLU)

Computing with Abstract Neurons McCollough-Pitts Neurons were initially used to model pattern classification size = small AND shape = round AND color = green AND location = on_tree => unripe linking classified patterns to behavior size = large OR motion = approaching => move_away size = small AND direction = above => move_above McCollough-Pitts Neurons can compute logical functions. AND, NOT, OR

Computing logical functions: the OR function x0 f i1 w01 y0 i2 b=1 w02 w0b i1 i2 y0 1 Assume a binary threshold activation (TLU) function. What should you set w01, w02 and w0b to be so that you can get the right answers for y0?

Many answers would work y = f (w01i1 + w02i2 + w0bb) recall the threshold function the separation happens when w01i1 + w02i2 + w0bb = 0 move things around and you get i2 = - (w01/w02)i1 - (w0bb/w02) i2 i1

Linearly separable patterns PERCEPTRON is an architecture which can solve this type of decision boundary problem. An "on" response in the output node represents one class, and an "off" response represents the other. Linearly Separable Patterns

Models of Learning Hebbian ~ coincidence Supervised ~ correction (backprop) Recruitment ~ one trial Reinforcement ~ delayed reward Unsupervised ~ similarity

Hebb’s rule is not sufficient What happens if the neural circuit fires perfectly, but the result is very bad for the animal, like eating something sickening? A pure invocation of Hebb’s rule would strengthen all participating connections, which can’t be good. On the other hand, it isn’t right to weaken all the active connections involved; much of the activity was just recognizing the situation – we would like to change only those connections that led to the wrong decision. No one knows how to specify a learning rule that will change exactly the offending connections when an error occurs. Computer systems, and presumably nature as well, rely upon statistical learning rules that tend to make the right changes over time. Early models included perceptrons and the Delta rule XOR example requires multiple layers – backprop (advanced math)

Linearly separable patterns An architecture for a Perceptron which can solve this type of decision boundary problem. An "on" response in the output node represents one class, and an "off" response represents the other. Linearly Separable Patterns

Perceptron Training Rule new weight increment old weight Wi  Wi + dWi dWi = n * (t – o) * Xi step size perceptron output input target increment

Converges, if… … step size n sufficiently small … training data linearly separable 

Boolean XOR 1 o h1 h2 x1 x2 1 -0.5 XOR -1 1 -0.5 OR 1 -1.5 AND input x1 input x2 output 1 1 -0.5 OR 1 -1.5 AND h1 h2 x1 x2

Diagonal NOT linear separable y = f (w01i1 + w02i2 + w0bb) recall the threshold function i2 i1

Multi-layer Feed-forward Network

Supervised Learning - Backprop How do we train the weights of the network Basic Concepts Use a continuous, differentiable activation function (Sigmoid) Use the idea of gradient descent on the error surface Extend to multiple layers

The Sigmoid Function y=a x=net

Learning Rule – Gradient Descent on an Root Mean Square (RMS) Learn wi’s that minimize squared error O = output layer

Gradient Descent on an error

An informal account of BackProp For each pattern in the training set: Compute the error at the output nodes Compute Dw for each wt in 2nd layer Compute delta (generalized error expression) for hidden units Compute Dw for each wt in 1st layer After amassing Dw for all weights and, change each wt a little bit, as determined by the learning rate

Pattern Separation and NN architecture

CNNs solved the signal-translation problem, because they would convolve each input signal with a detector, (kernel), and thus be sensitive to the same feature, but everywhere.