Introduction to Neural Networks 2. Overview  The McCulloch-Pitts neuron  Pattern space  Limitations  Learning.

Slides:



Advertisements
Similar presentations
God. NEURAL NETWORKS M. Alborzi, Ph. D. Petroleum University of Technology October, 2001.
Advertisements

Cellular Networks.
Perceptron Lecture 4.
Bożena Kunka Tutor: dr inż.. Jan Matuszewski The Application of Neural Networks in Radar Signals Recognition.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Artificial Neural Networks (1)
Perceptron Learning Rule
NEURAL NETWORKS Perceptron
Artificial Intelligence Techniques. Aims: Section fundamental theory and practical applications of artificial neural networks.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
An Overview of the Application of Neural Networks to the Monitoring of Civil Engineering Structures By Brian Walsh & Arturo González With thanks thanks.
1 Satellite Link Equations Introduction to Space Systems and Spacecraft Design Space Systems Design.
Neural Networks An Introduction.
The McCulloch-Pitts Neuron. Characteristics The activation of a McCulloch Pitts neuron is binary. Neurons are connected by directed weighted paths. A.
1 COMP305. Part I. Artificial neural networks.. 2 The McCulloch-Pitts Neuron (1943). McCulloch and Pitts demonstrated that “…because of the all-or-none.
December 1, 2011 A presentation by USFpedals.org A project of Environmental Studies students in “Humans & Environmental Change” (ENVA 109) University of.
Artificial Neural Network
Lecture Plan of Neural Networks Purpose From introduction to advanced topic To cover various NN models –Basic models Multilayer Perceptron Hopfield.
1 Introduction to Artificial Neural Networks Andrew L. Nelson Visiting Research Faculty University of South Florida.
Neural Networks.
Pulsed Neural Networks Neil E. Cotter ECE Department University of Utah.
Establishing the Equivalence between Recurrent Neural Networks and Turing Machines. Ritesh Kumar Sinha(02d05005) Kumar Gaurav Bijay( )
Mestrado em Ciência de Computadores Mestrado Integrado em Engenharia de Redes e Sistemas Informáticos VC 14/15 – TP19 Neural Networks & SVMs Miguel Tavares.
Presented by Scott Lichtor An Introduction to Neural Networks.
PSY105 Neural Networks 2/5 2. “A universe of numbers”
Radoslav Forgáč, Igor Mokriš Pulse Coupled Neural Network Models for Dimension Reduction of Classification Space Institute of Informatics Slovak Academy.
Automation of Engineering Design Aids using Neural Networks Siripong Malasri and Jittapong Malasri Christian Brothers University Kriangsiri Malasri Georgia.
Neural Network of C. elegans is a Small-World Network Masroor Hossain Wednesday, February 29 th, 2012 Introduction to Complex Systems.
ICS 586: Neural Networks Dr. Lahouari Ghouti Information & Computer Science Department.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Biological Neural Network & Nonlinear Dynamics Biological Neural Network Similar Neural Network to Real Neural Networks Membrane Potential Potential of.
NEURAL NETWORKS LECTURE 1 dr Zoran Ševarac FON, 2015.
Artificial Intelligence Methods Neural Networks Lecture 1 Rakesh K. Bissoondeeal Rakesh K.
Spiking Neural Networks Banafsheh Rekabdar. Biological Neuron: The Elementary Processing Unit of the Brain.
Vilalta/Norvig/Eick: History of AI Origins McCulloch and Pitts (1943) Model of Artificial Neurons. Donald Hebb (1949) Hebbian Learning Conference at Dartmouth.
Mini-course on Artificial Neural Networks and Bayesian Networks Michal Rosen-Zvi Mini-course on ANN and BN, The Multidisciplinary Brain Research center,
Introduction to Connectionism Jaap Murre Universiteit van Amsterdam en Universiteit Utrecht
1 Introduction to Neural Networks Recurrent Neural Networks.
Chapter 7: Introduction to Data Communications and Networking
Richard Dorrance Literature Review: 1/11/13
Joost N. Kok Universiteit Leiden
Convolutional Neural Networks
شبکه عصبی تنظیم: بهروز نصرالهی-فریده امدادی استاد محترم: سرکار خانم کریمی دانشگاه آزاد اسلامی واحد شهرری.
مناهــــج البحث العلمي
Artificial Intelligence Methods
Chapter 1: Introduction
CSE (c) S. Tanimoto, 2004 Neural Networks
نجاح وفشل المنشآت الصغيرة
Corso su Sistemi complessi:
Face Recognition with Neural Networks
Introduction to Neural Networks And Their Applications - Basics
Pulsed Neural Networks
CSE (c) S. Tanimoto, 2001 Neural Networks
McCulloch–Pitts Neuronal Model :
CSE (c) S. Tanimoto, 2002 Neural Networks
Mihir Patel and Nikhil Sardana
Lecture 02: Perceptron By: Nur Uddin, Ph.D.
Research Paper Overview.
III. Introduction to Neural Networks And Their Applications - Basics
FLIPPED CLASSROOM ACTIVITY CONSTRUCTOR – USING EXISTING CONTENT
ARTIFICIAL NEURAL NETWORK Intramantra Global Solution PVT LTD, Indore
CSE (c) S. Tanimoto, 2007 Neural Nets
Introduction to Neural Network
Introduction to Neural Networks
Artificial Neural Network
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Introduction to Neural Networks 2

Overview  The McCulloch-Pitts neuron  Pattern space  Limitations  Learning

McCulloch-Pitts model X1 X2 X3 W1 W2 W3 T Y Y=1 if W1X1+W2X2+W3X3  T Y=0 if W1X1+W2X2+W3X3<T

Introduce the bias Take the threshold over to the other side of the equation and replace it with a weight W0 which equals -T, and include a constant input X0 which equals 1. Y=1 if W1X1+W2X2+W3X3 - T  0 Y=0 if W1X1+W2X2+W3X3 -T <0 Y=1 if W1X1+W2X2+W3X3 +W0X0  0 Y=0 if W1X1+W2X2+W3X3 +W0X0 <0

Logic functions - OR X1 X2 1 1 Y Y = X1 OR X2 X0

Logic functions - AND X1 X2 1 1 Y Y = X1 AND X2 X0 -2

Logic functions - NOT X1 Y Y = NOT X1 X0 0

The weighted sum  The weighted sum, Σ WiXi is called the “net” sum.  Net = Σ WiXi  y=1 if net  0  y=0 if net < 0

Hard-limiter The threshold function is known as a hard-limiter. y 0 1 net When net is zero or positive, the output is 1, when net is negative the output is 0.

Example Original image Weights +1 0 Net = 14

Example with bias With a bias of -14, the weighted sum, net, is 0. Any pattern other than the original will produce a sum that is less than 0. If the bias is changed to -13, then patterns with 1 bit different from the original will give a sum that is 0 or more, so an output of 1.

Generalisation  The neuron can respond to the original image and to small variations  The neuron is said to have generalised because it recognises patterns that it hasn’t seen before

Pattern space  To understand what a neuron is doing, the concept of pattern space has to be introduced  A pattern space is a way of visualizing the problem  It uses the input values as co-ordinates in a space

Pattern space in 2 dimensions X2 X1 X1 X2 Y The AND function 1 0

Linear separability The AND function shown earlier had weights of -2, 1 and 1. Substituting into the equation for net gives: net = W0X0+W1X1+W2X2 = -2X0+X1+X2 Also, since the bias, X0, always equals 1, the equation becomes: net = -2+X1+X2

Linear separability The change in the output from 0 to 1 occurs when: net = -2+X1+X2 = 0 This is the equation for a straight line. X2 = -X1 + 2 Which has a slope of -1 and intercepts the X2 axis at 2. This line is known as a decision surface.

Linear separability X2 X1 X1 X2 Y The AND function

Linear separability  When a neuron learns it is positioning a line so that all points on or above the line give an output of 1 and all points below the line give an output of 0  When there are more than 2 inputs, the pattern space is multi-dimensional, and is divided by a multi-dimensional surface (or hyperplane) rather than a line

 Demo Demo  Demonstration on linearly/non-linearly separatability  Image analysis Image analysis  Generality demonstration

Are all problems linearly separable?  No  For example, the XOR function is non-linearly separable  Non-linearly separable functions cannot be implemented on a single neuron

Exclusive-OR (XOR) X2 X1 X1 X2 Y ? ?

Learning  A single neuron learns by adjusting the weights  The process is known as the delta rule  Weights are adjusted in order to minimise the error between the actual output of the neuron and the desired output  Training is supervised, which means that the desired output is known

Delta rule The equation for the delta rule is: ΔWi = ηXiδ = ηXi(d-y) where d is the desired output and y is the actual output. The Greek “eta”, η, is a constant called the learning coefficient and is usually less than 1. ΔWi means the change to the weight, Wi.

Delta rule  The change to a weight is proportional to Xi and to d-y.  If the desired output is bigger than the actual output then d - y is positive  If the desired output is smaller than the actual output then d - y is negative  If the actual output equals the desired output the change is zero

Changes to the weight

Example  Assume that the weights are initially random  The desired function is the AND function  The inputs are shown one pattern at a time and the weights adjusted

The AND function

Example Start with random weights of 0.5, -1, 1.5 When shown the input pattern the weighted sum is: net = 0.5 x 1 + (-1) x x 0 = 0.5 This goes through the hard-limiter to give an output of 1. The desired output is 0. So the changes to the weights are: W0negative W1zero W2zero

Example New value of weights (with η equal to 0.1) of 0.4, -1, 1.5 When shown the input pattern the weighted sum is: net = 1 x (-1) x x 1 = 1.9 This goes through the hard-limiter to give an output of 1. The desired output is 0. So the changes to the weights are: W0negative W1zero W2negative

Example New value of weights of 0.3, -1, 1.4 When shown the input pattern the weighted sum is: net = 1 x (-1) x x 0 = -0.7 This goes through the hard-limiter to give an output of 0. The desired output is 0. So the changes to the weights are: W0zero W1zero W2zero

Example New value of weights of 0.3, -1, 1.4 When shown the input pattern the weighted sum is: net = 1 x (-1) x x 1 = 0.7 This goes through the hard-limiter to give an output of 1. The desired output is 1. So the changes to the weights are: W0zero W1zero W2zero

Example - with η = 0.5

Example

What happened in pattern space X2 X

What happened in pattern space X2 X

What happened in pattern space X2 X

What happened in pattern space X2 X

What happened in pattern space X2 X

What happened in pattern space X2 X

What happened in pattern space X2 X

Conclusions  A single neuron can be trained to implement any linearly separable function  Training is achieved using the delta rule which adjusts the weights to reduce the error  Training stops when there is no error  Training is supervised

Conclusions  To understand what a neuron is doing, it help to picture what’s going on in pattern space  A linearly separable function can divide the pattern space into two areas using a hyperplane  If a function is not linearly separable, networks of neurons are needed