Download presentation
Presentation is loading. Please wait.
1
Perceptrons Introduced in1957 by Rosenblatt
Used for pattern recognition Name is in use both for a particular artificial neuron model and for entire systems built from these neurons Introduced as a model for the visual system Heavily criticized by Minsky and Papert (1969) this caused a recession in ANN-research that lasted for more than a decade, until the advent of BP-learning for MLFF networks (Rumelhart e.a. 1986) and RNN-networks (Hopfield e.a ) 16-Jan-19 Rudolf Mak TU/e Computer Science
2
Single-layer Perceptrons
A discrete-neuron single-layer perceptron consists of an input layer of n real-valued input nodes (not neurons) an output layer of m neurons the output of a discrete neuron can only have the values zero (non firing) and one (firing) each neuron has a real-valued threshold and fires if and only if its accumulated input exceeds that threshold each connection from an input node j to an output neuron i has a real-valued weight wij It computes a vector function f: Rn ! {0,1}m 16-Jan-19 Rudolf Mak TU/e Computer Science
3
Questions Since a perceptron with n input nodes and m output nodes computes a function Rn ! {0,1}m, we therefore study the questions: Which functions can be computed? Does there exist a learning method, i.e. is there an algorithm that optimizes the weights? 16-Jan-19 Rudolf Mak TU/e Computer Science
4
Single-layer Single-output Perceptron
We start with the simplest configuration: A single-layer single-output perceptron consists of a single neuron whose output is either zero or one, and is given by -w0 is called the threshold 16-Jan-19 Rudolf Mak TU/e Computer Science
5
Where do we put the threshold
Heaviside function Linear combiner Heaviside + threshold Affine combiner Standard Heaviside 16-Jan-19 Rudolf Mak TU/e Computer Science
6
Artificial Neuron affine combiner transfer function 16-Jan-19
Synonyms Adder, integrator i.p.v. linear combiner Activation function i.p.v. transfer function squashing function W0 is the threshold also called bias V = Sum wk xk + w0 local field of activation potential affine combiner transfer function 16-Jan-19 Rudolf Mak TU/e Computer Science
7
Form affine to linear combiners
16-Jan-19 Rudolf Mak TU/e Computer Science
8
Boolean Function: AND logical geometrical X Y X Æ Y 1 2x + 2y > 3
1 2x + 2y > 3 2x + 2y < 3 16-Jan-19 Rudolf Mak TU/e Computer Science
9
Boolean Function: OR X Y X Ç Y 1 16-Jan-19
1 16-Jan-19 Rudolf Mak TU/e Computer Science
10
Boolean Functions: XOR
Y X © Y 1 16-Jan-19 Rudolf Mak TU/e Computer Science
11
Linearly Separable Sets
A set X 2 Rn £ {0,1} is called (absolutely) linearly separable if there exists a vector w 2 Rn+1 such that for each pair (x,t) 2 X : A training set X is correctly classified by a perceptron if for each (x,t) 2 X the output of the perceptron with input x is also t. A finite set X can be classified correctly by a one-layer perceptron if and only if it is linearly separable. 16-Jan-19 Rudolf Mak TU/e Computer Science
12
A Linearly Separable Set (in 2D)
16-Jan-19 Rudolf Mak TU/e Computer Science
13
Not linearly separable set (in 2D)
16-Jan-19 Rudolf Mak TU/e Computer Science
14
One-layer Perceptron Learning
Since the output neurons of a one-layer perceptron are independent, it suffices to study perceptron with a single output. Consider a finite set also called a training set. We say that such a set X is correctly classified by a perceptron, if for each pair (x,t) in X the output of the perceptron with input x is t. A finite set X can be classified correctly by a one-layer perceptron if and only if it is linearly separable. 16-Jan-19 Rudolf Mak TU/e Computer Science
15
Perceptron Learning Rule (incremental version)
16-Jan-19 Rudolf Mak TU/e Computer Science
16
Geometric Interpretation
< 0 > 0 The weights are modified such that the angle with the input vector is decreased. 16-Jan-19 Rudolf Mak TU/e Computer Science
17
Geometric Interpretation
The weights are modified such that the angle with the input vector is increased. 16-Jan-19 Rudolf Mak TU/e Computer Science
18
Perceptron Convergence Theorem
Let X be a finite, linearly separable training set. Let the initial weight vector and the learning parameter be chosen an arbitrary positive number. Then for each infinite sequence of training pairs from X, the sequence of weight vectors obtained by applying the perceptron learning rule converges in a finite number of steps. 16-Jan-19 Rudolf Mak TU/e Computer Science
19
Proof sketch 1 16-Jan-19 Rudolf Mak TU/e Computer Science
20
Proof sketch 2 16-Jan-19 Rudolf Mak TU/e Computer Science
21
Proof sketch 3 16-Jan-19 Rudolf Mak TU/e Computer Science
22
Remarks The perceptron learning algorithm is a form of reinforcement learning and is due to Rosenblatt By adjusting the weights sufficiently the network may learn the current training vector. Other vectors, however, may be unlearned Although the learning algorithm converges for any positive learning parameter , faster convergence can be obtained by a suitable choice, possible dependent on the observed error Scaling of the input vectors can also be beneficial to the convergence of the algorithm 16-Jan-19 Rudolf Mak TU/e Computer Science
23
Perceptron Learning Rule (batch version)
See lecture notes for a proof of convergence (similar to the incremental version). The inner loop is called training for an epoch (Matlab) 16-Jan-19 Rudolf Mak TU/e Computer Science
24
Learning by Error Minimization
Consider the error function Then the gradient of E (w) is given by Hence the weight updates (batch version) are given by 16-Jan-19 Rudolf Mak TU/e Computer Science
25
Capacity of One-layer Perceptrons
The number of boolean functions of n arguments is 2(2n) Each boolean function defines a dichotomy of the points of an n-dimensional hypercube The number of linear dichotomies Bn of the corner points of the hypercube is bounded by C(2n, n), where C(m, n) is the number of linear dichotomies of m points in Rn (in general position) which is given by Results should be known. Derivation not. No questions about section 2.3 in the exam. 16-Jan-19 Rudolf Mak TU/e Computer Science
26
# bool fie versus # lin. sep. dichotomies
Bn 2 1 4 16 14 3 256 104 65536 1882 5 94572 6 7 For n = 2 there are two boolean functions that are not given by linear dichotomies. Which are these? How many linear dichotomies are there for the 5 corners of a regular pentagon? Hint they are in general position. 16-Jan-19 Rudolf Mak TU/e Computer Science
27
Multi-layer Perceptrons
A discrete-neuron multi-layer perceptron consists of an input layer of n real-valued input nodes (not neurons) an output layer of m neurons several intermediate (hidden) layers consisting of one or more neurons. with exception of the last layer the nodes of each layer serve as inputs to the nodes of the next layer each connection from node j in layer k-1 to node i in layer k has a real valued weight wijk It computes a function f: Rn ! {0,1}m 16-Jan-19 Rudolf Mak TU/e Computer Science
28
Graphical representation
input nodes output nodes edge direction left to right not drawn hidden layers 16-Jan-19 Rudolf Mak TU/e Computer Science
29
Discrete Multi-layer Perceptrons
The computational capabilities of multi-layer perceptrons for two and three layers are given by Every boolean function can be computed by a two-layer perceptron Every region in Rn that is bounded by a finite number of n-1 dimensional hyperplanes can be classified by a three-layer perceptron Unfortunately there is no simple learning algorithm for multi-layer perceptrons 16-Jan-19 Rudolf Mak TU/e Computer Science
30
Disjunctive Normal Form
x1 x2 x3 x4 x5 f(x1,x2,x3,x4,x5) 1 Clause Cj x1x2 x3 x4 x5 literals Also known as sum-of-products by electrical engineers Logic table for f 16-Jan-19 Rudolf Mak TU/e Computer Science
31
Perceptron for a Clause
In fact this is a McCulloch Pitts neuron. Edges with weight +1 are excitatory edges. Edges with weight -1 are called inhibitory edges. This is a generalized And-gate 16-Jan-19 Rudolf Mak TU/e Computer Science
32
2-layer perceptron for a boolean function
Neuron in the second layer is a generalized or-gate How would it look if negated variables are allowed (threshold 0.5 – s). 16-Jan-19 Rudolf Mak TU/e Computer Science
33
XOR revisited 16-Jan-19 Rudolf Mak TU/e Computer Science
Use the standard construction 16-Jan-19 Rudolf Mak TU/e Computer Science
34
XOR revisited again 16-Jan-19 Rudolf Mak TU/e Computer Science
Non-layered neural net strictly not a perceptron 16-Jan-19 Rudolf Mak TU/e Computer Science
35
Minsky Papert observation
No diameter limited perceptron can determine whether a geometric figure is connected A and D not connected B and C connected If necessary figures are stretched in the horizontal direction to exceed the limits of perception of the individual neurons A B C D 16-Jan-19 Rudolf Mak TU/e Computer Science
36
Diameter limited perceptron
Each circle is stands for a clause that has a limited number of inputs (receptive fields) Clauses (also called predicates) can be classified in three groups Left group G1 distinguishes a pattern from {A, C} from a pattern from {B, D} Right group G3 distinguishes a pattern from {A, B} from a pattern from {C, D} Middle group has no discriminating power. 16-Jan-19 Rudolf Mak TU/e Computer Science
37
16-Jan-19 Rudolf Mak TU/e Computer Science
38
Star Region 16-Jan-19 Rudolf Mak TU/e Computer Science
39
3-layer perceptron for star region
Set up equations for the lines. Let the origin lie at the center of region R6 For each line let the origin lie in the positive half-space Let l1: 1-x2 = 0 then (0,0) in positive half-space W0 = (1, 1, -1) T Normal vector for line l1 is (1, -1) Rotate this vector over angle f = 2pi/5 to obtain the normal vectors of the other lines. Premultiply by matrix [ cos f -sin f ] counterclockwise [ sin f cos f ] 16-Jan-19 Rudolf Mak TU/e Computer Science
40
Summary One-layer perceptrons have limited computational capabilities. Only linearly separable sets can be classified. For one-layer perceptrons there exists a learning algorithm with robust convergence properties. Multi-layer perceptrons have larger computational capabilities (all boolean functions for two-layer perceptrons), but for those there does not exist a simple learning algorithm. 16-Jan-19 Rudolf Mak TU/e Computer Science
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.