Download presentation
Presentation is loading. Please wait.
Published byMaria Long Modified over 9 years ago
1
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple nodes communicate with each other through synapses that interconnect them Imitative of structure of biological nervous systems. Anns are more accurately described as a class of parallel algorithms
2
Knowledge in Anns Long term knowledge is stored in the networks in the states of the synaptic interconnections – in anns as weights between nodes Short term knowledge is temporarily stored in the on/off states of the nodes Both kinds of stored information determine how the network will respond to to inputs
3
Training of ANNS Networks are organized by by automated training methods, this simplifies the development of specific applications There is big advantage in all situations where no clear set of logical rules are given The inherent fault tolerance of nns is also a big advantage Nns can also be made to be tolerant against noise in the input : with increased noise the quality of the output only degrades slowly. (Graceful degradation)
4
Training of Networks A network will begin with no memories of the input space A network needs to go through a training phase in which it classifies input vectors
5
One of the major advantages of nns is their ability to generalize
One of the major advantages of nns is their ability to generalize. This means that a trained net could classify data from the same class as the learning data that it has never seen before Training set – used to train the net Validation set – used to determine the the performance of the net on patterns not trained during learning phase A test set for finally checking the over all performance of a NN
6
Mcculloch Pitts Model Acts a feature detector N inputs N weights
M outputs Threshold θ
7
Mcculloch Pitts Model Input is modulated by weighting the value of the connection Input is then integrated by the unit to produce the stimulation signal to the unit This becomes the activation If activation if >= θ the neuron fires Inhibitory input is absolute in keeping a neuron off
8
Mcculloch Pitts Model The mcculloch pitts model is severely limited
Can only categorize linearly separable domains No training regime
9
Rosenblatt’s Perceptron
Single layer network Again can categorize patterns However a training algorithm exist to adjust weights within the network which causes the network to ‘learn’ Hebbian learning algorithm used
10
Hebbian Learning Donald hebb proposed a learning theory
If a neuron X in a nervous system repeatedly stimulates a second neuron, Y to fire, then the pathways from X to Y becomes increasingly efficient in conveying that stimulus The perceptron is essentially mcculloch pitt’s model with hebbian training model
11
Problem With the Perceptron
Minksy & papert proved the perceptron couldn’t categorize a problem as simple as the XOR problem
12
Introduction of a Hidden Layer
The hidden layer solves the issue of linear separability This introduces the idea of a multi-layer network, the multi-perceptron is born
13
The Dawn of New Networks
Back-propagation Hopfield Kohonen SOM – self organizing maps
14
The Back-propagation Algorithm
Essentially a multi-layer perceptron with A different threshold function A more robust capable learning rule Backprop acts as a classifier/predictor eg. Loan evaluation A trained network is guaranteed to find relationships between input and output presented to it
15
Neural networks are universal approximators
Backprop has been shown to always find the right model it will always converge
16
Training a Back Propagation Network
An input pattern is propagated forward in the net until activation reaches the output layer The output at the output layer is compared with the teaching input The error j (if there is one) is propagated back through the network from the output through the hidden layers weights are adjusted so that all those nodes that contribute to the error are adjusted
17
Backprop In online learning the change in weights are applied to the network after each training pattern after each forward and back pass In offline or batch learning the weight changes are cumulated for all patterns in the input set after one full cycle (epoch) through the training pattern file In the back propagation algorithm online training is usually faster than batch training especially in the case of large training sets with many similar examples
18
Limitations of Neural Networks
Scalability Neural networks can become unstable when applied to larger problems
19
A simple example - OR (single layer see diagrams below)
We want a network to be able to respond to learn the OR input pattern Input Output The network has 2 inputs, and one output. All are binary. The output is determined by 1 if W0 *I0 + W1 * I1 > 0 0 if W0 *I0 + W1 * I1 <= 0 We want it to learn simple OR: output a 1 if either I0 or I1 is 1.
20
Single layer Input 1 output Input 2
21
Inside the neuron stage 1
W1 * input1 Input 1 output + = Activation Input 2 W2 W2 * input2
22
Inside the neuron stage 2
If activation (a) >= threshold () output = 1 else output =0 Input 1 output Input 2
23
Neuron fundamentals activation a = wixi threshold = output = 1 if a >= threshold output = 0 if a <threshold -1 > wi < 1
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.