Neural network architectures and learning algorithms Author : Bogdan M. Wilamowski Source : IEEE INDUSTRIAL ELECTRONICS MAGAZINE Date : 2011/11/22 Presenter.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

EE 690 Design of Embodied Intelligence
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Computing Gradient Vector and Jacobian Matrix in Arbitrarily Connected Neural Networks Author : Bogdan M. Wilamowski, Fellow, IEEE, Nicholas J. Cotton,
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Fall 2004 Shreekanth Mandayam ECE Department Rowan University.
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2008 Shreekanth Mandayam ECE Department Rowan University.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Fall 2004 Shreekanth Mandayam ECE Department Rowan University.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2010 Shreekanth Mandayam ECE Department Rowan University.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks / Spring 2002 Shreekanth Mandayam Robi Polikar ECE Department.
Multi Layer Perceptrons (MLP) Course website: The back-propagation algorithm Following Hertz chapter 6.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2006 Shreekanth Mandayam ECE Department Rowan University.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Neural networks.
Traffic Sign Recognition Using Artificial Neural Network Radi Bekker
Revision Michael J. Watts
Research on Advanced Training Algorithms of Neural Networks Hao Yu Ph.D Defense Aug 17 th 2011 Supervisor: Bogdan Wilamowski Committee Members: Hulya Kirkici.
Soft Computing Colloquium 2 Selection of neural network, Hybrid neural networks.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Cascade Correlation Architecture and Learning Algorithm for Neural Networks.
Artificial Neural Networks
© N. Kasabov Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering, MIT Press, 1996 INFO331 Machine learning. Neural networks. Supervised.
Multi Layer NN and Bit-True Modeling of These Networks SILab presentation Ali Ahmadi September 2007.
Waqas Haider Khan Bangyal. Multi-Layer Perceptron (MLP)
Lecture 3 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 3/1 Dr.-Ing. Erwin Sitompul President University
Classification / Regression Neural Networks 2
Artificial Intelligence Techniques Multilayer Perceptrons.
From Machine Learning to Deep Learning. Topics that I will Cover (subject to some minor adjustment) Week 2: Introduction to Deep Learning Week 3: Logistic.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
Last lecture summary. biologically motivated synapses Neuron accumulates (Σ) positive/negative stimuli from other neurons. Then Σ is processed further.
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
11 1 Backpropagation Multilayer Perceptron R – S 1 – S 2 – S 3 Network.
CS621 : Artificial Intelligence
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 32: sigmoid neuron; Feedforward.
Fundamentals of Artificial Neural Networks Chapter 7 in amlbook.com.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
NEURAL NETWORKS LECTURE 1 dr Zoran Ševarac FON, 2015.
Neural Networks 2nd Edition Simon Haykin
Perceptrons Michael J. Watts
Previous Lecture Perceptron W  t+1  W  t  t  d(t) - sign (w(t)  x)] x Adaline W  t+1  W  t  t  d(t) - f(w(t)  x)] f’ x Gradient.
Dimensions of Neural Networks Ali Akbar Darabi Ghassem Mirroshandel Hootan Nokhost.
Intro. ANN & Fuzzy Systems Lecture 11. MLP (III): Back-Propagation.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Machine Learning 12. Local Models.
Neural networks.
Multilayer Perceptrons
CSE 473 Introduction to Artificial Intelligence Neural Networks
Derivation of a Learning Rule for Perceptrons
Dr. Unnikrishnan P.C. Professor, EEE
CS621: Artificial Intelligence
Lecture 11. MLP (III): Back-Propagation
Synaptic DynamicsII : Supervised Learning
network of simple neuron-like computing elements
Backpropagation.
Multilayer Perceptron: Learning : {(xi, f(xi)) | i = 1 ~ N} → W
2. Matrix-Vector Formulation of Backpropagation Learning
3. Feedforward Nets (1) Architecture
Backpropagation.
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
Artificial Neural Networks / Spring 2002
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Neural network architectures and learning algorithms Author : Bogdan M. Wilamowski Source : IEEE INDUSTRIAL ELECTRONICS MAGAZINE Date : 2011/11/22 Presenter : 林哲緯 1

Outline Neural Architectures Parity-N Problem Suitable Architectures Use Minimum Network Size Conclusion 2

Neural Architectures 3 Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Neural Architectures 4 Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

Neural Architectures 5 Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

error back propagation(EBP) algorithm – multilayer perceptron (MLP) 6 Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0)

multilayer perceptron (MLP) 7 Neural network architectures and learning algorithms, Wilamowski, B.M. MLP-type architecture (without connections across layers)

neuron by neuron(NBN) algorithm – bridged multilayer perceptron (BMLP) – fully connected cascade (FCC) 8 Neural network architectures and learning algorithms, Wilamowski, B.M. arbitrarily connected network

neuron by neuron(NBN) algorithm Levenberg–Marquardt(LM) algorithm – Improve nonlinear function of least square – Forward & Backward Computation Jacobian Matrix – Forward-Only Computation 9

bridged multilayer perceptron (BMLP) 10 Neural network architectures and learning algorithms, Wilamowski, B.M. BMLP architecture 3=3=4=1(with connections across layers marked by dotted lines)

fully connected cascade (FCC) 11 Neural network architectures and learning algorithms, Wilamowski, B.M. Bipolar neural network for parity-8 problem in a FCC architecture

Outline Neural Architectures Parity-N Problem Suitable Architectures Use Minimum Network Size Conclusion 12

parity-8 problem MLP 8*9 + 9 = 81 weights BMLP 4* = 49 weights 13 Neural network architectures and learning algorithms, Wilamowski, B.M.

parity-8 problem = 42 weights 14 Neural network architectures and learning algorithms, Wilamowski, B.M.

parity-17 problem MLP architecture needs 18 neurons BMLP architecture with connections across hidden layers needs 9 neurons FCC architecture needs only 5 neurons 15

parity-N problem MLP architectures BMLP architectures FCC architectures nn = neurons nw = weights 16 Neural network architectures and learning algorithms, Wilamowski, B.M.

Outline Neural Architectures Parity-N Problem Suitable Architectures Use Minimum Network Size Conclusion 17

suitable architectures For a limited number of neurons, FCC neural networks are the most powerful architectures, but this does not mean that they are the only suitable architectures 18

suitable architectures if the two weights marked by red dotted lines – signal has to be propagated by fewer layers 19 Neural network architectures and learning algorithms, Wilamowski, B.M.

Outline Neural Architectures Parity-N Problem Suitable Architectures Use Minimum Network Size Conclusion 20

Use Minimum Network Size receive a close-to-optimum answer for all patterns that were never used in training generalization abilities 21

Case Study 22 Neural network architectures and learning algorithms, Wilamowski, B.M. TSK fuzzy controller: (a) Required control surface (b) 8*6 = 48 defuzzification rules TSK fuzzy controller: (a) Trapezoidal membership functions (b) Triangular membership functions

Case Study 23 Neural network architectures and learning algorithms, Wilamowski, B.M. (a) 3 neurons in cascade (12 weights), training error = (b) 4 neurons in cascade (18 weights), training error = (a) 5 neurons in cascade (25 weights), training error = (b) 8 neurons in cascade (52 weights), training error = 1.118E-005

time complexity NBN algorithm can train neural networks 1,000 times faster than the EBP algorithm. 24 Neural network architectures and learning algorithms, Wilamowski, B.M. (a)EBP algorithm, average solution time of 4.2s, and average iterations (b)NBN algorithm, average solution time of 2.4ms, and average 5.73 iterations

two-spiral problem 25 Neural network architectures and learning algorithms, Wilamowski, B.M. NBN algorithm using FCC architecture 244 iterations and 0.913s EBP algorithm using FCC architecture 30,8225 iterations and 342.7s

Outline Neural Architectures Parity-N Problem Suitable Architectures Use Minimum Network Size Conclusion 26

Conclusions FCC or BMLP architectures are not only more powerful but also easier to train use networks with a minimum number of neurons NBN have to invert a nw*nw matrix, but 500 weights are limit now. 27