Multi-Layer Perceptrons Michael J. Watts

Slides:



Advertisements
Similar presentations
1 Image Classification MSc Image Processing Assignment March 2003.
Advertisements

G53MLE | Machine Learning | Dr Guoping Qiu
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
also known as the “Perceptron”
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Biological and Artificial Neurons Michael J. Watts
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Overview over different methods – Supervised Learning
Lecture 14 – Neural Networks
Simple Neural Nets For Pattern Classification
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Neural Networks Marco Loog.
MACHINE LEARNING 12. Multilayer Perceptrons. Neural Networks Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Artificial Neural Networks
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Radial Basis Function (RBF) Networks
Neural Networks Lecture 8: Two simple learning algorithms
Where We’re At Three learning rules  Hebbian learning regression  LMS (delta rule) regression  Perceptron classification.
Radial Basis Function Networks
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Neural Networks
Artificial Neural Networks
© N. Kasabov Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering, MIT Press, 1996 INFO331 Machine learning. Neural networks. Supervised.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
ANNs (Artificial Neural Networks). THE PERCEPTRON.
Machine Learning Chapter 4. Artificial Neural Networks
Classification / Regression Neural Networks 2
Radial Basis Function Networks:
Artificial Intelligence Techniques Multilayer Perceptrons.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Applying Neural Networks Michael J. Watts
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
Soft computing Lecture 7 Multi-Layer perceptrons.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
CS621 : Artificial Intelligence
For Friday No reading Take home exam due Exam 2. For Monday Read chapter 22, sections 1-3 FOIL exercise due.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Chapter 2 Single Layer Feedforward Networks
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall Perceptron Rule and Convergence Proof Capacity.
EEE502 Pattern Recognition
COMP53311 Other Classification Models: Neural Network Prepared by Raymond Wong Some of the notes about Neural Network are borrowed from LW Chan’s notes.
Perceptrons Michael J. Watts
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Machine Learning Supervised Learning Classification and Regression
Neural networks.
Applying Neural Networks
Other Classification Models: Neural Network
Real Neurons Cell structures Cell body Dendrites Axon
CS621: Artificial Intelligence
Classification Neural Networks 1
Chapter 3. Artificial Neural Networks - Introduction -
Artificial Intelligence Methods
Training a Neural Network
Neuro-Computing Lecture 4 Radial Basis Function Network
Lecture Notes for Chapter 4 Artificial Neural Networks
David Kauchak CS158 – Spring 2019
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Multi-Layer Perceptrons Michael J. Watts

Lecture Outline Perceptron revision Multi-Layer Perceptrons Terminology Advantages Problems

Perceptron Revision Two neuron layer networks Single layer of adjustable weights Feed forward network Cannot handle non-linearly-separable problems Supervised learning algorithm Mostly used for classification

Multi-Layer Perceptrons Otherwise known as MLPs Adds an additional layer (or layers) of neurons to a perceptron Additional layer called hidden (or intermediate) layer Additional layer of adjustable connections

Multi-Layer Perceptrons Proposed mid-1960s, but learning algorithms not available until mid-1980s Continuous inputs Continuous outputs Continuous activation functions used o e.g. sigmoid, tanh Feed forward networks

Multi-Layer Perceptrons

MLPs can also have biases A bias is an additional, external, input Input to bias node is always one Bias performs the task of a threshold

Multi-Layer Perceptrons

Able to model non-linearly separable functions Each hidden neuron is equivalent to a perceptron Each hidden neuron adds a hyperplane to the problem space Requires sufficient neurons to partition the input space appropriately

Terminology Some define MLP according to the number of connection layers Others define MLP according to the number of neuron layers Some don’t count the input neuron layer o doesn’t perform processing Some refer to the number of hidden neurons layers

Terminology Best to specify which o i.e. “three neuron layer MLP”, or o “two connection layer MLP”, or o “three neuron layer, one hidden layer MLP”

Advantages A MLP with one hidden layer of sufficient size can approximate any continuous function to any desired accuracy o Kolmogorov theorem MLP can learn conditional probabilities MLP are multivariate non-linear regression models

Advantages Learning models o several ways of training a MLP o backpropagation is the most common Universal function approximators Able to generalise to new data o can accurately identify / model previously unseen examples

Problems with MLP Choosing the number of hidden layers o how many are enough? o one should be sufficient, according to the Kolmogorov Theorem o two will always be sufficient

Problems with MLP Choosing the number of hidden nodes o how many are enough? o Usually, the number of connections in the network should be less than the number of training examples As number of connections approaches the number of training examples, generalisation decreases

Problems with MLP Representational capacity Catastrophic forgetting Occurs when a trained network is further trained on new data Network forgets what it learned about the old data Only knows about the new data

Problems with MLP Initialisation of weights Random initialisation can cause problems with training o start in a bad spot Can initialise using statistics of training data o PCA o Nearest neighbour

Problems with MLP Weight initialisation o can initialise using an evolutionary algorithm o initialisation from decision tress o initialisation from rules

Summary MLP overcome the linear separability problem of perceptrons Add an additional layer of neurons Additional layer of variable connections No agreement on terminology to use o be precise!

Summary Convergence is guaranteed over continuous functions Problems exist o but solutions also exist