Applying Neural Networks Michael J. Watts

Slides:



Advertisements
Similar presentations
Chapter 8 Geocomputation Part B:
Advertisements

Slides from: Doug Gray, David Poole
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
also known as the “Perceptron”
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
EA, neural networks & fuzzy systems Michael J. Watts
Kohonen Self Organising Maps Michael J. Watts
Biological and Artificial Neurons Michael J. Watts
Artificial Neural Networks - Introduction -
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Lecture 14 – Neural Networks
Decision Support Systems
Neural Networks Marco Loog.
Neural Networks. R & G Chapter Feed-Forward Neural Networks otherwise known as The Multi-layer Perceptron or The Back-Propagation Neural Network.
Introduction to Neural Networks Simon Durrant Quantitative Methods December 15th.
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
Artificial Neural Networks
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Radial Basis Function (RBF) Networks
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Revision Michael J. Watts
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Introduction to AI Michael J. Watts
Artificial Neural Networks
Evolving a Sigma-Pi Network as a Network Simulator by Justin Basilico.
Artificial Neural Networks
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
© Negnevitsky, Pearson Education, Will neural network work for my problem? Will neural network work for my problem? Character recognition neural.
Multi-Layer Perceptrons Michael J. Watts
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Chapter 9 Neural Network.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 23 Nov 2, 2005 Nanjing University of Science & Technology.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
Back-Propagation MLP Neural Network Optimizer ECE 539 Andrew Beckwith.
Classification / Regression Neural Networks 2
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Artificial Intelligence Techniques Multilayer Perceptrons.
Programming for Geographical Information Analysis: Advanced Skills Online mini-lecture: Introduction to Neural Nets Dr Andy Evans.
Non-Bayes classifiers. Linear discriminants, neural networks.
Fuzzy Systems Michael J. Watts
Neural Networks and Backpropagation Sebastian Thrun , Fall 2000.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Fundamentals of Artificial Neural Networks Chapter 7 in amlbook.com.
Neural Networks Demystified by Louise Francis Francis Analytics and Actuarial Data Mining, Inc.
Neural Networks - lecture 51 Multi-layer neural networks  Motivation  Choosing the architecture  Functioning. FORWARD algorithm  Neural networks as.
Data Mining and Decision Support
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Perceptrons Michael J. Watts
Artificial Neural Networks for Data Mining. Copyright © 2011 Pearson Education, Inc. Publishing as Prentice Hall 6-2 Learning Objectives Understand the.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Today’s Lecture Neural networks Training
Machine Learning Supervised Learning Classification and Regression
Neural networks.
Applying Neural Networks
Fuzzy Systems Michael J. Watts
Real Neurons Cell structures Cell body Dendrites Axon
Neural Networks Dr. Peter Phillips.
EA, neural networks & fuzzy systems
Classification / Regression Neural Networks 2
Artificial Intelligence Methods
Artificial Intelligence Lecture No. 28
Lecture Notes for Chapter 4 Artificial Neural Networks
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Presentation transcript:

Applying Neural Networks Michael J. Watts

Lecture Outline When to use an ANN Preparing the data Apportioning data What kind of ANN to use Training ANN

When to use an ANN When the rules of the problem aren’t known When there is an underlying function o ANN can’t model random processes! ANN learn from existing data Careful preparation of the data is essential!

Preparing the Data Need a numerical representation of inputs Beware when encoding discrete items! o nominal vs. ordinal Binary coding should be -1,1 not 0,1 o hyperplane geometry & small initial weights o most published application papers ignore this, however

Preparing the Data Identify essential variables o data analysis Redundant variables complicate the learning process o curse of dimensionality Need more training examples than inputs Remove / transform outliers

Preparing the Data Normalisation o not strictly necessary o speeds training  sigmoid, tanh functions o again, either -1,1 or 0,1 o set mean 0 STD 1 MATLAB MLP simulator normalises inputs automatically

Preparing the Data Two kinds of problems o classification o function approximation / time series Classification has binary outputs Function approximation has continuous outputs

Preparing the Data Two ways of encoding multiple classes One output per class o sparse encoding o e.g , 0 1 0, Combining outputs o dense encoding o e.g. 1 0, 0 1, 0 0

Preparing the Data Some people use 0.1 & 0.9 instead of 0 & 1 Used to speed up training o limits of sigmoid function This is a gimmick Cannot interpret outputs as probabilities Better training algorithm should be used

Apportioning Data Splitting up the available data into multiple sets Two or three sets o training o testing o validation Each set has a different purpose

Apportioning Data Training data set is used to train the network Larger of the three sets Usually 50-75% of the data Should represent a wide range of output classes / values o mustn’t omit any target classes

Apportioning Data Testing data set is used to evaluate the performance of the trained network Network has not been trained on testing data Accuracy over testing set is an indicator of generalisation ability of the network

Apportioning Data 25% of the total data if no validation set it used Otherwise, half of the remaining data Must also represent an adequate range of output classes / values

Apportioning Data Validation data set is used as a final test Performance over this set is evaluated once Guards against optimising the network to the training set

Apportioning Data Similar to over-training Validation set consists of all remaining data As before, must have a wide range of output classes / values

Apportioning Data Can’t just split a data set into three to form training, testing and validation sets Statistical parameters of each set should be similar Distribution of target classes / values should be similar o can’t have all of class A in the training set and all of class B in the testing set

What Kind of ANN to use Perceptrons are useful for linearly separable problems MLP handle problems perceptrons cannot o non-linearly separable Both use supervised learning o require known outputs for each example

What Kind of ANN to use Kohonen SOMs are good for classification Can’t handle function approximation very well Unsupervised learning o no outputs required Find clusters in the data

Training ANN Goal is to maximise generalisation accuracy Problems with over-training o low training error o high generalisation error Common practice is to train until a low training error, then evaluate generalisation on test set

Training ANN Can lead to optimising on test set as well If the network is optimised for performance on the testing set, then the network is only good for that data set Use of validation data set is intended to counter this

Training ANN Number of hidden neurons must be selected for MLPs Some heuristics exist for this o most are not reliable To avoid over-training, number of connections should be less than the number of training examples o this is itself a heuristic, though

Training ANN Selection of parameters for backpropagation training Learning rate o must be determined by trial and error Momentum o less critical Number of epochs o effects generalisation

Training ANN Heuristics exist for each of these parameters Trial and error is the only reliable method Care must be taken in assessing generalisation o selection of validation set

Summary Applying ANN is something of a “black art” Each problem is different Careful preparation of data is critical Trial and error is often the only way to find an optimal network