Neural and Evolutionary Computing - Lecture 9 1 Evolutionary Neural Networks Design  Motivation  Evolutionary training  Evolutionary design of the architecture.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

© Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems Introduction.
Population-based metaheuristics Nature-inspired Initialize a population A new population of solutions is generated Integrate the new population into the.
Slides from: Doug Gray, David Poole
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
1 Evolutionary Computational Inteliigence Lecture 6b: Towards Parameter Control Ferrante Neri University of Jyväskylä.
EA, neural networks & fuzzy systems Michael J. Watts
Machine Learning Neural Networks
Data Mining Techniques Outline
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Neural Networks Marco Loog.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
The Performance of Evolutionary Artificial Neural Networks in Ambiguous and Unambiguous Learning Situations Melissa K. Carroll October, 2004.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Radial-Basis Function Networks
Neural networks - Lecture 111 Recurrent neural networks (II) Time series processing –Networks with delayed input layer –Elman network Cellular networks.
Radial Basis Function Networks
A Genetic Algorithms Approach to Feature Subset Selection Problem by Hasan Doğu TAŞKIRAN CS 550 – Machine Learning Workshop Department of Computer Engineering.
Genetic Algorithms and Ant Colony Optimisation
Biointelligence Laboratory, Seoul National University
Evolving a Sigma-Pi Network as a Network Simulator by Justin Basilico.
Multiple-Layer Networks and Backpropagation Algorithms
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
Integrating Neural Network and Genetic Algorithm to Solve Function Approximation Combined with Optimization Problem Term presentation for CSC7333 Machine.
Soft Computing Lecture 18 Foundations of genetic algorithms (GA). Using of GA.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
Neural Networks Chapter 6 Joost N. Kok Universiteit Leiden.
Chapter 11 – Neural Networks COMP 540 4/17/2007 Derek Singer.
Introduction to Artificial Neural Network Models Angshuman Saha Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg.
Lecture 8: 24/5/1435 Genetic Algorithms Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Classification / Regression Neural Networks 2
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Zorica Stanimirović Faculty of Mathematics, University of Belgrade
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Neural and Evolutionary Computing - Lecture 6
Neural and Evolutionary Computing - Lecture 5 1 Evolutionary Computing. Genetic Algorithms Basic notions The general structure of an evolutionary algorithm.
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
1 Genetic Algorithms and Ant Colony Optimisation.
Genetic Algorithms Przemyslaw Pawluk CSE 6111 Advanced Algorithm Design and Analysis
Chapter 12 FUSION OF FUZZY SYSTEM AND GENETIC ALGORITHMS Chi-Yuan Yeh.
Neural Networks - Berrin Yanıkoğlu1 Applications and Examples From Mitchell Chp. 4.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
CITS7212: Computational Intelligence An Overview of Core CI Technologies Lyndon While.
Neural Networks - lecture 51 Multi-layer neural networks  Motivation  Choosing the architecture  Functioning. FORWARD algorithm  Neural networks as.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Waqas Haider Bangyal 1. Evolutionary computing algorithms are very common and used by many researchers in their research to solve the optimization problems.
Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg
Data Mining and Decision Support
Learning Neural Networks (NN) Christina Conati UBC
CHEE825 Fall 2005J. McLellan1 Nonlinear Empirical Models.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
129 Feed-Forward Artificial Neural Networks AMIA 2003, Machine Learning Tutorial Constantin F. Aliferis & Ioannis Tsamardinos Discovery Systems Laboratory.
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Chapter 12 Case Studies Part B. Control System Design.
Machine Learning Supervised Learning Classification and Regression
Evolutionary Algorithms Jim Whitehead
Deep Feedforward Networks
One-layer neural networks Approximation problems
EA, neural networks & fuzzy systems
Classification / Regression Neural Networks 2
Machine Learning Today: Reading: Maria Florina Balcan
CSC 578 Neural Networks and Deep Learning
Neural Network - 2 Mayank Vatsa
The use of Neural Networks to schedule flow-shop with dynamic job arrival ‘A Multi-Neural Network Learning for lot Sizing and Sequencing on a Flow-Shop’
Lecture 4. Niching and Speciation (1)
Presentation transcript:

Neural and Evolutionary Computing - Lecture 9 1 Evolutionary Neural Networks Design  Motivation  Evolutionary training  Evolutionary design of the architecture  Evolutionary design of the learning rules

Neural and Evolutionary Computing - Lecture 9 2 Evolutionary Neural Networks Design Motivation. Neural networks design consists of:  Choice of the architecture (network topology+connectivity)  Has an influence on the network ability to solve the problem  Usually is a trial-and-error process  Train the network  Is an optimization problem = find the parameters (weights) which minimize the error on the training set  The classical methos (ex: BackPropagation) have some drawbacks :  If the activation functions are not differentiable  Risk of getting stuck in local minima

Neural and Evolutionary Computing - Lecture 9 3 Evolutionary Neural Networks Design Idea: use an evolutionary process  Inspired by the biological evolution of the brain  The system is not explicitely designed but its structure derives by an evolutionary process involving a population of encoded neural networks –Genotype = the network codification (structural description) –Phenotype = the network itself, which can be simulated (functional description)

Neural and Evolutionary Computing - Lecture 9 4 Evolutionary Training  Use an evolutionary algorithm to solve the problem of minimizing the mean squared error on the training set  Parameters: synaptic weights and biases

Neural and Evolutionary Computing - Lecture 9 5 Evolutionary Training Evolutionary algorithm components:  Encoding: each element of the population is a real vector containing all adaptive parameters (similar to the case of evolution strategies)  Evolutionary operators: typical to evolution strategies or evolutionary programming  Evaluation: the quality of an element depends on the mean squared error (MSE) on the training/validation set; an element is better if the MSE is smaller

Neural and Evolutionary Computing - Lecture 9 6 Evolutionary Training Aplications:  For networks with non-differentiable or non-continuous activation functions  For recurrent networks (the output value cannot be explicitely computing from the input value, thus the derivative based learning algorithms cannot be applied Drawbacks:  More costly than traditional non-evolutionary training  It is not appropriate for fine tuning the parameters Hybrid versions:  Use an EA to explore the parameter space and a local search technique to refine the values of the parameters

Neural and Evolutionary Computing - Lecture 9 7 Evolutionary Training Remark. EAs can be used to preprocess the training set Selection of attributes Selection of examples

Neural and Evolutionary Computing - Lecture 9 8 Evolutionary Pre-processing Selection of attributes (for classification problems) Motivation: if the number of attributes is large the training is difficult It is important when some of the attributes are not relevant for the classification task The aim is to select the relevant attributes For initial data having N attributes the encoding could be a vector of N binary values (0 – not selected, 1 – selected) The evaluation is based on training the network for the selected attributes (this corresponds to a wrapper-like technique of attributes selection)

Neural and Evolutionary Computing - Lecture 9 9 Evolutionary Pre-processing Example: identify patients with cardiac risk Total set of attributes: (age, weight, height, body mass index, blood pressure, cholesterol, glucose level) Population element: (1,0,0,1,1,1,0) Corresponding subset of attributes: (age, body mass index, blood pressure, cholesterol) Evaluation: train the network using the subset of selected attributes and compute the accuracy; the fitness value will be proportional to the accuracy Remark: This technique can be applied also for non neural classifiers (ex: Nearest-Neigbhor) It is called “wrapper based attribute selection”

Neural and Evolutionary Computing - Lecture 9 10 Evolutionary Pre-processing Selection of examples Motivation: if the training set is large the training process is costly and there is a higher risk of overfitting It is similar to attribute selection Binary encoding (0 – not selected, 1 – selected) The evaluation is based on training the network (using any training algorithm) for the subset specified by the binary encoding

Neural and Evolutionary Computing - Lecture 9 11 Evolving architecture Non-evolutionary approaches:  Growing networks  Start with a small size network  If the training stagnates add a new unit  The assimilation of the new unit is based on adjusting, in a first stage, only its weights  Pruning networks  Start with a large size network  The units and connection which do not influence the training process are eliminated

Neural and Evolutionary Computing - Lecture 9 12 Evolving architecture Elements which can be evolved:  Number of units  Connectivity  Activation function type Encoding variants:  Direct  Indirect

Neural and Evolutionary Computing - Lecture 9 13 Evolving architecture Direct encoding: each element of the architecture appears explicitly in the encoding Network architecture = oriented graph The network can be encoded by the adjacency matrix Rmk. For feedforward networks the units can be numbered such that the unit i receives signals only from units j, such that j inferior triangular matrix Architecture Adjacency matrix Chromosome

Neural and Evolutionary Computing - Lecture 9 14 Evolving architecture Operators Crossover similar to that used for genetic algorithms

Neural and Evolutionary Computing - Lecture 9 15 Evolving architecture Operators: Mutation similar to that used for genetic algorithms

Neural and Evolutionary Computing - Lecture 9 16 Evolving architecture Evolve the number of units an connections Hypothesis: N – maximal number of units Encoding: Binary vector with N elements –0: inactivated unit –1: activ unit Adjacency matrix NxN –For a zero element in the unit vector the corresponding row and column in the matrix are ignored.

Neural and Evolutionary Computing - Lecture 9 17 Evolving architecture Evolving the activation function type: Encoding : Binary vector with N elements –0: inactivated unit –1: active unit with activation function of type 1 (ex: tanh) –2: active unit with activation function of type 2 (ex: logistic) –3: active unit with activation function of type 3 (ex: linear) Evolution of weights The adjacency matrix is replaced with the matrix of weights –0: no coonection –<>0: weight value

Neural and Evolutionary Computing - Lecture 9 18 Evolving architecture Evaluation: The network is trained The training error is estimated (E a ) The validation error is estimated (E v ) The fitness is inverse proportional to: –Training error –Validation error –Network size

Neural and Evolutionary Computing - Lecture 9 19 Evolving architecture Drawbacks of the direct encoding: It is not scalable Can lead to different representations of the same network (permutation problem) It is not appropriate for modular networks

Neural and Evolutionary Computing - Lecture 9 20 Evolving architecture Indirect encoding: Biological motivation Parametric encoding –The network is described by a set of characteristics (fingerprint) –Particular case: feedforward network with variable number of hidden units –The fingerprint is instantiated in a network only for evaluation Rules-based encoding

Neural and Evolutionary Computing - Lecture 9 21 Evolving architecture Parametric encoding Instantiation: random choice of connections according to the specified characteristics

Neural and Evolutionary Computing - Lecture 9 22 Evolving architecture Example: Operators: Mutation: change the network characteristics Recombination: combine characteristics of layers Nr. of layers Param. BP Info. layer 1 Info. layer 2

Neural and Evolutionary Computing - Lecture 9 23 Evolving architecture Rule-based encoding (similar to Grammar Evolution) : General rule Examples: Structure of an element:

Neural and Evolutionary Computing - Lecture 9 24 Evolving architecture Deriving an architecture:

Neural and Evolutionary Computing - Lecture 9 25 Learning rules General form of a local adjustement rule x i,x j – input signals y i,y j – output signals α – control parameters (ex: learning rate) δ i,δ j – error signal Example: BackPropagation

Neural and Evolutionary Computing - Lecture 9 26 Learning rules Elements which can be evolved: Parameters of the learning process (ex: learning rate, momentum coefficient) The adjusting expression (see Genetic Programming) Evaluation: Train networks using the corresponding rule Drawback: very high cost

Neural and Evolutionary Computing - Lecture 9 27 Summary General structure Levels: Weights Learning rules Architecture