Aula 4 Radial Basis Function Networks

Slides:



Advertisements
Similar presentations
Aula 3 Single Layer Percetron
Advertisements

Multi-Layer Perceptron (MLP)
Support Vector Machines
Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Neural Networks II CMPUT 466/551 Nilanjan Ray. Outline Radial basis function network Bayesian neural network.
Radial-Basis Function Networks CS/CMPE 537 – Neural Networks.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
6/10/ Visual Recognition1 Radial Basis Function Networks Computer Science, KAIST.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Aula 5 Alguns Exemplos PMR5406 Redes Neurais e Lógica Fuzzy.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
2806 Neural Computation Radial Basis Function Networks Lecture 5
Radial Basis-Function Networks. Back-Propagation Stochastic Back-Propagation Algorithm Step by Step Example Radial Basis-Function Networks Gaussian response.
Radial Basis Functions
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2008 Shreekanth Mandayam ECE Department Rowan University.
Radial Basis Function Networks 표현아 Computer Science, KAIST.
An Illustrative Example
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
Prediction Networks Prediction –Predict f(t) based on values of f(t – 1), f(t – 2),… –Two NN models: feedforward and recurrent A simple example (section.
Supervised Learning Networks. Linear perceptron networks Multi-layer perceptrons Mixture of experts Decision-based neural networks Hierarchical neural.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks I PROF. DR. YUSUF OYSAL.
Artificial Neural Network
Radial Basis Function (RBF) Networks
Radial Basis Function G.Anuradha.
Last lecture summary.
Radial-Basis Function Networks
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Radial Basis Function Networks
8/10/ RBF NetworksM.W. Mak Radial Basis Function Networks 1. Introduction 2. Finding RBF Parameters 3. Decision Surface of RBF Networks 4. Comparison.
Last lecture summary.
Neural networks.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Radial Basis Function Networks
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
Artificial Neural Networks Shreekanth Mandayam Robi Polikar …… …... … net k   
DIGITAL IMAGE PROCESSING Dr J. Shanbehzadeh M. Hosseinajad ( J.Shanbehzadeh M. Hosseinajad)
Neural Networks Artificial neural network (ANN) is a machine learning approach inspired by the way in which the brain performs a particular learning task.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Radial Basis Function Networks:
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
EE459 Neural Networks Examples of using Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University.
Non-Bayes classifiers. Linear discriminants, neural networks.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
CS621 : Artificial Intelligence
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 32: sigmoid neuron; Feedforward.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
EEE502 Pattern Recognition
Neural Networks 2nd Edition Simon Haykin
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Machine Learning Supervised Learning Classification and Regression
Radial Basis Function G.Anuradha.
CS621: Artificial Intelligence
Lecture 9 MLP (I): Feed-forward Model
Chapter 3. Artificial Neural Networks - Introduction -
Neuro-Computing Lecture 4 Radial Basis Function Network
Neural Network - 2 Mayank Vatsa
Lecture Notes for Chapter 4 Artificial Neural Networks
Introduction to Radial Basis Function Networks
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
Presentation transcript:

Aula 4 Radial Basis Function Networks PMR5406 Redes Neurais e Lógica Fuzzy Aula 4 Radial Basis Function Networks Baseado em: Neural Networks, Simon Haykin, Prentice-Hall, 2nd edition Slides do curso por Elena Marchiori, Vrije University

Radial-Basis Function Networks A function is approximated as a linear combination of radial basis functions (RBF). RBF’s capture local behaviour of functions. Biological motivation: RBF’s represent local receptors: PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network ARCHITECTURE Input layer: source of nodes that connect the NN with its environment. x1 w1 x2 y wm1 xm Hidden layer: applies a non-linear transformation from the input space to the hidden space. Output layer: applies a linear transformation from the hidden space to the output space. PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

φ-separability of patterns Hidden function Hidden space A (binary) partition, also called dichotomy, (C1,C2) of the training set C is φ-separable if there is a vector w of dimension m1 such that: PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Examples of φ-separability Separating surface: Examples of separable partitions (C1,C2): Linearly separable: Quadratically separable: Polynomial type functions Spherically separable: PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Cover’s Theorem (1) size of feature space φ =<1, …, m1> P(N, m1) - Probability that a particular partition (C1,C2) of the training set C picked at random is φ-separable Cover’s theorem. Under suitable assumptions on C = {x1, …, xN} and on the partitions (C1,C2) of C: PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Cover’s Theorem (2) Essentially P(N,m1) is a cumulative binomial distribution that corresponds to the probability of picking N points C = {x1, …, xN} (each one has a probability P(C1)=P(C2)=1/2) which are φ-separable using m1-1 or fewer degrees of freedom. PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Cover’s Theorem (3) P(N,m1) tends to 1 with the increase of m1 (size of feature space φ =<1, …, m1>). More flexibility with more functions in the feature space φ =<1, …, m1> PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Cover’s Theorem (4) A complex pattern-classification problem cast in a high-dimensional space non-linearly is more likely to be linearly separable than in a low-dimensional space. Corollary: The expected maximum number of randomly assigned patterns that are linearly separable in a space of dimension m1 is equal to 2m1 PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network HIDDEN NEURON MODEL Hidden units: use a radial basis function φ( || x - t||2) the output depends on the distance of the input x from the center t x1 φ( || x - t||2) t is called center  is called spread center and spread are parameters x2 xm PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Hidden Neurons A hidden neuron is more sensitive to data points near its center. This sensitivity may be tuned by adjusting the spread . Larger spread  less sensitivity Biological example: cochlear stereocilia cells have locally tuned frequency responses. PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Gaussian Radial Basis Function φ center φ :  is a measure of how spread the curve is: Large  Small  PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Types of φ Multiquadrics: Inverse multiquadrics: Gaussian functions: PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Example: the XOR problem (1,1) (0,1) (0,0) (1,0) x1 x2 Input space: Output space: Construct an RBF pattern classifier such that: (0,0) and (1,1) are mapped to 0, class C1 (1,0) and (0,1) are mapped to 1, class C2 y 1 PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Example: the XOR problem In the feature (hidden) space: When mapped into the feature space < 1 , 2 >, C1 and C2 become linearly separable. So a linear classifier with 1(x) and 2(x) as inputs can be used to solve the XOR problem. φ1 φ2 1.0 (0,0) 0.5 (1,1) Decision boundary (0,1) and (1,0) PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Learning Algorithms Parameters to be learnt are: centers spreads weights Different learning algorithms PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Learning Algorithm 1 Centers are selected at random center locations are chosen randomly from the training set Spreads are chosen by normalization: PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Learning Algorithm 1 Weights are found by means of pseudo-inverse method Desired response Pseudo-inverse of PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Learning Algorithm 2 Hybrid Learning Process: Self-organized learning stage for finding the centers Spreads chosen by normalization Supervised learning stage for finding the weights, using LMS algorithm PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Learning Algorithm 2: Centers K-means clustering algorithm for centers Initialization: tk(0) random k = 1, …, m1 Sampling: draw x from input space C Similarity matching: find index of best center Updating: adjust centers Continuation: increment n by 1, goto 2 and continue until no noticeable changes of centers occur PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Learning Algorithm 3 Supervised learning of all the parameters using the gradient descent method Modify centers Instantaneous error function Learning rate for Depending on the specific function can be computed using the chain rule of calculus PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Radial Basis Function Network Learning Algorithm 3 Modify spreads Modify output weights PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Comparison with multilayer NN RBF-Networks are used to perform complex (non-linear) pattern classification tasks. Comparison between RBF networks and multilayer perceptrons: Both are examples of non-linear layered feed-forward networks. Both are universal approximators. Hidden layers: RBF networks have one single hidden layer. MLP networks may have more hidden layers. PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Comparison with multilayer NN Neuron Models: The computation nodes in the hidden layer of a RBF network are different. They serve a different purpose from those in the output layer. Typically computation nodes of MLP in a hidden or output layer share a common neuron model. Linearity: The hidden layer of RBF is non-linear, the output layer of RBF is linear. Hidden and output layers of MLP are usually non-linear. PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network

Comparison with multilayer NN Activation functions: The argument of activation function of each hidden unit in a RBF NN computes the Euclidean distance between input vector and the center of that unit. The argument of the activation function of each hidden unit in a MLP computes the inner product of input vector and the synaptic weight vector of that unit. Approximations: RBF NN using Gaussian functions construct local approximations to non-linear I/O mapping. MLP NN construct global approximations to non-linear I/O mapping. PMR5406 Redes Neurais e Lógica Fuzzy Radial Basis Function Network