RBF TWO-STAGE LEARNING NETWORKS: EXPLOITATION OF SUPERVISED DATA IN THE SELECTION OF HIDDEN UNIT PARAMETERS An application to SAR data classification.

Slides:



Advertisements
Similar presentations
Computational Intelligence Winter Term 2009/10 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Advertisements

Computational Intelligence Winter Term 2011/12 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Computational Intelligence Winter Term 2014/15 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Introduction to Neural Networks Computing
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Support Vector Machines
Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
1 Introduction to Bio-Inspired Models During the last three decades, several efficient machine learning tools have been inspired in biology and nature:
Neural Nets Using Backpropagation Chris Marriott Ryan Shirley CJ Baker Thomas Tannahill.
Neural Networks II CMPUT 466/551 Nilanjan Ray. Outline Radial basis function network Bayesian neural network.
Radial-Basis Function Networks CS/CMPE 537 – Neural Networks.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Decision Support Systems
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
6/10/ Visual Recognition1 Radial Basis Function Networks Computer Science, KAIST.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Radial Basis Function Networks 표현아 Computer Science, KAIST.
Prediction Networks Prediction –Predict f(t) based on values of f(t – 1), f(t – 2),… –Two NN models: feedforward and recurrent A simple example (section.
Supervised Learning Networks. Linear perceptron networks Multi-layer perceptrons Mixture of experts Decision-based neural networks Hierarchical neural.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks I PROF. DR. YUSUF OYSAL.
CS Instance Based Learning1 Instance Based Learning.
Aula 4 Radial Basis Function Networks
Radial Basis Function (RBF) Networks
Radial-Basis Function Networks
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Radial Basis Function Networks
8/10/ RBF NetworksM.W. Mak Radial Basis Function Networks 1. Introduction 2. Finding RBF Parameters 3. Decision Surface of RBF Networks 4. Comparison.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Radial Basis Function Networks
Chapter 4 Supervised learning: Multilayer Networks II.
Artificial Neural Networks Shreekanth Mandayam Robi Polikar …… …... … net k   
Multi-Layer Perceptrons Michael J. Watts
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Radial Basis Function Networks:
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Well Log Data Inversion Using Radial Basis Function Network Kou-Yuan Huang, Li-Sheng Weng Department of Computer Science National Chiao Tung University.
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Chapter 2 Single Layer Feedforward Networks
Fast Learning in Networks of Locally-Tuned Processing Units John Moody and Christian J. Darken Yale Computer Science Neural Computation 1, (1989)
Computational Intelligence Winter Term 2015/16 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Chapter 6 Neural Network.
Artificial Neural Networks for Data Mining. Copyright © 2011 Pearson Education, Inc. Publishing as Prentice Hall 6-2 Learning Objectives Understand the.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning 12. Local Models.
Machine Learning Supervised Learning Classification and Regression
Predictive Learning from Data
Machine Learning Clustering: K-means Supervised Learning
Neural Networks Winter-Spring 2014
Chapter 2 Single Layer Feedforward Networks
General Aspects of Learning
Computational Intelligence
Neuro-Computing Lecture 4 Radial Basis Function Network
Machine Learning Ensemble Learning: Voting, Boosting(Adaboost)
Neural Networks Chapter 5
Computational Intelligence
Lecture Notes for Chapter 4 Artificial Neural Networks
Introduction to Radial Basis Function Networks
Computational Intelligence
Machine Learning Support Vector Machine Supervised Learning
COSC 4335: Part2: Other Classification Techniques
Machine Learning Perceptron: Linearly Separable Supervised Learning
Prediction Networks Prediction A simple example (section 3.7.3)
Computational Intelligence
Presentation transcript:

RBF TWO-STAGE LEARNING NETWORKS: EXPLOITATION OF SUPERVISED DATA IN THE SELECTION OF HIDDEN UNIT PARAMETERS An application to SAR data classification

Objectives Improve performance of RBF two- stage learning networks by introducing parameter adaptation criteria where supervised data are exploited to adjust hidden unit parameters (traditionally detected by unsupervised learning techniques)

Topics of discussion n Properties of RBF networks vs MLPs n One- and two-stage RBF network learning strategies n Recent developments in RBF two- stage learning algorithms (Bruzzone, IEEE TGARS, 1999) n Classification examples

RBF network MLP network (hard threshold perceptron)

MLP network n Advantages u distribution-free u importance-free (data fusion) u one-stage supervised learning algorithm (BP) n Disadvantages u slow to train u convergence to a local minimum u high output responses to input data that fall into regions of the input space where there are no training examples (extrapolation) u sensitive to outliers that affect every free parameters u network topology is not data-driven (model selection)

RBF network n Advantages u simple two-layer architecture to solve complex tasks u two-stage hybrid (supervised 1st stage + unsupervised 2nd stage) learning scheme F fast to train F closed form linear optimization of the output weights u localized BFs F low output responses to input data that fall into regions of the input space where there are no training examples F learning of each input sample affects a specialized subset of the network parameters (modularization, outliers do not affect every free parameter) F easy interpretation of the processing units

RBF network n Disadvantages u the classification error strongly depends on the selection of the number, centers and widths of BFs u molteplicity of one-stage (supervised [error-driven]) and two-stage (supervised, hybrid [error-driven + data- driven]) learning algorithms

RBF network learning strategies n One-stage supervised (error-driven) learning u GBF sum-of-squared error gradient descent (Bishop, 1995) F Disadvantages GBFs may not stay localized (    ) no effect on the positions (centers) of GBFs no model selection u new types of BFs suitable for gradient descent learning (Karayiannis, IEEE TNN, 1998) F Disadvantages no model selection u constructive learning (Fritzke, IEEE TNN, 1994) F Disadvantages unlimited growing unstable (small input variations cause large output changes)

RBF network learning strategies n Two-stage learning u hybrid learning (Moody and Darken, 1989) F first stage (hidden layer): data-driven BF centers –clustering BF spread parameters –p-nearest neighbor heuristic F second stage: error-driven gradient descent pseudo-inverse linear optimization (may be unstable) majority voting

RBF network learning strategies n Two-stage learning u hybrid learning (Moody and Darken, 1989) F Disadvantages no model selection (number of BFs) mixed clusters: unsupervised learning does not reflect the local complexity of the classification problem at hand if the number of BFs increases, then there is no guarantee of improving the system’s performance

RBF network learning strategies n Two-stage learning u constructive (Karayiannis, 1998): error-driven location of new BFs F Disadvantages only one unit is inserted per two-stage growing cycle u supervised learning F first stage (hidden layer) BF centers –Karayiannis, 1998: gradient descent of the “localized class- conditional activation variance” (LOCCAV) method –Bruzzone, IEEE TGARS, 1999: class-conditional (constructive) clustering BF spread parameters –Karayiannis: LOCCAV or “localized class-conditional quantization error” (LOCCEQ) –Bruzzone: class-conditional p-nearest neighbor heuristic

RBF network learning strategies n Two-stage learning u supervised learning F first stage (hidden layer) BF spread parameters –Bruzzone: class-conditional p-nearest neighbor heuristic. Given one BF center  i, which belongs to a class, if the 3- nearest BF centers  h,,,  k and  m belong to the same class then less conservative choice:  = (d(  i,  h ) + d(  i,  k )) / 2 otherwise more conservative choice:   LOCCEQ

Bruzzone’s RBF network two- stage learning strategy SAR ERS-1/ERS-2 tandem pair data classification task (IGARSS 2000)

n improvement in classification performance n enhanced performance stability with respect to changes in the number of hidden units n integration of class-conditional data with constructive unsupervised learning techniques to address the model selection issue Conclusions Simple heuristic techniques exploiting supervised data in learning hidden parameters of an RBF two- stage learning network may lead to: