Institute for Advanced Studies in Basic Sciences – Zanjan Mahdi Vasighi Supervised Kohonen Artificial Neural Networks.

Slides:



Advertisements
Similar presentations
Memristor in Learning Neural Networks
Advertisements

2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Unsupervised learning. Summary from last week We explained what local minima are, and described ways of escaping them. We investigated how the backpropagation.
Self Organization: Competitive Learning
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Kohonen Self Organising Maps Michael J. Watts
Artificial neural networks:
Unsupervised Networks Closely related to clustering Do not require target outputs for each input vector in the training data Inputs are connected to a.
Artificial Neural Networks
X0 xn w0 wn o Threshold units SOM.
Decision Support Systems
Simple Neural Nets For Pattern Classification
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation.
Un Supervised Learning & Self Organizing Maps Learning From Examples
November 9, 2010Neural Networks Lecture 16: Counterpropagation 1 Unsupervised Learning So far, we have only looked at supervised learning, in which an.
Introduction to Neural Networks Simon Durrant Quantitative Methods December 15th.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
October 7, 2010Neural Networks Lecture 10: Setting Backpropagation Parameters 1 Creating Data Representations On the other hand, sets of orthogonal vectors.
Neural Networks Lecture 17: Self-Organizing Maps
Radial Basis Function (RBF) Networks
Projection methods in chemistry Autumn 2011 By: Atefe Malek.khatabi M. Daszykowski, B. Walczak, D.L. Massart* Chemometrics and Intelligent Laboratory.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Microarray Gene Expression Data Analysis A.Venkatesh CBBL Functional Genomics Chapter: 07.
Principle Component Analysis (PCA) Networks (§ 5.8) PCA: a statistical procedure –Reduce dimensionality of input vectors Too many features, some of them.
KOHONEN SELF ORGANISING MAP SEMINAR BY M.V.MAHENDRAN., Reg no: III SEM, M.E., Control And Instrumentation Engg.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Self Organized Map (SOM)
Intrusion Detection Using Hybrid Neural Networks Vishal Sevani ( )
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Chapter 9 Neural Network.
Artificial Neural Network Unsupervised Learning
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Applying Neural Networks Michael J. Watts
Institute for Advanced Studies in Basic Sciences – Zanjan Kohonen Artificial Neural Networks in Analytical Chemistry Mahdi Vasighi.
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
381 Self Organization Map Learning without Examples.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Unsupervised Learning Networks 主講人 : 虞台文. Content Introduction Important Unsupervised Learning NNs – Hamming Networks – Kohonen’s Self-Organizing Feature.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A self-organizing map for adaptive processing of structured.
CHAPTER 14 Competitive Networks Ming-Feng Yeh.
Perceptrons Michael J. Watts
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
Artificial Neural Networks for Data Mining. Copyright © 2011 Pearson Education, Inc. Publishing as Prentice Hall 6-2 Learning Objectives Understand the.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Self-Organizing Network Model (SOM) Session 11
Data Mining, Neural Network and Genetic Programming
Unsupervised Learning Networks
Self organizing networks
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
Lecture 22 Clustering (3).
The Naïve Bayes (NB) Classifier
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Machine Learning – a Probabilistic Perspective
Artificial Neural Networks
Unsupervised Networks Closely related to clustering
Lecture 16. Classification (II): Practical Considerations
Patterson: Chap 1 A Review of Machine Learning
Presentation transcript:

Institute for Advanced Studies in Basic Sciences – Zanjan Mahdi Vasighi Supervised Kohonen Artificial Neural Networks

Data Analysis: Linear transformations and data reduction techniques like Principal Component Analysis (PCA) Advantage: 1. Projection from high dim. onto a low dim. coordinate system. 2. Does not require a high level of modelling expertise. Disadvantages: 1. Assuming the topology of the input data can be reduced in a linear fashion. 2. Outliers disturb the quality of the projection. 3. Visualization power deteriorates considerably if the number of relevant dimensions in the multivariate space remains high after a PCA analysis. Introduction

An alternative is nonlinear mapping: This explicitly aims to map objects in a low- dimensional space (usually two-dimensional) in such a way that the distances between objects are preserved in the mapping. The Kohonen maps (self organizing maps) incorporates in an unsupervised way the topology present in the data. The unsupervised problem means that one deals with a set of experimental data which have no specific associated answers (or supplemental information) attached.

A supervised modelling technique is able to capture the relationship between the input data (measurements, observations) and output data (properties of interest).  MLR, PLS, for regression problems  LDA for classification problems These modelling techniques fail if:  Nonlinearity or topologically incoherent relationship and Considerable number of outliers. ANNs and SVMs can tackle such nonlinear relationships in a convincing way. However, the visualization and interpretation of these models is severely hard due to the fact that they are more or less ‘black-box’ techniques. X (input) Y (output)

Kohonen Artificial Neural Networks The Kohonen network is probably the closest of all artificial neural networks architectures and learning schemes to the biological neuron network As a rule, the Kohonen type of net is based on a single layer of neurons arranged in a two- dimensional plane having a well defined topology A defined topology means that each neuron has a defined number of neurons as nearest neighbors, second-nearest neighbor, etc.

W Input vector Output Weight vector Similarity is the basis of selection of the winner neuron. In other words, there is a competition between neurons for winning. (competitive learning)

The Kohonen learning concept tries to map the input so that similar signals excite neurons that are very close together.

Top Map After the training process accomplished, the complete set of the training vectors is once more run through the KANN. In this last run the labeling of the neurons excited by the input vector is made into the table called top map. e XSXS d c b a Trained KANN e db c a Top Map

Weight Map The number of weights in each neuron is equal to the dimension m of the input vector. Hence, in each level of weight only data of one specific variable are handled. Trained KANN X S Input Vector LLLL LLL H HHHH HHH Top Map

Applications of Kohonen ANNs Sampling from large amount of data Disease Diagnosis QSAR & QSPR Analysis of genomics data Making model using corrupt object (missing values) Typical similarity criteria Specific similarity criteria Number of missing values

Counter Propagation network (CPN) CPANN has the same structure as Kohonen network with an additional output layer with same layout as input layer.

Output layer Input Kohonen layer Input Target Input Similarity map Winner Based on the location of the winning unit in the input map (i.e., the unit which is most similar or closest to the presented object X), the input map and the output map are updated simultaneously at the same spatial locations. If the CPN network is trained, it can be used for prediction. Simply, an unknown input object is presented to the network. The position of the winning unit in the input map then is used to look-up the predicted value of the corresponding unit in the output map.

Applications of CP-ANN Building predictive calibration and classification models Spectrum Class Membership Molecular Descriptor Medicinal Activity Process Descriptor Process Condition Reaction Descriptor Reaction Mechanism

Detection of faulty condition in a process

CPN was not able to model the inverse relationship between the output and the input.  The output properties are not involved in the formation of the directing Kohonen input map. Hence, the CPN model cannot be considered as being a true supervised method.  One can easily look inside the driving input layer and relationship with output. in a CPN network the flow of information is directed from the input layer units towards the output layer. For this reason, we prefer to denote the CPN as being a pseudo- supervised strategy.

Supervised Kohonen network (SKN) In a SKN network, the input layer and the output layer are ‘glued’ together, thereby forming a combined input-output layer. Because in a SKN information present in the objects X and Y is used explicitly during the update of the units in the map, the topological formation of the concatenated map is driven by X and Y in a truly supervised way.

After training, the input and output maps are decoupled. Then, for a new input object its class membership is estimated according to the procedure outlined for the CPN network. The variables of the objects X and Y in the training set must be scaled properly, but it is not trivial how to deal with the relative weight of the number of variables in X and the number of variables in Y during the training of a SKN network.  User must determine beforehand the proper balance between the influence of the input and output objects: in general, correct scaling of the input and output variables is of utmost importance.

Output layer Input Kohonen layer Input Target Output Similarity map Input Similarity map Fused Similarity map Winner By using a ‘fused’ similarity measure based on a weighted combination of the similarities between an object X and all units in the input layer, and the similarities between the corresponding target and the units in the output layer, the common winning unit for both maps is determined. α(t) decreases linearly in epochs Chemo. Int. lab. 83 (2006) 99–113 XY-Fused networks

Simulated data sets The first data set, referred to as Odd–Even, contains 400 rows (input objects) with 8 variables: each integer valued variable was varied randomly between 1 and 100. Odd–Even data set Data Matrix (8  400) Class (1  400) if per row the total number of even values was greater than the total number of odd values, the class membership for that row was assigned to be 1, otherwise the class membership was set to −1

Output map by C CC CPNN for Odd-Even data Output layer Input layer Input Class CPNN  Scattered pattern of output map  No real relationship between the multivariate topological structure present in the input space and the associated class membership of the objects.  Does not take into account the class information during the formation of the input and output maps.

Output map by S SS SKNN for Odd-Even data Input ClassSKNN Kohonen layer  Scattered pattern of output map  Imbalance between the number of input (8) and output (1) variables  Does not take into account the class information during the formation of the input and output maps.

Output layer Output Similarity map Fused Similarity map Input Similarity map Input Class Input Kohonen layer X-Y Fused Network Output map by X XX XYFN for Odd-Even data  Nice coherent output maps  Indicating a certain ‘hidden’ relationship present between the input and output space.

Class 1 Class 2 Class 3 Set 1 X Y Z Class 1 Data Matrix (3  450) Set 2Set 3 Class 2 Class Three normally distributed clouds of data points in three dimensions. the first 150 objects belong to a multivariate normal distribution around the origin (class 3), whereas the other two classes 1 and 2, each consisting of 150 objects as well, are normally distributed around the centroids (5,3,4) and (5.5, 3.5, 4.5). Overlap data set Input layer Output layer We will compares the quality of the 8×8 input and output maps for the CPN and XYF networks for the Overlap data set.

Xmap Ymap Class map CPNN Input layer Output layer

XY-Fused Network for Overlap data set Class map Ymap Xmap

Neural Networks For Chemists, An Introduction. (Weinheim/VCH Publishers ) Chem.Int.Lab.sys. 83 (2006) 99–113 Chem.Int.Lab.sys 90 (2008) 84–91 Chem.Int.Lab.sys. 38 (1997) 1-23 Current Computer-Aided Drug Design, 2005, 1, References

Thanks

×4× Input vector output Winner

Input vector × 0.9× 0.8×0.9× 0.6×0.9× × 0.4×0.9 a max =0.9 a min =0.1 t=1 (first epoch) Neighbor function: Linear winner  d