Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.

Slides:



Advertisements
Similar presentations
PARTITIONAL CLUSTERING
Advertisements

Neural Networks with Unsupervised Learning
Un Supervised Learning & Self Organizing Maps. Un Supervised Competitive Learning In Hebbian networks, all neurons can fire at the same time Competitive.
Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Self Organization: Competitive Learning
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
DATA-MINING Artificial Neural Networks Alexey Minin, Jass 2006.
Kohonen Self Organising Maps Michael J. Watts
Artificial neural networks:
Unsupervised Networks Closely related to clustering Do not require target outputs for each input vector in the training data Inputs are connected to a.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
X0 xn w0 wn o Threshold units SOM.
Self Organizing Maps. This presentation is based on: SOM’s are invented by Teuvo Kohonen. They represent multidimensional.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
November 9, 2010Neural Networks Lecture 16: Counterpropagation 1 Unsupervised Learning So far, we have only looked at supervised learning, in which an.
Neural Networks Chapter Feed-Forward Neural Networks.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Lecture 09 Clustering-based Learning
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Introduction to undirected Data Mining: Clustering
Evaluating Performance for Data Mining Techniques
Neural networks - Lecture 111 Recurrent neural networks (II) Time series processing –Networks with delayed input layer –Elman network Cellular networks.
Radial Basis Function Networks
Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Self Organizing Maps (SOM) Unsupervised Learning.
Self Organized Map (SOM)
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Artificial Neural Network Unsupervised Learning
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
NEURAL NETWORKS FOR DATA MINING
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
Machine Learning Neural Networks (3). Understanding Supervised and Unsupervised Learning.
Neural and Evolutionary Computing - Lecture 9 1 Evolutionary Neural Networks Design  Motivation  Evolutionary training  Evolutionary design of the architecture.
Institute for Advanced Studies in Basic Sciences – Zanjan Kohonen Artificial Neural Networks in Analytical Chemistry Mahdi Vasighi.
Ming-Feng Yeh1 CHAPTER 16 AdaptiveResonanceTheory.
UNSUPERVISED LEARNING NETWORKS
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Fast Learning in Networks of Locally-Tuned Processing Units John Moody and Christian J. Darken Yale Computer Science Neural Computation 1, (1989)
Self-Organizing Maps (SOM) (§ 5.5)
CHAPTER 14 Competitive Networks Ming-Feng Yeh.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 15: Mixtures of Experts Geoffrey Hinton.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Fuzzy Pattern Recognition. Overview of Pattern Recognition Pattern Recognition Procedure Feature Extraction Feature Reduction Classification (supervised)
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
Machine Learning 12. Local Models.
Chapter 5 Unsupervised learning
Semi-Supervised Clustering
Data Mining, Neural Network and Genetic Programming
One-layer neural networks Approximation problems
Unsupervised Learning Networks
Clustering (3) Center-based algorithms Fuzzy k-means
Unsupervised Learning and Neural Networks
Dr. Unnikrishnan P.C. Professor, EEE
Lecture 22 Clustering (3).
Neural Networks and Their Application in the Fields of Coporate Finance By Eric Séverin Hanna Viinikainen.
Competitive Networks.
Competitive Networks.
Text Categorization Berlin Chen 2003 Reference:
Feature mapping: Self-organizing Maps
Artificial Neural Networks
Unsupervised Networks Closely related to clustering
Presentation transcript:

Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering Vectorial quantization Topological mapping and Kohonen networks

Neural Networks - Lecture 82 Particularities of unsupervised learning Supervised learning The training set contains both inputs and correct answers Example: classification in predefined classes for which one knows examples of labeled data It is similar with the optimization of an error function which measures the difference between the true answers and the answers given by the network Unsupervised learning: The training set contains just input data Example: grouping data in categories based on the similarities between them Relies on the statistical properties of data Does not use an error concept but a clustering quality concept which should be maximized

Neural Networks - Lecture 83 Data clustering Data clustering = identifying natural groups (clusters) in the data set such that –Data in the same cluster are highly similar –Data belonging to different clusters are dissimilar enough Rmk: It does not exist an apriori labeling of the data (unlike supervised classification) and sometimes even the number of clusters is unknown Applications: Identification of user profiles in the case of e-commerce systems, e-learning systems etc.) Identification of homogeneous regions in digital images (image segmentation) Electronic document categorization Biological data analysis

Neural Networks - Lecture 84 Data clustering Example: image segmentation = identification the homogeneous regions in the image = reduction of the number of labels (colours) in the image in order to help the image analysis Rmk: satellite image obtained by combining three spectral bands

Neural Networks - Lecture 85 Data clustering A key element in data clustering is the similarity/dissimilarity measure The choice of an appropriate similarity/dissimilarity measure is influenced by the nature of attributes Numerical attributes Categorical attributes Mixed attributes

Neural Networks - Lecture 86 Data clustering Measures adequate to numerical data

Neural Networks - Lecture 87 Data clustering Clustering methods: Partitional methods: –Lead to a data partitions in disjoint or partially overlapping clusters –Each cluster is characterized by one or multiple prototypes Hierarchical methods: –Lead to a hierarchy of partitions which can be visualized as a tree-like structure called dendrogram.

Neural Networks - Lecture 88 Data clustering The prototypes can be: The average of data in the class The median of data in the class The data are assigned to clusters based on the nearest neighbour prototype

Neural Networks - Lecture 89 Neural networks for clustering Problem: unsupervised classification of N-dimensional data in K clusters Architecture: One input layer with N units Un linear layer with K output units Fully connectivity between the input and the output layers (the weights matrix, W, contains on row k the prototype corresponding to class k) Functioning: For an input data X compute the distances between X and all prototypes and find the closest one. This corresponds to the cluster to which X belongs. The unit having the closest weight vector to X is called winning unit W

Neural Networks - Lecture 810 Neural networks for clustering Training: Training set: {X 1,X 2,…, X L } Algorithms: the number of clusters (output units) is known – “Winner Takes All” algorithm (WTA) the number of clusters (output units) is unknown –“Adaptive Resonance Theory” (ART algorithms) Particularities of these algorithms: Only the weights corresponding to the winning unit are adjusted The learning rate is decreasing

Neural Networks - Lecture 811 Neural networks for clustering Examplu: WTA algorithm

Neural Networks - Lecture 812 Neural networks for clustering Remarks: Decreasing learning rate

Neural Networks - Lecture 813 Neural networks for clustering Remarks: “Dead” units: units which are never winners Cause: inappropriate initialization of prototypes Solutions: Using vectors from the training set as initial prototypes Adjusting not only the winning units but also the other units (using a much smaller learning rate) Penalizing the units which frequently become winners

Neural Networks - Lecture 814 Neural networks for clustering Penalizing the units which frequently become winners: change the criterion to establish if a unit is a winner or not

Neural Networks - Lecture 815 Neural networks for clustering It is useful to normalize both the input data and the weights (prototypes): The normalization of data from the training set is realized before the training The weights are normalized during the training:

Neural Networks - Lecture 816 Neural networks for clustering Adaptive Resonance Theory: gives solutions for the following problems arising in the design of unsupervised classification systems: Adaptability –Refers to the capacity of the system to assimilate new data and to identify new clusters (this usually means a variable number of clusters) Stability –Refers to the capacity of the system to conserve the clusters structures such that during the adaptation process the system does not radically change its output for a given input data

Neural Networks - Lecture 817 Neural networks for clustering Example: ART2 algorithm

Neural Networks - Lecture 818 Neural networks for clustering Remarks: The value of rho influences the number of output units (clusters) –A small value of ρ leads to a large number of clusters –A large value of ρ leads to a small number of clusters Main drawback: the presentation order of the training data influences the training process The main difference between this algorithm and that used to find the centers of a RBF network is represented just by the adjustment eq. There are also versions for binary data (alg ART1)

Neural Networks - Lecture 819 Vectorial quantization Aim of vectorial quantization: Mapping a region of R N to a finite set of prototypes Allows the partitioning of a N-dimensional region in a finite number of subregions such that each subregion is identified by a prototype The cuantization process allows to replace a N-dimensional vector with the index of the region which contains it leading to a very simple compression method with a loss of information. The aim is to minimize the loss of information The number of prototypes should be large in the dense regions and small in less dense regions

Neural Networks - Lecture 820 Vectorial quantization Example

Neural Networks - Lecture 821 Vectorial quantization If the number of regions is predefined then one can use the WTA algorithm There is also a supervised variant of vectorial quantization (LVQ - Learning Vector Quantization) Training set: {(X1,d1),…,(XL,dL)} LVQ algorithm: 1.Initialize the prototypes by applying a WTA algorithm to the set {X1,…,XL} 2.Identify the clusters based on the nearest neighbour criterion 3.Establish the label for each class by using the labels from the training set: for each class is assigned the most frequent label from d1,…dL

Neural Networks - Lecture 822 Vectorial quantization Iteratively adjust the prototypes by applying an algorithm similar to that of perceptron. At each iteration one applies

Neural Networks - Lecture 823 Topological mapping This is a variant of vectorial quantization which ensures a preservation of the neighbourhood relationship: –Similar input data either belongs to the same cluster or to neighboring clusters –There is necessary to define a neighborhood relation on the prototypes and on the output units of the network –The networks for topological mapping have the output units placed on the nodes of a geometrical grid (one, bi or three dimensional) –Networks having such an architecture are called Kohonen networks

Neural Networks - Lecture 824 Kohonen networks Architecture Two inputs and a bidimensional grid Square grid Hexagonal grid

Neural Networks - Lecture 825 Kohonen networks Functioning: similar to all competitive networks – by using the nearest neighbor criterion one finds the winning unit In order to define a topology for the output level one needs : –A position vector for each output unit –A distance between position vectors

Neural Networks - Lecture 826 Kohonen networks The neighbourhood of radius s of unit p is: Example: for a squared grid the neighborhood of radius 1 defined by the previous three distances are: