Self-organizing map Speech and Image Processing Unit Department of Computer Science University of Joensuu, FINLAND Pasi Fränti Clustering Methods: Part.

Slides:



Advertisements
Similar presentations
Sheeraza Bandi Sheeraza Bandi CS-635 Advanced Machine Learning Zahid Irfan February 2004 Picture © Greg Martin,
Advertisements

Variable Metric For Binary Vector Quantization UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE JOENSUU, FINLAND Ismo Kärkkäinen and Pasi Fränti.
2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
November 18, 2010Neural Networks Lecture 18: Applications of SOMs 1 Assignment #3 Question 2 Regarding your cascade correlation projects, here are a few.
Un Supervised Learning & Self Organizing Maps. Un Supervised Competitive Learning In Hebbian networks, all neurons can fire at the same time Competitive.
Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Self Organization: Competitive Learning
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
DATA-MINING Artificial Neural Networks Alexey Minin, Jass 2006.
Computational Intelligence: Methods and Applications Lecture 10 SOM and Growing Cell Structures Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Self-Organizing Map (SOM). Unsupervised neural networks, equivalent to clustering. Two layers – input and output – The input layer represents the input.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
X0 xn w0 wn o Threshold units SOM.
Competitive Networks. Outline Hamming Network.
Self-Organizing Hierarchical Neural Network
CONTENT BASED FACE RECOGNITION Ankur Jain 01D05007 Pranshu Sharma Prashant Baronia 01D05005 Swapnil Zarekar 01D05001 Under the guidance of Prof.
1 Kunstmatige Intelligentie / RuG KI2 - 7 Clustering Algorithms Johan Everts.
Neural Network Homework Report: Clustering of the Self-Organizing Map Professor : Hahn-Ming Lee Student : Hsin-Chung Chen M IEEE TRANSACTIONS ON.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
November 24, 2009Introduction to Cognitive Science Lecture 21: Self-Organizing Maps 1 Self-Organizing Maps (Kohonen Maps) In the BPN, we used supervised.
Neural Networks Lecture 17: Self-Organizing Maps
A Hybrid Self-Organizing Neural Gas Network James Graham and Janusz Starzyk School of EECS, Ohio University Stocker Center, Athens, OH USA IEEE World.
Lecture 09 Clustering-based Learning
Introduction to undirected Data Mining: Clustering
 C. C. Hung, H. Ijaz, E. Jung, and B.-C. Kuo # School of Computing and Software Engineering Southern Polytechnic State University, Marietta, Georgia USA.
Project reminder Deadline: Monday :00 Prepare 10 minutes long pesentation (in Czech/Slovak), which you’ll present on Wednesday during.
Clustering Methods: Part 2d Pasi Fränti Speech & Image Processing Unit School of Computing University of Eastern Finland Joensuu, FINLAND Swap-based.
Lecture 12 Self-organizing maps of Kohonen RBF-networks
Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Self Organized Map (SOM)
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
Self-organizing Maps Kevin Pang. Goal Research SOMs Research SOMs Create an introductory tutorial on the algorithm Create an introductory tutorial on.
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Cut-based & divisive clustering Clustering algorithms: Part 2b Pasi Fränti Speech & Image Processing Unit School of Computing University of Eastern.
Community Architectures for Network Information Systems
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
Self-Organizing Maps Corby Ziesman March 21, 2007.
Genetic Algorithm Using Iterative Shrinking for Solving Clustering Problems UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE FINLAND Pasi Fränti and.
Machine Learning Neural Networks (3). Understanding Supervised and Unsupervised Learning.
Self Organizing Feature Map CS570 인공지능 이대성 Computer Science KAIST.
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
UNSUPERVISED LEARNING NETWORKS
Genetic algorithms (GA) for clustering Pasi Fränti Clustering Methods: Part 2e Speech and Image Processing Unit School of Computing University of Eastern.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Rival-Model Penalized Self-Organizing Map Yiu-ming Cheung.
381 Self Organization Map Learning without Examples.
Unsupervised Learning Networks 主講人 : 虞台文. Content Introduction Important Unsupervised Learning NNs – Hamming Networks – Kohonen’s Self-Organizing Feature.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Fast Learning in Networks of Locally-Tuned Processing Units John Moody and Christian J. Darken Yale Computer Science Neural Computation 1, (1989)
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Adaptive FIR Neural Model for Centroid Learning in Self-Organizing.
Self-Organizing Maps (SOM) (§ 5.5)
CHAPTER 14 Competitive Networks Ming-Feng Yeh.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Multilevel thresholding by fast PNN based algorithm UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE FINLAND Olli Virmajoki and Pasi Fränti.
Genetic Algorithms for clustering problem Pasi Fränti
Computational Intelligence: Methods and Applications Lecture 9 Self-Organized Mappings Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Chapter 5 Unsupervised learning
Self-Organizing Network Model (SOM) Session 11
Data Mining, Neural Network and Genetic Programming
Self Organizing Maps: Parametrization of Parton Distribution Functions
Unsupervised Learning and Neural Networks
Lecture 22 Clustering (3).
Self-Organizing Maps Corby Ziesman March 21, 2007.
Competitive Networks.
Competitive Networks.
Self-Organizing Maps (SOM) (§ 5.5)
Pasi Fränti and Sami Sieranoja
Introduction to Cluster Analysis
Artificial Neural Networks
Presentation transcript:

Self-organizing map Speech and Image Processing Unit Department of Computer Science University of Joensuu, FINLAND Pasi Fränti Clustering Methods: Part 9

SOM main principles Self-organizing map (SOM) is a clustering method suitable especially for visualization. Clustering represented by centroids organized in a 1-d or 2-d network. Dimensionality reduction and visualization possibility achieved as side product. Clustering performed by competitive learning principle.

Self-organizing map Initial configuration M nodes, one for each cluster Initial locations not important Nodes connected by network topology (1-d or 2-d)

Self-organizing map Final configuration Node locations adapt during learning stage Network keeps neighbor vectors close to each other Network limits the movement of vectors during learning

SOM pseudo code (1/2) Learning stage

SOM pseudo code Update centroids (2/2)

Competitive learning Each data vector is processed once. Find nearest centroid: The centroid is updated by moving it towards the data vector by: Learning stage similar to k-means but centroid update has different principle.

Decreases with time  movement is large in the beginning but eventually stabilizes. Linear decrease of weighting: Learning rate (  ) Exponential decrease of weighting:

Neighborhood (d) Neighboring centroids are also updated: Effect is stronger for nearby centroids:

Weighting of the neighborhood Weighting decreases exponentially

Number of iterations T –Convergence of SOM is rather slow  Should be set as high as possible –Roughly iterations at minimum. Size of the initial neighborhood D max –Small enough to allow local adaption. –Value D=0 indicates no neighbor structure Maximum learning rate A –Higher values have mostly random effect. –Most critical are the final stages (D  2) –Optimal choices of A and D max highly correlated. Parameter setup

Difficulty of parameter setup Fixing total number of iterations (T  D max ) to 20, 40 and 80. Optimal parameter combination non-trivial.

To reduce the effect of parameter set- up, should be as high as possible. Enough time to adapt at the cost of high time complexity. Adaptive number of iterations: Adaptive number of iterations For D max =10 and T max =100: Ti = {1, 1, 1, 1, 2, 3, 6, 13, 25, 50, 100}

Example of SOM (1-d) One cluster too many One cluster missing

Example of SOM (2-d) (to appear sometime in future)

1.T. Kohonen, Self-Organization and Associative Memory. Springer- Verlag, New York, N.M. Nasrabadi and Y. Feng, "Vector quantization of images based upon the Kohonen self-organization feature maps", Neural Networks, 1 (1), 518, P. Fränti, "On the usefulness of self-organizing maps for the clustering problem in vector quantization", 11th Scandinavian Conf. on Image Analysis (SCIA’99), Kangerlussuaq, Greenland, vol. 1, , Literature