Efficient SOM Learning by Data Order Adjustment

Slides:



Advertisements
Similar presentations
Gate Sizing for Cell Library Based Designs Shiyan Hu*, Mahesh Ketkar**, Jiang Hu* *Dept of ECE, Texas A&M University **Intel Corporation.
Advertisements

2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Organizing a spectral image database by using Self-Organizing Maps Research Seminar Oili Kohonen.
Intelligent Database Systems Lab Advisor : Dr.Hsu Graduate : Keng-Wei Chang Author : Gianfranco Chicco, Roberto Napoli Federico Piglione, Petru Postolache.
Graduate : Sheng-Hsuan Wang
X0 xn w0 wn o Threshold units SOM.
Self Organizing Maps. This presentation is based on: SOM’s are invented by Teuvo Kohonen. They represent multidimensional.
1 Module 2: Fundamental Concepts Problems Programs –Programming languages.
Lecture 2: Fundamental Concepts
1 Module 2: Fundamental Concepts Problems Programs –Programming languages.
On the Construction of Energy- Efficient Broadcast Tree with Hitch-hiking in Wireless Networks Source: 2004 International Performance Computing and Communications.
1 Kunstmatige Intelligentie / RuG KI2 - 7 Clustering Algorithms Johan Everts.
A Hybrid Self-Organizing Neural Gas Network James Graham and Janusz Starzyk School of EECS, Ohio University Stocker Center, Athens, OH USA IEEE World.
CoNA : Dynamic Application Mapping for Congestion Reduction in Many-Core Systems 2012 IEEE 30th International Conference on Computer Design (ICCD) M. Fattah,
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
Efficient Model Selection for Support Vector Machines
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Student : Sheng-Hsuan Wang Department.
-Artificial Neural Network- Chapter 9 Self Organization Map(SOM) 朝陽科技大學 資訊管理系 李麗華 教授.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A novel genetic algorithm for automatic clustering Advisor.
An Efficient Approach to Clustering in Large Multimedia Databases with Noise Alexander Hinneburg and Daniel A. Keim.
2002/4/10IDSL seminar Estimating Business Targets Advisor: Dr. Hsu Graduate: Yung-Chu Lin Data Source: Datta et al., KDD01, pp
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A Comparison of SOM Based Document Categorization Systems.
Clustering Spatial Data Using Random Walks Author : David Harel Yehuda Koren Graduate : Chien-Ming Hsiao.
Self Organization of a Massive Document Collection Advisor : Dr. Hsu Graduate : Sheng-Hsuan Wang Author : Teuvo Kohonen et al.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Topology Preservation in Self-Organizing Feature Maps: Exact.
Color Image Segmentation Speaker: Deng Huipeng 25th Oct , 2007.
A genetic approach to the automatic clustering problem Author : Lin Yu Tseng Shiueng Bien Yang Graduate : Chien-Ming Hsiao.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 New Unsupervised Clustering Algorithm for Large Datasets.
A New Similarity Measure of Generalized Fuzzy Numbers Based on Geometric-Mean Averaging Operator Author: Shi-Jay Chen Speaker: Shih-Hua Wei 2006 IEEE International.
Identifying Patterns in Time Series Data Daniel Lewis 04/06/06.
Hierarchical Clustering of Gene Expression Data Author : Feng Luo, Kun Tang Latifur Khan Graduate : Chien-Ming Hsiao.
SimRank : A Measure of Structural-Context Similarity
The Supervised Network Self- Organizing Map for Classification of Large Data Sets Authors: Papadimitriou et al, Advisor: Dr. Hsu Graduate: Yu-Wei Su.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 The Evolving Tree — Analysis and Applications Advisor.
1 On Optimal Worst-Case Matching Cheng Long (Hong Kong University of Science and Technology) Raymond Chi-Wing Wong (Hong Kong University of Science and.
August 30, 2004STDBM 2004 at Toronto Extracting Mobility Statistics from Indexed Spatio-Temporal Datasets Yoshiharu Ishikawa Yuichi Tsukamoto Hiroyuki.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A modified version of the K-means algorithm with a distance.
1. Find the instantaneous velocity for the following 1. Find the instantaneous velocity for the following time—distance function at time = 2 seconds. time—distance.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Graduate : Sheng-Hsuan Wang Authors :
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Rival-Model Penalized Self-Organizing Map Yiu-ming Cheung.
CUNY Graduate Center December 15 Erdal Kose. Outlines Define SOMs Application Areas Structure Of SOMs (Basic Algorithm) Learning Algorithm Simulation.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology O( ㏒ 2 M) Self-Organizing Map Algorithm Without Learning.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A self-organizing map for adaptive processing of structured.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Adaptive FIR Neural Model for Centroid Learning in Self-Organizing.
Joint Replication-Migration-based Routing in Delay Tolerant Networks Yunsheng Wang and Jie Wu Temple University Zhen Jiang Feng Li West Chester Unveristy.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A hierarchical clustering algorithm for categorical sequence.
Image Processing A Study in Pixel Averaging Building a Resolution Pyramid With Parallel Computing Denise Runnels and Farnaz Zand.
Lloyd Algorithm K-Means Clustering. Gene Expression Susumu Ohno: whole genome duplications The expression of genes can be measured over time. Identifying.
Efficient Rule-Based Attribute-Oriented Induction for Data Mining Authors: Cheung et al. Graduate: Yu-Wei Su Advisor: Dr. Hsu.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Growing Hierarchical Tree SOM: An unsupervised neural.
Intelligent Database Systems Lab Advisor : Dr. Hsu Graduate : Yu Cheng Chen Author : Yongqiang Cao Jianhong Wu 國立雲林科技大學 National Yunlin University of Science.
ViSOM - A Novel Method for Multivariate Data Projection and Structure Visualization Advisor : Dr. Hsu Graduate : Sheng-Hsuan Wang Author : Hujun Yin.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Graduate : Chun Kai Chen Author : Andrew.
Intelligent Database Systems Lab Advisor : Dr. Hsu Graduate : Jian-Lin Kuo Author : Aristidis Likas Nikos Vlassis Jakob J.Verbeek 國立雲林科技大學 National Yunlin.
Mining Top-n Local Outliers in Large Databases Author: Wen Jin, Anthony K. H. Tung, Jiawei Han Advisor: Dr. Hsu Graduate: Chia- Hsien Wu.
Neural network based hybrid computing model for wind speed prediction K. Gnana Sheela, S.N. Deepa Neurocomputing Volume 122, 25 December 2013, Pages 425–429.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Graduate : Yu Cheng Chen Author: Michael.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Intelligent Exploration for Genetic Algorithms Using Self-Organizing.
Self-Organizing Network Model (SOM) Session 11
CACTUS-Clustering Categorical Data Using Summaries
Data Mining, Neural Network and Genetic Programming
A modified hyperplane clustering algorithm allows for efficient and accurate clustering of extremely large datasets Ashok Sharma, Robert Podolsky, Jieping.
KMeans Clustering on Hadoop Fall 2013 Elke A. Rundensteiner
Self-organizing map numeric vectors and sequence motifs
A connectionist model in action
Donghui Zhang, Tian Xia Northeastern University
Presentation transcript:

Efficient SOM Learning by Data Order Adjustment Authors: Miyoshi et al. Advisor: Dr. Hsu Graduate :Yu-Wei Su

Outline Motivation Objective SOM Data order and learning convergence Ending point of convergence Experiments Conclusion opinion

Motivation In SOM, there are many factors to aggravate computational load and competition

Objective Reducing the competition load and increasing the speed of SOM

SOM SOM is a algorithm that map dimension from high to low, always two dimension Step 1: finding BMU of each datum Step 2: modifying the value of BMU and neighborhood nodes

Data order and learning convergence The SOM spends lots of time to learn because of large map size, large quantity of input data and many dimensions in data et al. In the beginning stage of learning process, SOM map is dynamically and widely and that is depended on the distance of each input data

Data order and learning convergence( cont.) Adjusting data order based on the distance between data classes to reduce the competition load

Data order To change order of input data, using class distance that is calculated by class center First select typical data as class center in each class And calculate Euclidian distance between all class centers as class distance

Data order( cont.) Order 1: random order Order 2: the largest distance order based on previous data class scli : selected data class i ucli : still unselected data class i cd(A,B) :class distance between A and B

Data order( cont.) Order 3: the smallest distance order based on previous data class Order 4: the smallest distance order based on all classes scli : selected data class i ucli : still unselected data class i cd(A,B) :class distance between A and B

Data order( cont.) Order 5: the largest distance order based on all classes Order 6: average distance order based on all classes scli : selected data class i ucli : still unselected data class i cd(A,B) :class distance between A and B

Ending point of convergence Definition of converging point of learning Keep maximum Euclidian distance for all nodes in each step of learning Test the difference between (|xdn-xdn-1|) whether (|xdn-xdn-1|) is smaller than Th1 If it is, test how long the distance are continued smaller If it continues long enough than Th2 it is determined as the ending point of learning

Ending point of convergence (cont.) Th1 : threshold of difference Th2 : threshold of period xdn : n-th max distance through all input data and output nodes ed(A,B) : Euclidian distance between A and B dti : input data I onj: output node j dmax : total of input data nmax : total of output nodes

Experiments Experiment data Parameters Synthetic data, 5 dim, 7 classes each 49 data Parameters Size of map 8x8, initial neighborhood from 3 to 5, initial learning rate from 0.2 to 0.8, 300 maps that initialized at random

Experiments ( cont.) Learning rate function Neighborhood function

Experiments ( cont.)

Experiments ( cont.)

Conclusion The data stream of small distance makes maximum 9% improvement The data stream of large distance still similar with conventional SOM All order make no remarkable difference in result map

Opinion No experiments of comparison with others The terminal condition is a good idea