Reporter: Ming-Jui Kuo D

Slides:



Advertisements
Similar presentations
K-MEANS Michael Jones ENEE698Q Fall Overview  Introduction  Problem Formulation  How K-Means Works  Pros and Cons of Using K-Means  How to.
Advertisements

Fast Algorithms For Hierarchical Range Histogram Constructions
INTELLIGENT EDITOR FOR ANDROID MOBILES PHASE 1 : HANDWRITING RECOGNITION ADVANCED MOBILE SYSTEMS ENGINEERING RESEARCH PROJECT BY NITYATA N KUMAR AND AASHRAY.
Unsupervised learning. Summary from last week We explained what local minima are, and described ways of escaping them. We investigated how the backpropagation.
Kohonen Self Organising Maps Michael J. Watts
Procedures of Extending the Alphabet for the PPM Algorithm Radu Rădescu George Liculescu Polytechnic University of Bucharest Faculty of Electronics, Telecommunications.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Complexity 7-1 Complexity Andrei Bulatov Complexity of Problems.
Self Organizing Maps. This presentation is based on: SOM’s are invented by Teuvo Kohonen. They represent multidimensional.
CPSC 411, Fall 2008: Set 12 1 CPSC 411 Design and Analysis of Algorithms Set 12: Undecidability Prof. Jennifer Welch Fall 2008.
UNIVERSITY OF JYVÄSKYLÄ Yevgeniy Ivanchenko Yevgeniy Ivanchenko University of Jyväskylä
Complexity 5-1 Complexity Andrei Bulatov Complexity of Problems.
Tracking a moving object with real-time obstacle avoidance Chung-Hao Chen, Chang Cheng, David Page, Andreas Koschan and Mongi Abidi Imaging, Robotics and.
Aho-Corasick String Matching An Efficient String Matching.
Programming Logic and Design, Introductory, Fourth Edition1 Understanding Computer Components and Operations (continued) A program must be free of syntax.
Lecture 09 Clustering-based Learning
Cmpt-225 Simulation. Application: Simulation Simulation  A technique for modeling the behavior of both natural and human-made systems  Goal Generate.
Copyright © Cengage Learning. All rights reserved.
Super-Orthogonal Space- Time BPSK Trellis Code Design for 4 Transmit Antennas in Fast Fading Channels Asli Birol Yildiz Technical University,Istanbul,Turkey.
A Shaft Sensorless Control for PMSM Using Direct Neural Network Adaptive Observer Authors: Guo Qingding Luo Ruifu Wang Limei IEEE IECON 22 nd International.
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A Comparison of SOM Based Document Categorization Systems.
Chih-Ming Chen, Student Member, IEEE, Ying-ping Chen, Member, IEEE, Tzu-Ching Shen, and John K. Zao, Senior Member, IEEE Evolutionary Computation (CEC),
Machine Learning Using Support Vector Machines (Paper Review) Presented to: Prof. Dr. Mohamed Batouche Prepared By: Asma B. Al-Saleh Amani A. Al-Ajlan.
On Distinguishing the Multiple Radio Paths in RSS-based Ranging Dian Zhang, Yunhuai Liu, Xiaonan Guo, Min Gao and Lionel M. Ni College of Software, Shenzhen.
1 Pattern Matching Using n-gram Sampling Of Cumulative Algebraic Signatures : Preliminary Results Witold Litwin[1], Riad Mokadem1, Philippe Rigaux1 & Thomas.
Self Organization of a Massive Document Collection Advisor : Dr. Hsu Graduate : Sheng-Hsuan Wang Author : Teuvo Kohonen et al.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Visualizing Ontology Components through Self-Organizing.
Fast and accurate text classification via multiple linear discriminant projections Soumen Chakrabarti Shourya Roy Mahesh Soundalgekar IIT Bombay
Fundamentals of Algorithms MCS - 2 Lecture # 8. Growth of Functions.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. Externally growing self-organizing maps and its application to database visualization and exploration.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Bell Ringer  1. What is science and how is it related to technology?  2. List 3 forms of technology used in your home and describe how tasks were completed.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Extensions of vector quantization for incremental clustering.
CUNY Graduate Center December 15 Erdal Kose. Outlines Define SOMs Application Areas Structure Of SOMs (Basic Algorithm) Learning Algorithm Simulation.
CHAPTER 1: Introduction. 2 Why “Learn”? Machine learning is programming computers to optimize a performance criterion using example data or past experience.
Feature Selection in k-Median Clustering Olvi Mangasarian and Edward Wild University of Wisconsin - Madison.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. Mining massive document collections by the WEBSOM method Presenter : Yu-hui Huang Authors :Krista Lagus,
Flat clustering approaches
What are the advantages of using bar code scanner?  Fast  It is fast  It is fast for reading data  It is fast for data input  Accurate  The advantage.
Lower Bounds for Embedding Edit Distance into Normed Spaces A. Andoni, M. Deza, A. Gupta, P. Indyk, S. Raskhodnikova.
WHAT IS THIS? Clue…it’s a drink SIMPLE SEQUENCE CONTROL STRUCTURE Introduction A computer is an extremely powerful, fast machine. In less than a second,
CS307P-SYSTEM PRACTICUM CPYNOT. B13107 – Amit Kumar B13141 – Vinod Kumar B13218 – Paawan Mukker.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Mismatch String Kernals for SVM Protein Classification Christina Leslie, Eleazar Eskin, Jason Weston, William Stafford Noble Presented by Pradeep Anand.
Fast algorithm and implementation of dissimilarity self-organizing maps Reporter: Ming-Jui Kuo D
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Today’s Lecture Neural networks Training
Presented by Niwan Wattanakitrungroj
Real-Time Soft Shadows with Adaptive Light Source Sampling
Self Organizing Maps: Parametrization of Parton Distribution Functions
A modified hyperplane clustering algorithm allows for efficient and accurate clustering of extremely large datasets Ashok Sharma, Robert Podolsky, Jieping.
QianZhu, Liang Chen and Gagan Agrawal
Introduction The average running speed of a human being is 5–8 mph. Imagine you set out on an 8-mile run that takes you no longer than 1 hour. You run.
PSG College of Technology
DSS & Warehousing Systems
Introduction to Problem Solving and Programming CS140: Introduction to Computing 1 8/19/13.
Theory of Algorithms Introduction.
Spatial Online Sampling and Aggregation
A weight-incorporated similarity-based clustering ensemble method based on swarm intelligence Yue Ming NJIT#:
State Space Analysis and Controller Design
A Fault-Tolerant Routing Strategy for Fibonacci-Class Cubes
Lecture 22 Clustering (3).
Neural Networks and Their Application in the Fields of Coporate Finance By Eric Séverin Hanna Viinikainen.
Clustering 77B Recommender Systems
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
Topological Signatures For Fast Mobility Analysis
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Inductive Clustering: A technique for clustering search results Hieu Khac Le Department of Computer Science - University of Illinois at Urbana-Champaign.
Presentation transcript:

Reporter: Ming-Jui Kuo D9515007 Fast algorithm and implementation of dissimilarity self-organizing maps Reporter: Ming-Jui Kuo D9515007

Outline Introduction The DSOM Simulation Results

Introduction

A drawback in standard SOM Since vectors from a fixed and finite-dimensional vector space. Unfortunately, many real-world data depart strongly from this model. It is quite common, for instance, to have variable-sized data. They are natural, for example, in online handwriting recognition where the representation of a character drawn by the user can vary in length because of the drawing conditions. Other data, such as texts for instance, are strongly non-numerical and have a complex internal structure: they are very difficult to represent accurately in a vector space.

Related papers [1] Teuvo Kohonen*, Panu Somervuo, “Self-organizing maps of symbol strings,” Neurocomputing, vol 21, pp. 19-30, 1998 [2] Teuvo Kohonen*, Panu Somervuo, “How to make large self-organizing maps for nonvectorial data,” vol. 15, pp. 945-152, 2002 [3] Aïcha El Golli , Brieuc Conan-Guez, and Fabrice Rossi, “A Self Organizing Map for dissimilarity data,” IFCS, 2004, Proceedings. [4] Aïcha El Golli , “Speeding up the self organizing map for dissimilarity data”

Alias Name Median self-organizing map : Median SOM [2] Dissimilarity self-organizing map : DSOM [3]

Related web sites http://apiacoa.org/ http://lists.gforge.inria.fr/pipermail/somlib-commits by Fabrice Rossi

The goal of this paper A major drawback of the DSOM is that its running time can be very high, especially when compared to the standard vector SOM. It is well known that the SOM algorithm behaves linearly with the number of input data. In contrast, the DSOM behaves quadratically with this number.

The goal of this paper (cont’d) In this paper, the authors propose several modifications of the basic algorithm that allow a much faster implementation. The quadratic nature of the algorithm cannot be avoided, essentially because dissimilarity data are intrinsically described by a quadratic number of one-to-one dissimilarities.

The goal of this paper (cont’d) The standard DSOM algorithm cost is proportional to, where N is the number of observations and M the number of clusters that the algorithm has to produce, whereas the modifications of this paper lead to a cost proportional to. (save in the representation phase) An important property of all modifications in this paper is that the obtained algorithm produces exactly the same results as the standard DSOM algorithm.

Dissimilarity Data In a given data set X, use the dissimilarity measure to measure the dissimilarity between data instances (one-to-one, pairwise). Sometimes, the distance measurement can be used.

Dissimilarity Data (cont’d)

Dissimilarity Data (cont’d)

The DSOM

The DSOM algorithm 1: choose initial values for {Initialization phase} 2: for l = 1 to L do 3: for all do {Template for the affectation phase} 4: compute 5: end for 6: for all do {Template for the representation phase} 7: compute 8: end for 9: end for

The DSOM

Partial Sums

Early Stopping

Simulation Results

Thank You for your attention!!