1 Abstract This paper presents a novel modification to the classical Competitive Learning (CL) by adding a dynamic branching mechanism to neural networks.

Slides:



Advertisements
Similar presentations
Aggregating local image descriptors into compact codes
Advertisements

Perceptron Learning Rule
Computer Vision Lecture 18: Object Recognition II
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Proposed concepts illustrated well on sets of face images extracted from video: Face texture and surface are smooth, constraining them to a manifold Recognition.
Pei Fan*, Ji Wang, Zibin Zheng, Michael R. Lyu Toward Optimal Deployment of Communication-Intensive Cloud Applications 1.
Fingerprint Minutiae Matching Algorithm using Distance Histogram of Neighborhood Presented By: Neeraj Sharma M.S. student, Dongseo University, Pusan South.
Self Organization: Competitive Learning
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 On Rival Penalization Controlled Competitive Learning.
A Novel Scheme for Video Similarity Detection Chu-Hong Hoi, Steven March 5, 2003.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Local Clustering Algorithm DISCOVIR Image collection within a client is modeled as a single cluster. Current Situation.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Radial Basis Functions
Reinforcement Learning Rafy Michaeli Assaf Naor Supervisor: Yaakov Engel Visit project’s home page at: FOR.
NORM BASED APPROACHES FOR AUTOMATIC TUNING OF MODEL BASED PREDICTIVE CONTROL Pastora Vega, Mario Francisco, Eladio Sanz University of Salamanca – Spain.
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
VARIABLE PRESELECTION LIST LENGTH ESTIMATION USING NEURAL NETWORKS IN A TELEPHONE SPEECH HYPOTHESIS-VERIFICATION SYSTEM J. Macías-Guarasa, J. Ferreiros,
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
Handwritten Thai Character Recognition Using Fourier Descriptors and Robust C-Prototype Olarik Surinta Supot Nitsuwat.
Multiple Object Class Detection with a Generative Model K. Mikolajczyk, B. Leibe and B. Schiele Carolina Galleguillos.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
05/06/2005CSIS © M. Gibbons On Evaluating Open Biometric Identification Systems Spring 2005 Michael Gibbons School of Computer Science & Information Systems.
Birch: An efficient data clustering method for very large databases
Radial-Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
 C. C. Hung, H. Ijaz, E. Jung, and B.-C. Kuo # School of Computing and Software Engineering Southern Polytechnic State University, Marietta, Georgia USA.
Presented by Tienwei Tsai July, 2005
Hierarchical Distributed Genetic Algorithm for Image Segmentation Hanchuan Peng, Fuhui Long*, Zheru Chi, and Wanshi Siu {fhlong, phc,
Face Recognition using Parallel Associative Memory 2009/12/31 學生:羅國育.
Rotation Invariant Neural-Network Based Face Detection
CSE 185 Introduction to Computer Vision Pattern Recognition 2.
COMPARISON OF IMAGE ANALYSIS FOR THAI HANDWRITTEN CHARACTER RECOGNITION Olarik Surinta, chatklaw Jareanpon Department of Management Information System.
Combined Central and Subspace Clustering for Computer Vision Applications Le Lu 1 René Vidal 2 1 Computer Science Department, Johns Hopkins University,
AUTOWISARD: Unsupervised Modes for the WISARD Authors: Iuri Wickert Felipe M. G. França Computer Systems Engineering.
CHAPTER 7: Clustering Eick: K-Means and EM (modified Alpaydin transparencies and new transparencies added) Last updated: February 25, 2014.
Machine Learning Neural Networks (3). Understanding Supervised and Unsupervised Learning.
So Far……  Clustering basics, necessity for clustering, Usage in various fields : engineering and industrial fields  Properties : hierarchical, flat,
The Application of The Improved Hybrid Ant Colony Algorithm in Vehicle Routing Optimization Problem International Conference on Future Computer and Communication,
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman modified by Hanne Jarmer.
CURE: An Efficient Clustering Algorithm for Large Databases Sudipto Guha, Rajeev Rastogi, Kyuseok Shim Stanford University Bell Laboratories Bell Laboratories.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
A Trust Based Distributed Kalman Filtering Approach for Mode Estimation in Power Systems Tao Jiang, Ion Matei and John S. Baras Institute for Systems Research.
DATA CLUSTERING WITH KERNAL K-MEANS++ PROJECT OBJECTIVES o PROJECT GOAL  Experimentally demonstrate the application of Kernel K-Means to non-linearly.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
K-Means Algorithm Each cluster is represented by the mean value of the objects in the cluster Input: set of objects (n), no of clusters (k) Output:
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A modified version of the K-means algorithm with a distance.
Face Image-Based Gender Recognition Using Complex-Valued Neural Network Instructor :Dr. Dong-Chul Kim Indrani Gorripati.
Vector Quantization Vector quantization is used in many applications such as image and voice compression, voice recognition (in general statistical pattern.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Adaptive FIR Neural Model for Centroid Learning in Self-Organizing.
Database Management Systems, R. Ramakrishnan 1 Algorithms for clustering large datasets in arbitrary metric spaces.
Faculty of Information Engineering, Shenzhen University Liao Huilian SZU TI-DSPs LAB Aug 27, 2007 Optimizer based on particle swarm optimization and LBG.
Application of the GA-PSO with the Fuzzy controller to the robot soccer Department of Electrical Engineering, Southern Taiwan University, Tainan, R.O.C.
ViSOM - A Novel Method for Multivariate Data Projection and Structure Visualization Advisor : Dr. Hsu Graduate : Sheng-Hsuan Wang Author : Hujun Yin.
Iterative K-Means Algorithm Based on Fisher Discriminant UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE JOENSUU, FINLAND Mantao Xu to be presented.
Similarity Measurement and Detection of Video Sequences Chu-Hong HOI Supervisor: Prof. Michael R. LYU Marker: Prof. Yiu Sang MOON 25 April, 2003 Dept.
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Intelligent Database Systems Lab Advisor : Dr. Hsu Graduate : Jian-Lin Kuo Author : Aristidis Likas Nikos Vlassis Jakob J.Verbeek 國立雲林科技大學 National Yunlin.
Cluster Analysis What is Cluster Analysis? Types of Data in Cluster Analysis A Categorization of Major Clustering Methods Partitioning Methods.
1 Munther Abualkibash University of Bridgeport, CT.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Self-Organizing Network Model (SOM) Session 11
A modified hyperplane clustering algorithm allows for efficient and accurate clustering of extremely large datasets Ashok Sharma, Robert Podolsky, Jieping.
Rutgers Intelligent Transportation Systems (RITS) Laboratory
Image Processing for Physical Data
Pei Fan*, Ji Wang, Zibin Zheng, Michael R. Lyu
Presentation transcript:

1 Abstract This paper presents a novel modification to the classical Competitive Learning (CL) by adding a dynamic branching mechanism to neural networks so that the number of neurons can be increased over time until the networks reaches a good estimation of the cluster number in a data set. The algorithm, called Branching Competitive Learning (BCL), shows a fast convergence of the synaptic vectors to cluster centroids, and more importantly, shows the ability to automatically detect cluster number in a data distribution. We illustrate the formulation of the Branching Criteria and demonstrate the efficiency of BCL for data clustering through a set of experiments.

2 Introduction Some applications of data clustering: Pattern recognition Vector quantization image coding Image database indexing Key problems of most clustering algorithms: The number of clusters must be appropriately preselected, e.g., K-mean and classical CL Sensitive to the preselected cluster number and the initialization of synaptic vectors, e.g., RPCL

3 Contributions 1Propose a neuron branching mechanism to estimate cluster number and cluster data 2Present the Branching Criteria 3Present a new way of hierarchical data clustering, i.e., multi-resolution clustering The advantages of BCL for clustering: The ability to automatically detect cluster number Fast convergence of synaptic vectors Convenience to implement multiresolution data clustering

4 The Branching Criterion There are two conditions for a neuron to spawn a new one: The Angle Criteria — Based on the angle between current moving direction and the previous moving direction of a synaptic vector: (1) The Distance Criteria — Based on the distance between the input sample and winner: (2) where: : A randomly selected sample at t, the current step : The winner in current competition, : Angle and distance thresholds

5 The Algorithm of BCL for Clustering  Initialize the first synaptic vector 2Randomly take a sample from the data set, find the winner of current competition in the synaptic vector set, i.e.,, where the is the frequency that has won the competition up to now 3If satisfies the branching criterion above, a new neuron is spawn off from : otherwise, update by

6 Illustrations of BCL Figure 1: (a) An illustration of the procedure of the BCL algorithm, where: (1) Initialization of the first synaptic vector (2) Branching points of synaptic vectors (3) Final convergence of synaptic vectors (b) An illustration of a branching point, where: and

7 Experiments We conduct three sets of experiments: I The first set of experiments examines the ability of BCL to detect cluster number automatically II The second set of experiments shows a multiresolution clustering in BCL scheme III The third set of experiments compares the performance of BCL and RPCL The experimental environment: Pentium II PC with 128 Meg of internal memory running on Windows98 Implement BCL and RPCL algorithm in Visual C++6.0

8 Experiment I (1) Cluster Number Detection (a) (b) Figure 2: The learning and branching traces on a data set, which contains four Gaussian clusters (1,000 samples) with  =0.5 and centered at (-2, 0), (2, 0), (0, -2), and (0, 2) respectively. The first synaptic vector is initialized by point (4, 4) in (a) or by point (0, 0) in (b).

9 Experiment I (2) (c) (d) Figure 3: The learning and branching traces on a data set, which contains four overlapping Gaussian clusters (1,000 samples) with  =0.5 and centered at (-1.5, 0), (1.5, 0), (0, -1.5), and (0, 1.5) res- pectively. The first synaptic vector is initialized by point (3, 3) in (c) or by point (0, 0) in (d).

10 Experiment II (1) Multiresolution Data Clustering Figure 4: A multiresolution data set, where: (a) A view of the data set in a small resolution (b) A view of the data set in a large resolution Data clustering and cluster number detection are resolution-dependent

11 Experiment II (2) Figure 5: The learning and branching traces on the multiresolution data set, where: (c) Presents the learning trace in level 1 (d) Presents the learning trace in 2-level BCL

12 Experiment II (3) Figure 6: The error between the centroids of data clusters and the estimated cluster centroids.

13 Experiment III (1) Comparison of BCL and RPCL Two measures used for comparing BCL and RPCL: The average accuracy of data clustering The average speed or the average running time of the algorithms Table I. 5-Dimensional data sets Data Set Distribution Sample Cluster Overlapping No. Numb. Numb. 1 Uniform 2, none 2 Uniform 2, small 3 Gaussian 2, small 4 Gaussian 2, larger

14 Experiment III (2) Experimental results (over 20 trials) Table II. Average Accuracy RPCL BCL Data Set % % Data Set % % Data Set % 97.16% Data Set % 96.27% Table II. Average Speed (s) RPCL BCL Data Set Data Set Data Set Data Set

15 Discussion & Conclusion Discussion: Robust to various initial conditions The can be seen as a resolution control How to choose ? The advantages of BCL for clustering: The ability to automatically detect cluster number Fast convergence of synaptic vectors Convenience to implement multiresolution data clustering