Lecture 14, CS5671 Clustering Algorithms Density based clustering Self organizing feature maps Grid based clustering Markov clustering.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Artificial Neural Networks (1)
2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Graphs Graphs are the most general data structures we will study in this course. A graph is a more general version of connected nodes than the tree. Both.
November 18, 2010Neural Networks Lecture 18: Applications of SOMs 1 Assignment #3 Question 2 Regarding your cascade correlation projects, here are a few.
Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Unsupervised learning. Summary from last week We explained what local minima are, and described ways of escaping them. We investigated how the backpropagation.
Kohonen Self Organising Maps Michael J. Watts
Unsupervised Networks Closely related to clustering Do not require target outputs for each input vector in the training data Inputs are connected to a.
Self-Organizing Map (SOM). Unsupervised neural networks, equivalent to clustering. Two layers – input and output – The input layer represents the input.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
. Markov Chains as a Learning Tool. 2 Weather: raining today40% rain tomorrow 60% no rain tomorrow not raining today20% rain tomorrow 80% no rain tomorrow.
X0 xn w0 wn o Threshold units SOM.
CSE 522 – Algorithmic and Economic Aspects of the Internet Instructors: Nicole Immorlica Mohammad Mahdian.
Support Vector Machines and Kernel Methods
Neural Networks Basic concepts ArchitectureOperation.
Neural Networks. R & G Chapter Feed-Forward Neural Networks otherwise known as The Multi-layer Perceptron or The Back-Propagation Neural Network.
Neural Networks in Electrical Engineering Prof. Howard Silver School of Computer Sciences and Engineering Fairleigh Dickinson University Teaneck, New Jersey.
Hypercubes and Neural Networks bill wolfe 10/23/2005.
Markov Cluster Algorithm
November 24, 2009Introduction to Cognitive Science Lecture 21: Self-Organizing Maps 1 Self-Organizing Maps (Kohonen Maps) In the BPN, we used supervised.
Clustering An overview of clustering algorithms Dènis de Keijzer GIA 2004.
Neural Networks Lecture 17: Self-Organizing Maps
A Hybrid Self-Organizing Neural Gas Network James Graham and Janusz Starzyk School of EECS, Ohio University Stocker Center, Athens, OH USA IEEE World.
Lecture 09 Clustering-based Learning
Neural Networks Lecture 8: Two simple learning algorithms
Graph mining in bioinformatics Laur Tooming. Graphs in biology Graphs are often used in bioinformatics for describing processes in the cell Vertices are.
CS623: Introduction to Computing with Neural Nets (lecture-20) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
Self-organizing Maps Kevin Pang. Goal Research SOMs Research SOMs Create an introductory tutorial on the algorithm Create an introductory tutorial on.
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
Liang Ge.  Introduction  Important Concepts in MCL Algorithm  MCL Algorithm  The Features of MCL Algorithm  Summary.
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
Neural Network Introduction Hung-yi Lee. Review: Supervised Learning Training: Pick the “best” Function f * Training Data Model Testing: Hypothesis Function.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Machine Learning Neural Networks (3). Understanding Supervised and Unsupervised Learning.
Self Organizing Feature Map CS570 인공지능 이대성 Computer Science KAIST.
Optimization with Neural Networks Presented by: Mahmood Khademi Babak Bashiri Instructor: Dr. Bagheri Sharif University of Technology April 2007.
Markov Cluster (MCL) algorithm Stijn van Dongen.
UNSUPERVISED LEARNING NETWORKS
Trust Management for the Semantic Web Matthew Richardson1†, Rakesh Agrawal2, Pedro Domingos By Tyrone Cadenhead.
TreeSOM :Cluster analysis in the self- organizing map Neural Networks 19 (2006) Special Issue Reporter 張欽隆 D
381 Self Organization Map Learning without Examples.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Computational Biology Clustering Parts taken from Introduction to Data Mining by Tan, Steinbach, Kumar Lecture Slides Week 9.
Semiconductors, BP&A Planning, DREAM PLAN IDEA IMPLEMENTATION.
John Lafferty Andrew McCallum Fernando Pereira
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Adaptive FIR Neural Model for Centroid Learning in Self-Organizing.
Self-Organizing Maps (SOM) (§ 5.5)
Self Organizing Maps: Clustering With unsupervised learning there is no instruction and the network is left to cluster patterns. All of the patterns within.
CHAPTER 14 Competitive Networks Ming-Feng Yeh.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
A Self-organizing Semantic Map for Information Retrieval Xia Lin, Dagobert Soergel, Gary Marchionini presented by Yi-Ting.
 Introduction  Important Concepts in MCL Algorithm  MCL Algorithm  The Features of MCL Algorithm  Summary.
Machine Learning 12. Local Models.
Fall 2004 Backpropagation CS478 - Machine Learning.
Self-Organizing Network Model (SOM) Session 11
Data Mining, Neural Network and Genetic Programming
Other Applications of Energy Minimzation
Lecture 22 Clustering (3).
CSE572, CBS598: Data Mining by H. Liu
Graph Clustering based on Random Walk
Self-Organizing Maps (SOM) (§ 5.5)
Self Organizing Maps A major principle of organization is the topographic map, i.e. groups of adjacent neurons process information from neighboring parts.
Feature mapping: Self-organizing Maps
Artificial Neural Networks
Presentation transcript:

Lecture 14, CS5671 Clustering Algorithms Density based clustering Self organizing feature maps Grid based clustering Markov clustering

Lecture 14, CS5672 Density based clustering Rationale –“Definition of city limits” –General version of single-linkage clustering (How?) –A cluster is a region where many points are concentrated –Density within each cluster is higher than cut-off –Sparse areas excluded from clustering Algorithm –Given: {i} and minDensity = minimum number of points within radius r of i –Pick all i that surpass minDensity to form initial clusters –Merge all clusters that can be connected through any j that surpasses minDensity Advantages –Don’t need to decide on number of clusters beforehand –Clusters of arbitrary shape can be picked up Limitations –Need to decide on radius and minDensity –Doesn’t differentiate between clusters that barely meet minDensity versus those that are much denser

Lecture 14, CS5673 Grid based clustering Rationale –“’Till death do us apart’, said the point to its neighbor” Algorithm –Given: {i} and grid of resolution s –Replace all i in each cell by summarized information, e.g., mean/mode –Cluster the grid cells rather than each object Advantages –Scalability, scalability, scalability –Performance independent of number of objects, as long as they exist in the same space “Takes equal amount of time to cluster New Yorkers and Icelanders” Limitations –Resolution confined to that of grid –Cluster boundary always corresponds to orientation of grid cell surfaces

Lecture 14, CS5674 Self-organizing feature maps (SOM) Rationale –“If you are truly my neighbor, say yes when I do, and say no when I do, and together we shall re-organize the world” –“Untangling puppet strings” Algorithm –Given: Input vectors {i} and NN with multiple output neurons {o} –Every time an output neuron o is activated by winner take-all, The neuron and its immediate neighbors are biased towards responding more strongly to the same example in future W new = W old - α(W old ) + α(P) (Kohonen) α(P) term makes weights approach the corresponding input vector per se Negative α(W old ) term ensures that weights do not increase infinitely Backpropagation not applicable here (Why?) –Finally, neighborhoods of output neurons map to each input vector Advantages –Unsupervised Neural Network (No targets necessary) –Used for expression array analysis (Different differences in expression level between multiple experiments are mapped onto different sets of output neurons) Limitations –Like the human brain, i.e., slow

Lecture 14, CS5675 Markov clustering algorithm (van Dongen) Rationale –“Let several drunks loose in a graph (Markov chain) with some chalk. Come back the next day and declare each well traveled area as a cluster” Principles: –No. of multi-step paths between two nodes is higher within clusters than between –A random walk between nodes will linger in a cluster before moving to a different one –Bridge edges will figure with high frequency in shortest paths between all nodes in data set Algorithm –Given: Markov chain of states (nodes) with symmetric probabilities of transitions –Alternate expansion (matrix squaring; “spread riches, but only within each kingdom”) and inflation (entry-wise squaring with rescaling for stochasticity; “more power to the mighty, and rob the weak of what little they have left”) till transition matrix converges –Each set of connected nodes constitutes a cluster Advantages –Numbers of clusters need not be specified explicitly –Scales well Limitations –Should be possible to transform distances into stochastic weight matrix