Iterative Optimization and Simplification of Hierarchical Clusterings Doug Fisher Department of Computer Science, Vanderbilt University Journal of Artificial.

Slides:



Advertisements
Similar presentations
Discrimination amongst k populations. We want to determine if an observation vector comes from one of the k populations For this purpose we need to partition.
Advertisements

Conceptual Clustering
Social network partition Presenter: Xiaofei Cao Partick Berg.
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Iterative Optimization and Simplification of Hierarchical Clusterings Doug Fisher Department of Computer Science, Vanderbilt University Journal of Artificial.
K Means Clustering , Nearest Cluster and Gaussian Mixture
Clustering Clustering of data is a method by which large sets of data is grouped into clusters of smaller sets of similar data. The example below demonstrates.
Experiments on Query Expansion for Internet Yellow Page Services Using Log Mining Summarized by Dongmin Shin Presented by Dongmin Shin User Log Analysis.
Introduction to Bioinformatics
1 Using Search in Problem Solving Part II. 2 Basic Concepts Basic concepts: Initial state Goal/Target state Intermediate states Path from the initial.
Iterative Optimization of Hierarchical Clusterings Doug Fisher Department of Computer Science, Vanderbilt University Journal of Artificial Intelligence.
Creating Concept Hierarchies in a Customer Self-Help System Bob Wall CS /29/05.
Cluster Analysis.  What is Cluster Analysis?  Types of Data in Cluster Analysis  A Categorization of Major Clustering Methods  Partitioning Methods.
L16: Micro-array analysis Dimension reduction Unsupervised clustering.
Semantic text features from small world graphs Jure Leskovec, IJS + CMU John Shawe-Taylor, Southampton.
4. Ad-hoc I: Hierarchical clustering
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Gene Expression 1. Methods –Unsupervised Clustering Hierarchical clustering K-means clustering Expression data –GEO –UCSC EPCLUST 2.
Expectation-Maximization
1 Topology Design of Structured Campus Networks by Habib Youssef Sadiq M. SaitSalman A. Khan Department of Computer Engineering King Fahd University of.
Revision (Part II) Ke Chen COMP24111 Machine Learning Revision slides are going to summarise all you have learnt from Part II, which should be helpful.
Clustering Unsupervised learning Generating “classes”
Evaluating Performance for Data Mining Techniques
Image Segmentation Image segmentation is the operation of partitioning an image into a collection of connected sets of pixels. 1. into regions, which usually.
CPSC 335 BTrees Dr. Marina Gavrilova Computer Science University of Calgary Canada.
Game Trees: MiniMax strategy, Tree Evaluation, Pruning, Utility evaluation Adapted from slides of Yoonsuck Choe.
Data Mining Chun-Hung Chou
Unsupervised Learning. CS583, Bing Liu, UIC 2 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate.
Surface Simplification Using Quadric Error Metrics Michael Garland Paul S. Heckbert.
Presented by Tienwei Tsai July, 2005
Clustering Analysis of Spatial Data Using Peano Count Trees Qiang Ding William Perrizo Department of Computer Science North Dakota State University, USA.
The BIRCH Algorithm Davitkov Miroslav, 2011/3116
Computational Intelligence: Methods and Applications Lecture 30 Neurofuzzy system FSM and covering algorithms. Włodzisław Duch Dept. of Informatics, UMK.
What is Genetic Programming? Genetic programming is a model of programming which uses the ideas (and some of the terminology) of biological evolution to.
tch?v=Y6ljFaKRTrI Fireflies.
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
Game Playing. Introduction One of the earliest areas in artificial intelligence is game playing. Two-person zero-sum game. Games for which the state space.
Information Retrieval Lecture 6 Introduction to Information Retrieval (Manning et al. 2007) Chapter 16 For the MSc Computer Science Programme Dell Zhang.
Clustering I. 2 The Task Input: Collection of instances –No special class label attribute! Output: Clusters (Groups) of instances where members of a cluster.
CLUSTER ANALYSIS Introduction to Clustering Major Clustering Methods.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Clustering.
Sparse Signals Reconstruction Via Adaptive Iterative Greedy Algorithm Ahmed Aziz, Ahmed Salim, Walid Osamy Presenter : 張庭豪 International Journal of Computer.
Prepared by: Mahmoud Rafeek Al-Farra
Cluster Analysis Potyó László. Cluster: a collection of data objects Similar to one another within the same cluster Similar to one another within the.
Clustering Clustering is a technique for finding similarity groups in data, called clusters. I.e., it groups data instances that are similar to (near)
CS 8751 ML & KDDData Clustering1 Clustering Unsupervised learning Generating “classes” Distance/similarity measures Agglomerative methods Divisive methods.
Machine Learning Queens College Lecture 7: Clustering.
Community Discovery in Social Network Yunming Ye Department of Computer Science Shenzhen Graduate School Harbin Institute of Technology.
CLUSTERING PARTITIONING METHODS Elsayed Hemayed Data Mining Course.
CZ5211 Topics in Computational Biology Lecture 4: Clustering Analysis for Microarray Data II Prof. Chen Yu Zong Tel:
Adversarial Search 2 (Game Playing)
Clustering Wei Wang. Outline What is clustering Partitioning methods Hierarchical methods Density-based methods Grid-based methods Model-based clustering.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
Intelligent Database Systems Lab Advisor : Dr. Hsu Graduate : Jian-Lin Kuo Author : Aristidis Likas Nikos Vlassis Jakob J.Verbeek 國立雲林科技大學 National Yunlin.
Cluster Analysis What is Cluster Analysis? Types of Data in Cluster Analysis A Categorization of Major Clustering Methods Partitioning Methods.
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Semi-Supervised Clustering
Iterative Deepening A*
SRAM Yield Rate Optimization EE201C Final Project
Announcements Homework 3 due today (grace period through Friday)
Clustering.
Minimax strategies, alpha beta pruning
Text Categorization Berlin Chen 2003 Reference:
The Rich/Knight Implementation
Data Mining CSCI 307, Spring 2019 Lecture 24
Minimax strategies, alpha beta pruning
The Rich/Knight Implementation
Presentation transcript:

Iterative Optimization and Simplification of Hierarchical Clusterings Doug Fisher Department of Computer Science, Vanderbilt University Journal of Artificial Intelligence Research, 4 (1996) Presented by: Biyu Liang

2 Outline Introduction Generating Initial Hierarchical Clustering Iterative Optimization Methods and Comparison Simplification of Hierarchical Clustering Conclusion

3 Introduction Clustering is a process of unsupervised learning, which groups objects into clusters. Major Clustering Methods  Partitioning  Hierarchical  Density-based  Grid-based  Model-based

4 Introduction (Continued) Clustering systems differ in objective function control strategy Usually a search strategy cannot be both computationally inexpensive and give any guarantee about the quality.

5 Introduction (Continued)  This paper discusses the use of iterative optimization and simplification to construct clusters that satisfy both conditions: High quality Computationally inexpensive  The suggested method involves 3 steps: Constructing a initial clustering inexpensively Iterative optimization to improve the clustering Retrospective simplification of the clustering

6 Outline Introduction Generating Initial Hierarchical Clustering Iterative Optimization Methods and Experiments Simplification of Hierarchical Clustering Conclusion

7 Category Utility CU(C K ) = P(C k )  i  j [P(A i = V ij |C K ) 2 -P(A i = V ij ) 2 ] PU({C 1, C 2, … C N }) =  k CU(C K )/N Where an observation is a vector of V ij along attributes(or variables) A i This measure rewards clusters C k, that increases the predictability of V ij within C k (i.e. P(A i =V ij |C k )) relative to their predictability in the population as a whole (i.e. P(A i = V ij ))

8 P(Color=gre|C1)

9 Hierarchical Sorting Given an observation and current partition, evaluate the quality of the clusterings that result from  Placing the observation in each of the existing clusters  Creating a new cluster that only covers the new observation Select the option that yields the highest quality score (PU)

10

11 Outline Introduction Generating Initial Hierarchical Clustering Iterative Optimization Methods and Comparison Simplification of Hierarchical Clustering Conclusion

12 Iterative Optimization Methods Reorder-resort (Cluster/2): seed selection, reordering, and re-clustering. Iterative redistribution of single observation: moving single observation one by one. Iterative hierarchical redistribution: moving clusters together with its sub-tree.

13 Reorder-resort (k-mean) k random seeds are selected, and k clusters are growing around these attractors the centroids of the clusters are picked as new seeds, new clusters are growing The process iterates until there is no further improvement in the quality of generated clustering

14 Reorder-resort (k-mean) con’t Ordering data to make consecutive observations dissimilar leads to good clusterings Extracting biased “dissimilarity” ordering from the hierarchical clustering Initial sorting, extraction dissimilarity ordering, re-clustering

15 Iterative Redistribution of Single Observations Moves single observations from cluster to cluster A cluster contains only one observation is removed and its single observation is resorted Iterate until two consecutive iterations yield the same clustering

16 The ISODATA algorithm determines a target cluster for each observation but does not move the cluster until targets for all observations have been determined A sequential version that moves each observation as its target is identified through sorting Single Observation Redistribution Variations

17 Iterative Hierarchical Redistribution Takes large steps in the search for a better clustering Remove and resorts sub-tree instead of single observation Requires update variable value counts of ancestor clusters and host cluster

18 Scheme Given an existing hierarchical clustering, a recursive loop examines sibling clusters in the hierarchy in a depth first fashion. An inner, iterative loop reclassifies each sibling based on the objective function. And repeats until two consecutive iterations lead to the same set of siblings.

19 (Continued) The recursive loop then turns its attention to the children of each of these remaining siblings. Finally the leaves will be reached and resorted. The recursive loop will be applied several times until there are no changes that occur from one pass to the next.

20

21 Main findings from the experiments Hierarchical redistribution achieves the highest mean PU scores in most cases Reordering and re-clustering comes closest to hierarchical redistribution’s performance in all cases Single-observation redistribution modestly improves an initial sort, and is substantially worse than the other two optimization methods

22 Outline Introduction Generating Initial Hierarchical Clustering Iterative Optimization Methods and Comparison Simplification of Hierarchical Clustering Conclusion

23 Simplifying Hierarchical Clustering Simplify hierarchical clustering and minimize classification cost Minimize Error Rate Validation set to identify the frontier of clusters for prediction of each variable Node lies below the frontier of every variable would be pruned

24 Validation For each variable, A i, the objects from the validation set are each classified through the hierarchical clustering with the value of variable A i “masked” for purposes of classification. At each cluster encountered during classification, prediction correct if the observation’s value for A i is equal to the most frequent value for A i at the cluster. A Count of all correct predictions for each variable at a cluster is maintained. A preferred frontier for each variable is identified that maximizes the number of correct counts for the variable.

25

26 Concluding Remarks There are three phases in searching the space of hierarchical clusterings:  Inexpensive generation of an initial clustering  Iterative optimization for clusterings  Retrospective simplification of generated clusterings Experiments found that the new method, hierarchical redistribution optimization works well

Thanks! Question?

28 Final Exam Questions The main idea in this paper is to construct clusterings which satisfy two conditions.  Name the conditions (p.5)  name the two steps to satisfy the conditions Discribe the three iterative methods for clustering optimization (p.12-20) The cluster is better when the relative CU score is a) big, b) small, c) equal to 0 (p.7) Which sorting method is better? a) random sorting, b) similarity sorting (p.14)