Two Technique Papers on High Dimensionality Allan Rempel December 5, 2005.

Slides:



Advertisements
Similar presentations
Self-Organizing Maps Projection of p dimensional observations to a two (or one) dimensional grid space Constraint version of K-means clustering –Prototypes.
Advertisements

Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
Multimedia DBs. Multimedia dbs A multimedia database stores text, strings and images Similarity queries (content based retrieval) Given an image find.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
Clustering and Dimensionality Reduction Brendan and Yifang April
MDSteer: Steerable and Progressive Multidimensional Scaling Matt Williams and Tamara Munzner University of British Columbia Imager Lab.
Graph Drawing Zsuzsanna Hollander. Reviewed Papers Effective Graph Visualization via Node Grouping Janet M. Six and Ioannis G. Tollis. Proc InfoVis 2001.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
DIMENSIONALITY REDUCTION BY RANDOM PROJECTION AND LATENT SEMANTIC INDEXING Jessica Lin and Dimitrios Gunopulos Ângelo Cardoso IST/UTL December
Dimensionality Reduction and Embeddings
Dimension reduction : PCA and Clustering Agnieszka S. Juncker Slides: Christopher Workman and Agnieszka S. Juncker Center for Biological Sequence Analysis.
Dimensionality Reduction
University of CreteCS4831 The use of Minimum Spanning Trees in microarray expression data Gkirtzou Ekaterini.
L16: Micro-array analysis Dimension reduction Unsupervised clustering.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman.
Scalability for High Cardinality in Steerable MDS CPSC 533C Project Presentation Allan Rempel December 19, 2005.
Force Directed Algorithm Adel Alshayji 4/28/2005.
Switch to Top-down Top-down or move-to-nearest Partition documents into ‘k’ clusters Two variants “Hard” (0/1) assignment of documents to clusters “soft”
Dimension reduction : PCA and Clustering Christopher Workman Center for Biological Sequence Analysis DTU.
Force Directed Algorithm Adel Alshayji 4/28/2005.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Y. Weiss (Hebrew U.) A. Torralba (MIT) Rob Fergus (NYU)
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Dimensionality Reduction
CS Instance Based Learning1 Instance Based Learning.
Clustering and MDS Exploratory Data Analysis. Outline What may be hoped for by clustering What may be hoped for by clustering Representing differences.
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
Projection methods in chemistry Autumn 2011 By: Atefe Malek.khatabi M. Daszykowski, B. Walczak, D.L. Massart* Chemometrics and Intelligent Laboratory.
Dimension Reduction and Visualization of Large High-Dimensional Data via Interpolation Seung-Hee Bae, Jong Youl Choi, Judy Qiu, and Geoffrey Fox School.
Microarray Gene Expression Data Analysis A.Venkatesh CBBL Functional Genomics Chapter: 07.
Fast Subsequence Matching in Time-Series Databases Christos Faloutsos M. Ranganathan Yannis Manolopoulos Department of Computer Science and ISR University.
The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang.
Unsupervised Learning and Clustering k-means clustering Sum-of-Squared Errors Competitive Learning SOM Pre-processing and Post-processing techniques.
Self-organizing Maps Kevin Pang. Goal Research SOMs Research SOMs Create an introductory tutorial on the algorithm Create an introductory tutorial on.
Boris Babenko Department of Computer Science and Engineering University of California, San Diego Semi-supervised and Unsupervised Feature Scaling.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Adaptive nonlinear manifolds and their applications to pattern.
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
1 Smashing Peacocks Further: Drawing Quasi-Trees from Biconnected Components Daniel Archambault and Tamara Munzner, University of British Columbia David.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
A hybrid SOFM-SVR with a filter-based feature selection for stock market forecasting Huang, C. L. & Tsai, C. Y. Expert Systems with Applications 2008.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
Multidimensional Scaling Vuokko Vuori Based on: Data Exploration Using Self-Organizing Maps, Samuel Kaski, Ph.D. Thesis, 1997 Multivariate Statistical.
Multidimensional Scaling by Deterministic Annealing with Iterative Majorization Algorithm Seung-Hee Bae, Judy Qiu, and Geoffrey Fox SALSA group in Pervasive.
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman modified by Hanne Jarmer.
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Manifold learning: MDS and Isomap
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
By Timofey Shulepov Clustering Algorithms. Clustering - main features  Clustering – a data mining technique  Def.: Classification of objects into sets.
University “Ss. Cyril and Methodus” SKOPJE Cluster-based MDS Algorithm for Nodes Localization in Wireless Sensor Networks Ass. Biljana Stojkoska.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
Data Mining Course 2007 Eric Postma Clustering. Overview Three approaches to clustering 1.Minimization of reconstruction error PCA, nlPCA, k-means clustering.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Application: Multiresolution Curves Jyun-Ming Chen Spring 2001.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
ViSOM - A Novel Method for Multivariate Data Projection and Structure Visualization Advisor : Dr. Hsu Graduate : Sheng-Hsuan Wang Author : Hujun Yin.
CSC321: Lecture 25: Non-linear dimensionality reduction Geoffrey Hinton.
Iterative K-Means Algorithm Based on Fisher Discriminant UNIVERSITY OF JOENSUU DEPARTMENT OF COMPUTER SCIENCE JOENSUU, FINLAND Mantao Xu to be presented.
Neural Network Approximation of High- dimensional Functions Peter Andras School of Computing and Mathematics Keele University
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A Nonlinear Mapping for Data Structure Analysis John W.
Out of sample extension of PCA, Kernel PCA, and MDS WILSON A. FLORERO-SALINAS DAN LI MATH 285, FALL
4.0 - Data Mining Sébastien Lemieux Elitra Canada Ltd.
SDE: Graph Drawing Using Spectral Distance Embedding
Dimension reduction : PCA and Clustering
Presentation transcript:

Two Technique Papers on High Dimensionality Allan Rempel December 5, 2005

Papers A. Morrison, G. Ross, and M. Chalmers. Fast multidimensional scaling through sampling, springs and interpolation. In Information Visualization, pages 68-77, F. Jourdan and G. Melançon. Multiscale hybrid MDS. In Intl. Conf. on Information Visualization (London), pages , well written, clear, appropriately detailed High-dim and MDS can be complicated

Morrison, Ross, Chalmers Dimensionality reduction Mapping high-dimensional data to 2D space Could be done many different ways Different techniques satisfy different goals Familiar example - projection of 3D to 2D preserves geometric relationships Abstract data may not need that

Multidimensional scaling (MDS) Display multivariate abstract point data in 2D  Data from bioinformatics, financial sector, etc.  No inherent mapping in 2D space  p-dim embedding of q-dim space (p < q) where inter-object relationships are approximated in low-dimensional space Proximity in high-D -> proximity in 2D  High-dim distance between points (similarity) determines relative (x,y) position  Absolute (x,y) positions are not meaningful Clusters show closely associated points

Multidimensional scaling (MDS) Eigenvector analysis of N x N matrix – O(N 3 )  Need to recompute if data changes slightly Iterative O(N 2 ) algorithm – Chalmers,1996 This paper – Next paper – O(N log N)

Multidimensional scaling (MDS) Proximity data  In social sciences, geology, archaeology, etc.  E.g. library catalogue query – find similar points Multi-dimensional scatterplot not possible  Want to see clusters, curves, etc. Features that stand out from the noise Distance function  Typically use Euclidean distance – intuitive

Spring models Used instead of statistical techniques (PCA)  Better convergence to optimal solution  Iterative – steerable – Munzner et al, 2004 Good aesthetic results – symmetry, edge lengths Basic algorithm – O(N 3 )  Start: place points randomly in 2D space  Springs reflect diff btwn high-D and 2D distance  #iterations required is generally O(N)

Chalmers’ 1996 algorithm Approximate solution works well Caching, stochastic sampling – O(N 2 )  Perform each iteration in O(N) instead of O(N 2 )  Keep constant-size set of neighbours  Constants as low as 5 worked well Still only worked on datasets up to few 1000s

Hybrid methods of clustering and layout Diff clustering algorithms have diff strengths  Kohonen’s self-organising feature maps (SOM)  K-means iterative centroid-based divisive alg. Hybrid methods have produced benefits Neural networks, machine learning literature

New hybrid MDS approach Start: run spring model on subset of size  Completes in O(N) For each remaining point:  Place close to closest ‘anchor’  Adjust by adding spring forces to other anchors Overall complexity

Experimental results 3-D data sets: 5000 – 50,000 points 13-D data sets: 2000 – 24,000 points Took less than 1/3 the time of the O(N 2 ) Achieved lower stress when done Also compared against original O(N 3 ) model  9 seconds vs. 577; and 24 vs  Achieved much lower stress (0.06 vs. 0.2)

Experimental results

Future work Hashing Pivots – Morrison, Chalmers, 2003  Achieved Dynamically resizing anchor set Proximity grid  Do MDS, then transform continuous layout into discrete topology

Jourdan and Melançon Multiscale hybrid MDS Extension of previous paper Achieves O(N log N) time complexity Good introduction of Chalmers et al papers Like Chalmers, begins by embedding subset S of size

Improving parent-finding strategy Select constant-size subset For each p in P create sorted list L p For each remaining point u, binary search L p for point u p as distant from p as u is  Implies that u and u p are very close Place u according to location of u p

Comparison Chalmers et al is better for N < 5500 Main diff is in parent-finding, represented by Fig. 3

Comparison Experimental study confirms theoretical results This technique becomes better for N > 70,000

Quality of output MDS theory uses stress to objectively determine quality of placement of points Subjective determinations can be made too  2D small world network example (500 – 80,000 nodes)

Multiscale MDS Recursively defining the initial kernel set of points can yield much better real-time performance

Conclusions and future work Series of results yielding progressively better time complexities for MDS 2D mappings provide good representations Further examination of multiscale approach User-steerable MDS could be fruitful