Raghuraman Gopalan1, and Jagan Sankaranarayanan2

Slides:



Advertisements
Similar presentations
Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
Advertisements

STOR 892 Object Oriented Data Analysis Radial Distance Weighted Discrimination Jie Xiong Advised by Prof. J.S. Marron Department of Statistics and Operations.
Christoph F. Eick Questions and Topics Review Dec. 10, Compare AGNES /Hierarchical clustering with K-means; what are the main differences? 2. K-means.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Machine learning continued Image source:
Discriminative, Unsupervised, Convex Learning Dale Schuurmans Department of Computing Science University of Alberta MITACS Workshop, August 26, 2005.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
Active Learning with Support Vector Machines
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
SVM (Support Vector Machines) Base on statistical learning theory choose the kernel before the learning process.
A Geometric Framework for Unsupervised Anomaly Detection: Detecting Intrusions in Unlabeled Data Authors: Eleazar Eskin, Andrew Arnold, Michael Prerau,
Linear hyperplanes as classifiers Usman Roshan. Hyperplane separators.
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
Combined Central and Subspace Clustering for Computer Vision Applications Le Lu 1 René Vidal 2 1 Computer Science Department, Johns Hopkins University,
Making classifiers by supervised learning
Extending the Multi- Instance Problem to Model Instance Collaboration Anjali Koppal Advanced Machine Learning December 11, 2007.
Intelligent Database Systems Lab Presenter : BEI-YI JIANG Authors : GUENAEL CABANES, YOUNES BENNANI, DOMINIQUE FRESNEAU ELSEVIER Improving the Quality.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
SUPPORT VECTOR MACHINES Presented by: Naman Fatehpuria Sumana Venkatesh.
Probability & Statistics Box Plots. Describing Distributions Numerically Five Number Summary and Box Plots (Box & Whisker Plots )
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
SUPPORT VECTOR MACHINES
Clustering Clustering definition: Partition a given set of objects into M groups (clusters) such that the objects of each group are ‘similar’ and ‘different’
Support Vector Machines
Probability & Statistics
Support Vector Machines
PREDICT 422: Practical Machine Learning
Support Vector Machine
Semi-Supervised Clustering
Hierarchical Segmentation of Polarimetric SAR Images
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
Ananya Das Christman CS311 Fall 2016
Simone Paolo Ponzetto University of Heidelberg Massimo Poesio
Constrained Clustering -Semi Supervised Clustering-
Network-Wide Bike Availability Clustering Using the College Admission Algorithm: A Case Study of San Francisco Bay Area Hesham Rakha, Ph.D., P.Eng. Samuel.
Data Mining: Concepts and Techniques (3rd ed.) — Chapter 12 —
Support Vector Machines
Unsupervised Learning - Clustering 04/03/17
Nonparametric Methods: Support Vector Machines
Basic machine learning background with Python scikit-learn
Machine Learning Basics
Howard Huang, Sivarama Venkatesan, and Harish Viswanathan
Unsupervised Learning - Clustering
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
K-means and Hierarchical Clustering
Pawan Lingras and Cory Butz
Support Vector Machines Introduction to Data Mining, 2nd Edition by
Support Vector Machines
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Dingding Liu* Yingen Xiong† Linda Shapiro* Kari Pulli†
Online Learning Kernels
Jianping Fan Dept of Computer Science UNC-Charlotte
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Support Vector Machines Most of the slides were taken from:
Mathematical Foundations of BME Reza Shadmehr
Machine Learning in Practice Lecture 26
COSC 4335: Other Classification Techniques
Presented by: Chang Jia As for: Pattern Recognition
Support Vector Machines
Support Vector Machines and Kernels
COSC 4368 Machine Learning Organization
Hairong Qi, Gonzalez Family Professor
Support Vector Machines
Support Vector Machines 2
What is Artificial Intelligence?
Presentation transcript:

Max-margin Clustering: Detecting Margins from Projections of Points on Lines Raghuraman Gopalan1, and Jagan Sankaranarayanan2 1Center for Automation Research, University of Maryland, College Park, MD USA 2NEC Labs, Cupertino, CA USA E-mail: {raghuram,jagan}@umiacs.umd.edu

Problem Statement Given an unlabelled set of points forming k clusters, find a grouping with maximum separating margin among the clusters Prior work: (Mostly) Establish feedback between different label proposals, and run a supervised classifier on it Goal: To understand the relation between data points and margin regions by analyzing projections of data on lines

Two-cluster Problem Assumptions Linearly separable clusters Kernel trick for non-linear case No outliers in data (max margin exist only between clusters) Enforce global cluster balance Proposition 1 SI* exists ONLY on line segments in margin region that are perpendicular to the separating hyperplane Such line segments directly provide cluster groupings

Multi-cluster Problem SI* doesn’t exist Location information of projected points (SI) alone is insufficient to detect margins

The Role of Distance of Projection Proposition 2 For line intervals in margin region, perpendicular to the separating hyperplane, Proposition 3 For line intervals inside a cluster of length more than Mm, Proposition 4 An interval with SI having no projected points with distance of projection less than Dmin*, can lie only outside a cluster; where γ1 γ2 γ3 CL1 CL2 CL3 Defn: Dmin of a line interval is the minimum distance of projection of points in that interval. No outlier assumption: Max margin between points within a cluster

A Pair-wise Similarity Measure for Clustering f(xi,xj)=1, iff xi=xj f(xi,xj)<<1, iff xi and xj are from different clusters, and Intij is perpendicular to their separating hyperplane

Max-margin Clustering Algorithm Draw lines between all pairs of points Estimate the probability of presence of margins between a pair of points xi and xj by computing f(xi,xj) Perform global clustering using f between all point-pairs

Results 3D 2D

Clustering Detecting margin regions Summary Clustering Detecting margin regions Obtaining statistics of location and distance of projection of points that are specific to line segments in margin regions (Prop. 1 to 4) A pair-wise similarity measure to perform clustering, which avoids some optimization-related challenges prevalent in most existing methods