A Binary Linear Programming Formulation of the Graph Edit Distance Presented by Shihao Ji Duke University Machine Learning Group July 17, 2006 Authors:

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Towards a Quadratic Time Approximation of Graph Edit Distance Fischer, A., Suen, C., Frinken, V., Riesen, K., Bunke, H. Contents Introduction Graph edit.
Active Appearance Models
1/22 Worst and Best-Case Coverage in Sensor Networks Seapahn Meguerdichian, Farinaz Koushanfar, Miodrag Potkonjak, and Mani Srivastava IEEE TRANSACTIONS.
ECG Signal processing (2)
Orthogonal Drawing Kees Visser. Overview  Introduction  Orthogonal representation  Flow network  Bend optimal drawing.
Improved Algorithms for Inferring the Minimum Mosaic of a Set of Recombinants Yufeng Wu and Dan Gusfield UC Davis CPM 2007.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Fast Algorithms For Hierarchical Range Histogram Constructions
Similarity Search on Bregman Divergence, Towards Non- Metric Indexing Zhenjie Zhang, Beng Chi Ooi, Srinivasan Parthasarathy, Anthony K. H. Tung.
Data Mining Classification: Alternative Techniques
Pattern Recognition and Machine Learning
Support Vector Machines
Exploiting Sparse Markov and Covariance Structure in Multiresolution Models Presenter: Zhe Chen ECE / CMR Tennessee Technological University October 22,
Coupling-Aware Length-Ratio- Matching Routing for Capacitor Arrays in Analog Integrated Circuits Kuan-Hsien Ho, Hung-Chih Ou, Yao-Wen Chang and Hui-Fang.
1 The Monte Carlo method. 2 (0,0) (1,1) (-1,-1) (-1,1) (1,-1) 1 Z= 1 If  X 2 +Y 2  1 0 o/w (X,Y) is a point chosen uniformly at random in a 2  2 square.
Convex Position Estimation in Wireless Sensor Networks
1 Efficient Subgraph Search over Large Uncertain Graphs Ye Yuan 1, Guoren Wang 1, Haixun Wang 2, Lei Chen 3 1. Northeastern University, China 2. Microsoft.
Second order cone programming approaches for handing missing and uncertain data P. K. Shivaswamy, C. Bhattacharyya and A. J. Smola Discussion led by Qi.
Lecture 21: Spectral Clustering
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Discrete random variables Probability mass function Distribution function (Secs )
Supervised Distance Metric Learning Presented at CMU’s Computer Vision Misc-Read Reading Group May 9, 2007 by Tomasz Malisiewicz.
Nearest Neighbor Retrieval Using Distance-Based Hashing Michalis Potamias and Panagiotis Papapetrou supervised by Prof George Kollios A method is proposed.
Visual Recognition Tutorial
A Study of the Relationship between SVM and Gabriel Graph ZHANG Wan and Irwin King, Multimedia Information Processing Laboratory, Department of Computer.
Small Subgraphs in Random Graphs and the Power of Multiple Choices The Online Case Torsten Mütze, ETH Zürich Joint work with Reto Spöhel and Henning Thomas.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Internet Traffic Engineering by Optimizing OSPF Weights Bernard Fortz (Universit é Libre de Bruxelles) Mikkel Thorup (AT&T Labs-Research) Presented by.
Efficient Gathering of Correlated Data in Sensor Networks
Preserving Link Privacy in Social Network Based Systems Prateek Mittal University of California, Berkeley Charalampos Papamanthou.
Network Aware Resource Allocation in Distributed Clouds.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Yaomin Jin Design of Experiments Morris Method.
Non-Parameter Estimation 主講人:虞台文. Contents Introduction Parzen Windows k n -Nearest-Neighbor Estimation Classification Techiques – The Nearest-Neighbor.
Swarup Acharya Phillip B. Gibbons Viswanath Poosala Sridhar Ramaswamy Presented By Vinay Hoskere.
Max-Margin Classification of Data with Absent Features Presented by Chunping Wang Machine Learning Group, Duke University July 3, 2008 by Chechik, Heitz,
Nearest Neighbor Queries Chris Buzzerd, Dave Boerner, and Kevin Stewart.
Optimal Bayes Classification
Mathematical Programming in Data Mining Author: O. L. Mangasarian Advisor: Dr. Hsu Graduate: Yan-Cheng Lin.
CSE 185 Introduction to Computer Vision Face Recognition.
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Stabbing balls and simplifying proteins Ovidiu Daescu and Jun Luo Department of Computer Science University of Texas at Dallas Richardson, TX
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
An Introduction to Support Vector Machine (SVM)
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Efficient Optimal Linear Boosting of a Pair of Classifiers.
Tony Jebara, Columbia University Advanced Machine Learning & Perception Instructor: Tony Jebara.
Efficient Computing k-Coverage Paths in Multihop Wireless Sensor Networks XuFei Mao, ShaoJie Tang, and Xiang-Yang Li Dept. of Computer Science, Illinois.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Occam’s Razor No Free Lunch Theorem Minimum.
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
Lower Bounds for Embedding Edit Distance into Normed Spaces A. Andoni, M. Deza, A. Gupta, P. Indyk, S. Raskhodnikova.
New Algorithms for Efficient High-Dimensional Nonparametric Classification Ting Liu, Andrew W. Moore, and Alexander Gray.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Optimal Relay Placement for Indoor Sensor Networks Cuiyao Xue †, Yanmin Zhu †, Lei Ni †, Minglu Li †, Bo Li ‡ † Shanghai Jiao Tong University ‡ HK University.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Spectral Methods for Dimensionality
Outline Introduction State-of-the-art solutions
Approximating the MST Weight in Sublinear Time
A New Support Vector Finder Method Based on Triangular Calculations
Probabilistic Data Management
Learning with information of features
Lecture 16: Earth-Mover Distance
Efficient Subgraph Similarity All-Matching
Using Manifold Structure for Partially Labeled Classification
Machine Learning – a Probabilistic Perspective
Presentation transcript:

A Binary Linear Programming Formulation of the Graph Edit Distance Presented by Shihao Ji Duke University Machine Learning Group July 17, 2006 Authors: Derek Justice & Alfred Hero (PAMI 2006)

Introduction to Graph Matching Proposed Method (binary linear program) Experimental Results (chemical graph matching) Outline

Graph Matching Objective: matching a sample input graph to a database of known prototype graphs.

Graph Matching (cont’d) A real example: face identification

Graph Matching (cont’d) Key issues: (1) representative graph generation (a) facial graph representations (b) chemical graphs

 Maximum Common Subgraph (MCS)  Graph Edit Distance (GED) Enumeration procedures (for small graphs) Probabilistic models (MAP estimates) Binary Linear Programming (BLP) Graph Matching (cont’d) Key issues: (2) graph distance metrics

Basic idea: define graph edit operations (such as insertion or deletion or relabeling of a vertex) along with costs associated with each operation. The GED between two graphs is the cost associated with the least costly series of edit operations needed to make the two graph isomorphic. Key issues: how to find the least costly series of edit operations? how to define edit costs? Graph Edit Distance

Graph Edit Distance (cont’d) How to compute the distance between G 0 and G 1 ? Edit Grid

Isomorphisms of G 0 on the edit grid State Vectors Graph Edit Distance (cont’d) standard placement

Definition: (if the cost function c is a metric) Objective function: binary linear program (NP-hard!!!) Graph Edit Distance (Cont’d)

Lower bound: linear program (polynomial time) Upper bound: assignment problem (polynomial time) Graph Edit Distance (cont’d)

Edit Cost Selection Goal: suppose there is a set of prototype graphs {G i } i=1,…,N and we classify a sample graph G 0 by a nearest neighbor classifier in the metric space defined by the graph edit distance. Prior informaiton: the prototypes should be roughly uniformly distributed in the metric space of graphs. Why: it minimizes the worst case classification error since it equalizes the probability of error under a nearest neighbor classifier.

Edit Cost Selection (cont’d) Objective: minimize the variance of pairwise NN distances Define unit cost function, i.e., c(0,1)=1, c( ,  )=1, c( ,  )=0 Solve the BLP (with unit cost) and find the NN pair Construct H k,i = the number of i th edit operation for the k th NN pair Objective function: (convex optimization)

Experimental Results Chemical Graph Recognition

1. edge edit 2. vertex deletion 3. vertex insertion 4. vertex relabeling 5. random (a) original graph Experiments Results (cont’d) (b) example perturbed graphs

Experiments Results (cont’d) Optimal Edit Costs

A: GEDo B: GEDu C: MCS1 D: MCS2 Experiments Results (cont’d) Classification Results

Present a binary linear programming formulation of the graph edit distance; Offer a minimum variance method for choosing a cost metric; Demonstrate the utility of the new method in the context of a chemical graph recognition. Conclusion