Improving Fuzzy Multilevel Graph Embedding through Feature Selection Technique Muhammad Muzzamil Luqman, Jean-Yves Ramel and Josep Lladós

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Mining customer ratings for product recommendation using the support vector machine and the latent class model William K. Cheung, James T. Kwok, Martin.
Employing structural representation for symbol detection, symbol spotting and indexation in line drawing document images Muhammad Muzzamil Luqman
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
Rule extraction in neural networks. A survey. Krzysztof Mossakowski Faculty of Mathematics and Information Science Warsaw University of Technology.
Pattern Recognition and Machine Learning
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.
Data preprocessing before classification In Kennedy et al.: “Solving data mining problems”
CSCI 347 / CS 4206: Data Mining Module 02: Input Topic 03: Attribute Characteristics.
Assessment. Schedule graph may be of help for selecting the best solution Best solution corresponds to a plateau before a high jump Solutions with very.
Pattern Recognition and Machine Learning
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Data classification based on tolerant rough set reporter: yanan yean.
Neural Networks Chapter Feed-Forward Neural Networks.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Feature Subset Selection using Minimum Cost Spanning Trees Mike Farah Supervisor: Dr. Sid Ray.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Machine Learning in Simulation-Based Analysis 1 Li-C. Wang, Malgorzata Marek-Sadowska University of California, Santa Barbara.
Evaluating Performance for Data Mining Techniques
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
Hubert CARDOTJY- RAMELRashid-Jalal QURESHI Université François Rabelais de Tours, Laboratoire d'Informatique 64, Avenue Jean Portalis, TOURS – France.
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
1 Template-Based Classification Method for Chinese Character Recognition Presenter: Tienwei Tsai Department of Informaiton Management, Chihlee Institute.
This week: overview on pattern recognition (related to machine learning)
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Unsupervised pattern recognition models for mixed feature-type.
Participation in the NIPS 2003 Challenge Theodor Mader ETH Zurich, Five Datasets were provided for experiments: ARCENE: cancer diagnosis.
Hierarchical Distributed Genetic Algorithm for Image Segmentation Hanchuan Peng, Fuhui Long*, Zheru Chi, and Wanshi Siu {fhlong, phc,
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering National Institute of Technology.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Data Mining Knowledge on rough set theory SUSHIL KUMAR SAHU.
1 Multiple Classifier Based on Fuzzy C-Means for a Flower Image Retrieval Keita Fukuda, Tetsuya Takiguchi, Yasuo Ariki Graduate School of Engineering,
Support Vector Machines Reading: Ben-Hur and Weston, “A User’s Guide to Support Vector Machines” (linked from class web page)
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
Kernels in Pattern Recognition. A Langur - Baboon Binary Problem m/2006/ /himplu s4.jpg … HA HA HA …
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Feature extraction using fuzzy complete linear discriminant analysis The reporter : Cui Yan
Support vector machine LING 572 Fei Xia Week 8: 2/23/2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A 1.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A self-organizing map for adaptive processing of structured.
Support Vector Machines and Kernel Methods for Co-Reference Resolution 2007 Summer Workshop on Human Language Technology Center for Language and Speech.
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
Hybrid Intelligent Systems for Network Security Lane Thames Georgia Institute of Technology Savannah, GA
2D-LDA: A statistical linear discriminant analysis for image matrix
A Kernel Approach for Learning From Almost Orthogonal Pattern * CIS 525 Class Presentation Professor: Slobodan Vucetic Presenter: Yilian Qin * B. Scholkopf.
Mete Ozay, Fatos T. Yarman Vural —Presented by Tianxiao Jiang
Nawanol Theera-Ampornpunt, Seong Gon Kim, Asish Ghoshal, Saurabh Bagchi, Ananth Grama, and Somali Chaterji Fast Training on Large Genomics Data using Distributed.
1 A Methodology for automatic retrieval of similarly shaped machinable components Mark Ascher - Dept of ECE.
Pattern Recognition. What is Pattern Recognition? Pattern recognition is a sub-topic of machine learning. PR is the science that concerns the description.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Fuzzy Multilevel Graph Embedding for Recognition, Indexing and Retrieval of Graphic Document Images presented by Muhammad Muzzamil LUQMAN
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
Feature learning for multivariate time series classification Mustafa Gokce Baydogan * George Runger * Eugene Tuv † * Arizona State University † Intel Corporation.
Big data classification using neural network
Guillaume-Alexandre Bilodeau
Learning Mid-Level Features For Recognition
Generating multidimensional embeddings based on fuzzy memberships
Hybrid Features based Gender Classification
Structure learning with deep autoencoders
Unsupervised Learning and Autoencoders
Liang Zheng and Yuzhong Qu
Nearest Neighbors CSC 576: Data Mining.
Ernest Valveny Computer Vision Center
Housam Babiker, Randy Goebel and Irene Cheng
Using Uneven Margins SVM and Perceptron for IE
Presentation transcript:

Improving Fuzzy Multilevel Graph Embedding through Feature Selection Technique Muhammad Muzzamil Luqman, Jean-Yves Ramel and Josep Lladós cvc.uab.es} Graph-TA workshop, 19th February 2013

2 Objectives of research  Problematic  Lack of efficient computational tools for graph based structural pattern recognition  Proposed solution  Transform graphs into numeric feature vectors and exploit computational strengths of state of the art statistical pattern recognition

3 Outline of presentation  Whats/whys of graph embedding?  Fuzzy multilevel graph embedding (FMGE)  Feature selection by ranking discriminatory features  Experimental results  Conclusions

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 4 Structural and statistical PR symbolic data structurenumeric feature vector Data structure Representational strength Fixed dimensionality Sensitivity to noise Efficient computational tools Pattern Recognition StructuralStatistical

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 5 Graph embedding (GEM) Structutal PR Expressive, convenient, powerful but computationally expensive representations Statistical PR Mathematically sound, mature, less expensive and computationally efficient models Graph embedding

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 6 Graph embedding (GEM)

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 7 Explicit and implicit GEM Explicit GEM  embeds each input graph into a numeric feature vector  provides more useful methods of GEM for PR  can be employed in a standard dot product for defining an implicit graph embedding function Implicit GEM  computes scalar product of two graphs in an implicitly existing vector space, by using graph kernels  does not permit all the operations that could be defined on vector spaces

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 8 State of the art (explicit GEM)  Graph probing based methods [Wiener, 1947] [Papadopoulos et al., 1999] [Gibert et al., 2011] [Sidere, 2012]  Spectral based graph embedding [Harchaoui, 2007] [Luo et al., 2003] [Robleskelly and Hancock, 2007]  Dissimilarity based graph embedding [Pekalska et al., 2005] [Ferrer et al., 2008] [Riesen, 2010] [Bunke et al., 2011]

9 Outline of presentation  Whats/whys of graph embedding?  Fuzzy multilevel graph embedding (FMGE)  Feature selection by ranking discriminatory features  Experimental results  Conclusions

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 10 FMGE  Fuzzy Multilevel Graph Embedding (FMGE) [Luqman et al., 2012]  Graph probing based explicit graph embedding method [Luqman et al., Fuzzy multilevel graph embedding, Pattern Recognition, 2012.]

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 11 FMGE  Input : Collection of attributed graphs  Output : Equal-size numeric feature vector for each input graph

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 12 Extracting details from graph  Multilevel analysis of graph Graph order Graph size Node degree Homogeneity of subgraphs in graph Node attributes Edge attributes Graph Level Information [macro details] Elementary Level Information [micro details] Structural Level Information [intermediate details]  Numeric feature vector embeds a graph, encoding: Numeric information by fuzzy histograms Symbolic information by crisp histograms

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 13 Homogeneity of subgraphs in a graph  Node-resemblance for an edge  Edge-resemblance for a node

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 14 Homogeneity of subgraphs in a graph a1b1a1b1 a2b2a2b2 a b  Node-resemblance for an edge  Edge-resemblance for a node

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 15 Homogeneity of subgraphs in a graph  Node-resemblance for an edge  Edge-resemblance for a node a1b1a1b1 a2b2a2b2 a b a3b3a3b3

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 16 Encoding extracted information  Unsupervised learning phase  Graph embedding phase

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 17 Unsupervised learning phase of FMGE

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 18 Unsupervised learning phase of FMGE  First fuzzy interval (- , - , …, …)  Last fuzzy interval (…, …, ,  )

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 19 Graph embedding phase of FMGE  Numeric information embedded by fuzzy histograms  Symbolic information embedded by crisp histograms

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 20 Fuzzy Structural Multilevel Feature Vector Graph order Graph size Fuzzy histograms of numeric node attributes Crisp histograms of symbolic node attributes Fuzzy histograms of numeric edge attributes Crisp histograms of symbolic edge attributes Fuzzy histogram of node degrees Fuzzy histograms of numeric resemblance attributes Crisp histograms of symbolic resemblanc e attributes Graph Level Information [macro details] Elementary Level Information [micro details] Structural Level Information [intermediate details]  Feature vector of FMGE

21 Outline of presentation  Whats/whys of graph embedding?  Fuzzy multilevel graph embedding (FMGE)  Feature selection by ranking discriminatory features  Experimental results  Conclusions

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 22 Why feature Selection? Large number of features enable FMGE to achieve generalization to diverse graphs in an unsupervised fashion BUT also results into high dimensionality and sparsity of feature vector!

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 23 Ranking discriminatory features  Relief algorithm based feature selection [Kira et al., 1992] For each instance of a given feature, the near-hit (closest value among elements of same class) and the near-miss (closest value among element of other classes) are computed A weight w f, is calculated for every feature in terms of the distances of each sample x, to its near-hit and near miss  n top-ranked (high discriminatory) features are selected

24 Outline of presentation  Whats/whys of graph embedding?  Fuzzy multilevel graph embedding (FMGE)  Feature selection by ranking discriminatory features  Experimental results  Conclusions and future research challenges

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 25 Graph database  IAM graph database [Riesen et al., 2008]

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 26 Parameter validation  Two parameters validated on “validation set”: 1.Number of fuzzy intervals for embedding numeric information o 2 to 25 (in steps of 1) 2.Selection of n top-ranked features o All subsets of high to low ranked features (i.e. top-1, top-2, top-3...)

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 27 Parameter validation Validation results showing the classification rates obtained for different subsets of n top-ranked features for the 25 configurations of FMGE. The horizontal axis gives top-n features (sorted high to low ranked by Relief algorithm).

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 28 Experimental results Results on test sets - IAM graph database. “CR” is the classification rate (%) obtained by k-nn classifier and “DIM” is the dimensionality of feature vector

29 Outline of presentation  Whats/whys of graph embedding?  Fuzzy multilevel graph embedding (FMGE)  Feature selection by ranking discriminatory features  Experimental results  Conclusions

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 30 Conclusions  Feature selection successfully improves the performance of FMGE Low dimensionality Reduced compuation time for classification Improved classification rate  Provides a deep insight into the groups of features that are discriminatory and the features that are not very useful for a graph dataset  More compact and discriminatory feature vector representation of graphs

What is graph embedding? Fuzzy Multilevel Graph Embedding Feature selection Experimentation Conclusions 31 Future research challenges  Other sophisticated feature selection techniques  Add more topological information to FMGE

Improving Fuzzy Multilevel Graph Embedding through Feature Selection Technique Muhammad Muzzamil Luqman, Jean-Yves Ramel and Josep Lladós cvc.uab.es} Graph-TA workshop, 19th February 2013