Date: 2011/1/11 Advisor: Dr. Koh. Jia-Ling Speaker: Lin, Yi-Jhen Mr. KNN: Soft Relevance for Multi-label Classification (CIKM’10) 1.

Slides:



Advertisements
Similar presentations
Learning Trajectory Patterns by Clustering: Comparative Evaluation Group D.
Advertisements

Multi-label Classification without Multi-label Cost - Multi-label Random Decision Tree Classifier 1.IBM Research – China 2.IBM T.J.Watson Research Center.
Decision trees for hierarchical multilabel classification A case study in functional genomics.
Data Mining Classification: Alternative Techniques
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
Data Mining Classification: Alternative Techniques
Christoph F. Eick Questions and Topics Review Nov. 22, Assume you have to do feature selection for a classification task. What are the characteristics.
Bring Order to Your Photos: Event-Driven Classification of Flickr Images Based on Social Knowledge Date: 2011/11/21 Source: Claudiu S. Firan (CIKM’10)
Patch to the Future: Unsupervised Visual Prediction
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
Classification and Decision Boundaries
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Fuzzy Support Vector Machines (FSVMs) Weijia Wang, Huanren Zhang, Vijendra Purohit, Aditi Gupta.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
2D1431 Machine Learning Boosting.
Lesson 8: Machine Learning (and the Legionella as a case study) Biological Sequences Analysis, MTA.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Ensemble Learning: An Introduction
Margin Based Sample Weighting for Stable Feature Selection Yue Han, Lei Yu State University of New York at Binghamton.
CHURN PREDICTION MODEL IN RETAIL BANKING USING FUZZY C- MEANS CLUSTERING Džulijana Popović Consumer Finance, Zagrebačka banka d.d. Consumer Finance, Zagrebačka.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
A k-Nearest Neighbor Based Algorithm for Multi-Label Classification Min-Ling Zhang
Tag Clouds Revisited Date : 2011/12/12 Source : CIKM’11 Speaker : I- Chih Chiu Advisor : Dr. Koh. Jia-ling 1.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Probability-based imputation method for fuzzy cluster analysis of gene expression microarray data Thanh Le, Tom Altman and Katheleen Gardiner University.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Presenter : Chien-Shing Chen Author: Tie-Yan.
Group Recommendations with Rank Aggregation and Collaborative Filtering Linas Baltrunas, Tadas Makcinskas, Francesco Ricci Free University of Bozen-Bolzano.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
Exploring Online Social Activities for Adaptive Search Personalization CIKM’10 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
Chapter 8 The k-Means Algorithm and Genetic Algorithm.
Special topics on text mining [ Part I: text classification ] Hugo Jair Escalante, Aurelio Lopez, Manuel Montes and Luis Villaseñor.
MINING MULTI-LABEL DATA BY GRIGORIOS TSOUMAKAS, IOANNIS KATAKIS, AND IOANNIS VLAHAVAS Published on July, 7, 2010 Team Members: Kristopher Tadlock, Jimmy.
1 Multiple Classifier Based on Fuzzy C-Means for a Flower Image Retrieval Keita Fukuda, Tetsuya Takiguchi, Yasuo Ariki Graduate School of Engineering,
Today Ensemble Methods. Recap of the course. Classifier Fusion
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Haojun Sun,ShengruiWang*,Qingshan Jiang Received 16 December 2002; received in revised form 29 March 2004; accepted 29 March 2004 Presenter Chia-Cheng.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Neural Text Categorizer for Exclusive Text Categorization Journal of Information Processing Systems, Vol.4, No.2, June 2008 Taeho Jo* 報告者 : 林昱志.
Prototype Classification Methods Fu Chang Institute of Information Science Academia Sinica ext. 1819
A new initialization method for Fuzzy C-Means using Fuzzy Subtractive Clustering Thanh Le, Tom Altman University of Colorado Denver July 19, 2011.
CS378 Final Project The Netflix Data Set Class Project Ideas and Guidelines.
Text Categorization by Boosting Automatically Extracted Concepts Lijuan Cai and Tommas Hofmann Department of Computer Science, Brown University SIGIR 2003.
Xiangnan Kong,Philip S. Yu An Ensemble-based Approach to Fast Classification of Multi-label Data Streams Dept. of Computer Science University of Illinois.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
{ Adaptive Relevance Feedback in Information Retrieval Yuanhua Lv and ChengXiang Zhai (CIKM ‘09) Date: 2010/10/12 Advisor: Dr. Koh, Jia-Ling Speaker: Lin,
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Fuzzy Pattern Recognition. Overview of Pattern Recognition Pattern Recognition Procedure Feature Extraction Feature Reduction Classification (supervised)
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Decision trees for hierarchical multi-label classification.
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Efficient Similarity Search : Arbitrary Similarity Measures, Arbitrary Composition Date: 2011/10/31 Source: Dustin Lange et. al (CIKM’11) Speaker:Chiang,guang-ting.
Fuzzy Logic in Pattern Recognition
Support Feature Machine for DNA microarray data
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
KDD 2004: Adversarial Classification
Basic machine learning background with Python scikit-learn
Cost-Sensitive Learning
Machine Learning Week 1.
Cost-Sensitive Learning
COSC 4335: Other Classification Techniques
A Unified View of Multi-Label Performance Measures
Feature Selection for Ranking
Concave Minimization for Support Vector Machine Classifiers
Physics-guided machine learning for milling stability:
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

Date: 2011/1/11 Advisor: Dr. Koh. Jia-Ling Speaker: Lin, Yi-Jhen Mr. KNN: Soft Relevance for Multi-label Classification (CIKM’10) 1

Preview Introduction Related Work Problem Transformation Methods Algorithm Adaptation Methods The ML-KNN (Multi-Label K Nearest Neighbor) Method Mr. KNN: Method Description Experimental Results Conclusion 2

Introduction Multi-label learning refers to learning tasks where each instance is assigned to one or more classes(labels). Multi-label classification is drawing increasing interest and emerging as a fast-growing research field. 3

Preview Introduction Related Work Problem Transformation Methods Algorithm Adaptation Methods The ML-KNN (Multi-Label K Nearest Neighbor) Method Mr. KNN: Method Description Experimental Results Conclusion 4

Related Work – Problem Transformation Methods 5 Freq.=(3, 5, 2, 4, 4) Select-max Select-min

Related Work – Problem Transformation Methods Another popular strategy is so-called binary relevance, which converts the problem into multiple single-label binary classification problems. Multi-label instances are forced into one single category without considering distribution. 6

Related Work – Algorithm Adaptation Methods Algorithm adaption methods modify standard single-label learning algorithm for multi-label classification. 7 single-label learning multi-label learning Algorithm Adaptation Decision treesadapted C4.5 Allowing leaves of a tree to represent a set of labels AdaBoost.MH Maintain a set of weights as a distribution over both training examples and associated labels SVMSVM-like optimization strategy Be treated as a ranking problem and a linear model that minimizes a ranking loss and maximizes a margin is developed

Related Work – The ML-KNN Method 8

9

Mr. KNN: Method Description Mr.KNN consists of two components Soft Relevance A modified fuzzy c-means (FCM)-based approach to produce soft relevance Mr.KNN: Volting-Margin Ratio Method A modified kNN for multi-label classification Fuzzy c-means algorithm (similar with k-means algorithm) In fuzzy clustering, each point has a degree of belonging to clusters, as in fuzzy logic, rather than belonging completely to just one cluster.fuzzy clusteringfuzzy logic We adapt the FCM algorithm to yield a soft relevance value for each instance with respect to each label 10

Soft Relevance 11

Soft Relevance 12 For 5-class multi-label classification c1~c5 If an instance xi belongs to class c1, c2, c4 Then u3i = u5i = 0 And u1i + u2i + u4i = 1

Soft Relevance To find the membership values, we minimize the cost function Jm with the constrains in previous slide, this leads to the following Lagrangian function: 13 Can be solved by the Gauss-Newton method

Mr.KNN: Voting-Margin Ratio Method 14 1

Mr.KNN: Voting-Margin Ratio Method To determine the optimal values of f in Minkowski distance and K in kNN, we introduce a new evaluation function, which is motivated by the margin concept (voting margin) Consider a 5-class learning problem with an instance belonging to two class labels: labels 2 and 3 The instance: the plus inside a circle A circle represents a voting value for the label marked by the number inside a circle 15 Correct voting Smaller margin Correct voting larger margin True label 3 is lower than false labels 4 & 5

Mr.KNN: Voting-Margin Ratio Method voting margin Ti : true label set Fi : false label set Our goal is to seek the combination of f and k that maximizes the average voting margin ratio The overall learning method for multi-label learning is called voting Margin Ration kNN, or Mr.KNN 16

Mr.KNN: Voting-Margin Ratio Method Mr.KNN consists of two steps: training and test. The procedures are as follow 17

Mr.KNN: Voting-Margin Ratio Method Mr.KNN consists of two steps: training and test. The procedures are as follow 18

Experimental Results – Data Description Three multi-label datasets are tested in this study Predict gene functions of yeast Detection of emotions in music Semantic scene classification 19

Experimental Results – Evaluation Criteria 20

Experimental Results – Evaluation Criteria Also use NDCG (normalized discounted cumulative gain) to evaluate the final ranking of labels for each instance For each instance, a label will receive a voting score Ideally, these true labels will rank higher than false labels The NDCG of a ranking list of labels at position n is 21

Experimental Results For each dataset select the f in Minkowski distance form 1, 2, 4, 6 K in kNN from 10, 15, 20, 25, 30, 35, 40, 45 Total 32 combinations of (f, k) 22

23

24

25

Conclusion We introduce the soft relevance strategy, in which each instance is assigned a relevance score with respect to a label Furthermore, it is used as a voting factor in a modified kNN algorithm Evaluated over three multi-label datasets, the proposed method outperforms ML-KNN 26