Cost Sensitive Evaluation Measures for F-term Classification

Slides:



Advertisements
Similar presentations
Developing Science Skills. Preparing for Tasks Level DLevel ELevel F individually or in small groups will identify two or three questions to investigate.
Advertisements

Mustafa Cayci INFS 795 An Evaluation on Feature Selection for Text Clustering.
~1~ Infocom’04 Mar. 10th On Finding Disjoint Paths in Single and Dual Link Cost Networks Chunming Qiao* LANDER, CSE Department SUNY at Buffalo *Collaborators:
Shape Classification Based on Skeleton Path Similarity Xingwei Yang, Xiang Bai, Deguang Yu, and Longin Jan Latecki.
Precision and Recall.
Modern Information Retrieval Chapter 2 Modeling. Can keywords be used to represent a document or a query? keywords as query and matching as query processing.
Semantic distance in WordNet: An experimental, application-oriented evaluation of five measures Presenter: Cosmin Adrian Bejan Alexander Budanitsky and.
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
A Comparative Study on Feature Selection in Text Categorization (Proc. 14th International Conference on Machine Learning – 1997) Paper By: Yiming Yang,
Selecting Informative Genes with Parallel Genetic Algorithms Deodatta Bhoite Prashant Jain.
Ordinal Decision Trees Qinghua Hu Harbin Institute of Technology
A Light-weight Approach to Coreference Resolution for Named Entities in Text Marin Dimitrov Ontotext Lab, Sirma AI Kalina Bontcheva, Hamish Cunningham,
Title Extraction from Bodies of HTML Documents and its Application to Web Page Retrieval Microsoft Research Asia Yunhua Hu, Guomao Xin, Ruihua Song, Guoping.
Authors: Ting Wang, Yaoyong Li, Kalina Bontcheva, Hamish Cunningham, Ji Wang Presented by: Khalifeh Al-Jadda Automatic Extraction of Hierarchical Relations.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
IR Evaluation Evaluate what? –user satisfaction on specific task –speed –presentation (interface) issue –etc. My focus today: –comparative performance.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
1 Efficient Search Ranking in Social Network ACM CIKM2007 Monique V. Vieira, Bruno M. Fonseca, Rodrigo Damazio, Paulo B. Golgher, Davi de Castro Reis,
Experiments of Opinion Analysis On MPQA and NTCIR-6 Yaoyong Li, Kalina Bontcheva, Hamish Cunningham Department of Computer Science University of Sheffield.
Context-Sensitive Information Retrieval Using Implicit Feedback Xuehua Shen : department of Computer Science University of Illinois at Urbana-Champaign.
Assessing The Retrieval A.I Lab 박동훈. Contents 4.1 Personal Assessment of Relevance 4.2 Extending the Dialog with RelFbk 4.3 Aggregated Assessment.
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
Semi-Supervised Time Series Classification & DTW-D REPORTED BY WANG YAWEN.
A genetic approach to the automatic clustering problem Author : Lin Yu Tseng Shiueng Bien Yang Graduate : Chien-Ming Hsiao.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
1 CS 430 / INFO 430 Information Retrieval Lecture 8 Evaluation of Retrieval Effectiveness 1.
Multilingual Information Retrieval using GHSOM Hsin-Chang Yang Associate Professor Department of Information Management National University of Kaohsiung.
Semi-Supervised Learning with Graph Transduction Project 2 Due Nov. 7, 2012, 10pm EST Class presentation on Nov. 12, 2012.
1 Context-Aware Ranking in Web Search (SIGIR 10’) Biao Xiang, Daxin Jiang, Jian Pei, Xiaohui Sun, Enhong Chen, Hang Li 2010/10/26.
LEARNING IN A PAIRWISE TERM-TERM PROXIMITY FRAMEWORK FOR INFORMATION RETRIEVAL Ronan Cummins, Colm O’Riordan (SIGIR’09) Speaker : Yi-Ling Tai Date : 2010/03/15.
ICS 353: Design and Analysis of Algorithms NP-Complete Problems King Fahd University of Petroleum & Minerals Information & Computer Science Department.
Data Structures and Algorithm Analysis Lecture 5
Clustering CSC 600: Data Mining Class 21.
Algorithms for Large Data Sets
Evaluating Classifiers
Sofus A. Macskassy Fetch Technologies
Walid Magdy Gareth Jones
Evaluation of IR Systems
Modeling Arithmetic, Computation, and Languages
Lesson Objectives Aims Understand the following “standard algorithms”:
Dipartimento di Ingegneria «Enzo Ferrari»,
COMBINED UNSUPERVISED AND SEMI-SUPERVISED LEARNING FOR DATA CLASSIFICATION Fabricio Aparecido Breve, Daniel Carlos Guimarães Pedronette State University.
Our Data Science Roadmap
CSE 4705 Artificial Intelligence
Location Recommendation — for Out-of-Town Users in Location-Based Social Network Yina Meng.
A Survey of Program Slicing Techniques: Section 4
Applying Key Phrase Extraction to aid Invalidity Search
GODFREY HODGSON HOLMES TARCA
Party-by-Night Problem
ICS 353: Design and Analysis of Algorithms
Critical Issues with Respect to Clustering
Ways to show a function Four ways to display/write a function
Department of Computer Science University of York
Text Categorization Document classification categorizes documents into one or more classes which is useful in Information Retrieval (IR). IR is the task.
Thesis Proposal Title Student Name Registration No. Supervisor Name
INFORMATION RETRIEVAL TECHNIQUES BY DR. ADNAN ABID
Walking the Interactome for Prioritization of Candidate Disease Genes
Automatic Extraction of Hierarchical Relations from Text
SVM Based Learning System for F-term Patent Classification
Perceptron Learning for Chinese Word Segmentation
Combining Keyword and Semantic Search for Best Effort Information Retrieval  Andrew Zitzelberger 1.
Feature Selection for Ranking
Using Uneven Margins SVM and Perceptron for IE
CSE4334/5334 Data Mining Lecture 7: Classification (4)
Roc curves By Vittoria Cozza, matr
Hierarchical, Perceptron-like Learning for OBIE
Our Data Science Roadmap
SVMs for Document Ranking
Precision and Recall.
Presentation transcript:

Cost Sensitive Evaluation Measures for F-term Classification Yaoyong Li, Kalina Bontcheva, Hamish Cunningham Department of Computer Science University of Sheffield {yaoyong,kalina,hamish}@dcs.shef.ac.uk http://gate.ac.uk/ http://nlp.shef.ac.uk/

Outline Measures for closeness of two concepts in taxonomy. Cost sensitive measures for document classification and information retrieval. Applying the new measures on the submitted runs of the F-term patent classification sub-task of NTCIR-6. 2(12)

Hierarchical Classification Class labels are organised in a hierarchical fashion. Cost of misclassification is dependent upon closeness of two classes. It’s better for evaluation measure to be closeness sensitive. 3(12)

Closeness of Concepts in Taxonomy Agirre and Rigau (1996) proposed the following criteria for measuring the closeness of two concepts in a taxonomy. C1. Be dependent on length of the shortest path connecting the two concepts. C2. Concepts in deeper part of taxonomy should be closer. C3. Concepts in denser part should be relatively closer. C4. Be independent of the number of concepts in the graph. 4(12)

Three Measures of Closeness Three measures and their compatibilities with the four criteria. C1 C2 C3 C4 Distance based √ Learning accuracy BDM 5(12)

Conventional Precision Given a particular class X and a subset S of test example with the predicted class labels, the conventional precision P =nmatch/|S| where nmatch is the number of the exactly matched examples. 6(12)

Cost Sensitive Precision Having the same assumption as in last slide, and also assume that S0 consists of the examples in test set S which belong to class X, and Ye is the predicted label of example e, then the cost sensitive precision is as P =∑e S0(CL(X,Ye)/|S| where CL(X,Ye) denotes the closeness of two classes X and Ye. 7(12)

Cost Sensitive Measures Three commonly used measures in document classification and information retrieval A-Precision, R-Precision, and F-measure. We can obtain the cost sensitive version of the three measures in a straight way, by using the cost sensitive precision defined in last slide. 8(12)

Measures for F-term classification F-term patent classification consists of hierarchical classification problems. We defined two cost sensitive measures for F-term classification, both based on the BDM scores In generous way: BDM_high. In conservative way: BDM_low. 9(12)

Results of F-term Classification Differences of the A-Precision between the BDM measures and the conventional one for all the official submitted runs of the NTCIR-6 F-term Classification sub-task. 10(12)

Effect on Ranking of Submitted Runs The Kendall’s tau for the pairs of rankings of all submitted runs using BDM score and binary score, respectively. A-Precision R-Precision F-measure BDM_high 0.887 0.861 0.614 BDM_low 0.912 0.848 0.752 11(12)

Conclusions Extend the measures for document classification to take into account the closeness of two concepts in a taxonomy. Apply the new measures to the submitted runs of the NTCIR-6 F-term patent classification Demonstrated the difference learning strategies better than conventional ones. Does have effect on the rankings. 12(12)