You Can’t Beat Frequency (Unless You Use Linguistic Knowledge) – A Qualitative Evaluation of Association Measures for Collocation and Term Extraction Joachim.

Slides:



Advertisements
Similar presentations
Natural Language Processing COLLOCATIONS Updated 16/11/2005.
Advertisements

Outline What is a collocation?
Introduction Goal of this work is to better understand Guelph’s 2007 LibQUAL+ comments (in aggregate), within the context of the quantitative findings.
Chapter 1 Displaying the Order in a Group of Numbers
Chinese Word Segmentation Method for Domain-Special Machine Translation Su Chen; Zhang Yujie; Guo Zhen; Xu Jin’an Beijing Jiaotong University.
January 12, Statistical NLP: Lecture 2 Introduction to Statistical NLP.
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
NORMAL CURVE Needed for inferential statistics. Find percentile ranks without knowing all the scores in the distribution. Determine probabilities.
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
Inferences About Means of Single Samples Chapter 10 Homework: 1-6.
1 LM Approaches to Filtering Richard Schwartz, BBN LM/IR ARDA 2002 September 11-12, 2002 UMASS.
Normal Distributions What is a Normal Distribution? Why are Many Variables Normally Distributed? Why are Many Variables Normally Distributed? How Are Normal.
Statistics for the Social Sciences Psychology 340 Fall 2006 Hypothesis testing.
Chapter 10 Hypothesis Testing
IENG 486 Statistical Quality & Process Control
Automatically Constructing a Dictionary for Information Extraction Tasks Ellen Riloff Proceedings of the 11 th National Conference on Artificial Intelligence,
Query Operations: Automatic Global Analysis. Motivation Methods of local analysis extract information from local set of documents retrieved to expand.
An Automatic Segmentation Method Combined with Length Descending and String Frequency Statistics for Chinese Shaohua Jiang, Yanzhong Dang Institute of.
Measuring the Unobservable
Testing Hypotheses I Lesson 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics n Inferential Statistics.
Natural Language Processing Spring 2007 V. “Juggy” Jagannathan.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Confidence Interval Estimation.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
Analyzing and Interpreting Quantitative Data
Scalable Statistical Bug Isolation Authors: B. Liblit, M. Naik, A.X. Zheng, A. Aiken, M. I. Jordan Presented by S. Li.
Part III Taking Chances for Fun and Profit Chapter 8 Are Your Curves Normal? Probability and Why it Counts.
Identifying Disease Diagnosis Factors by Proximity-based Mining of Medical Texts Rey-Long Liu *, Shu-Yu Tung, and Yun-Ling Lu * Dept. of Medical Informatics.
CONFIDENCE INTERVAL It is the interval or range of values which most likely encompasses the true population value. It is the extent that a particular.
Chapter 6 USING PROBABILITY TO MAKE DECISIONS ABOUT DATA.
Measures of central tendency are statistics that express the most typical or average scores in a distribution These measures are: The Mode The Median.
Sampling Distribution Tripthi M. Mathew, MD, MPH.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Pattern Discovery of Fuzzy Time Series for Financial Prediction -IEEE Transaction of Knowledge and Data Engineering Presented by Hong Yancheng For COMP630P,
N-Gram-based Dynamic Web Page Defacement Validation Woonyon Kim Aug. 23, 2004 NSRI, Korea.
Quick and Simple Statistics Peter Kasper. Basic Concepts Variables & Distributions Variables & Distributions Mean & Standard Deviation Mean & Standard.
1 Statistical NLP: Lecture 7 Collocations. 2 Introduction 4 Collocations are characterized by limited compositionality. 4 Large overlap between the concepts.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Methods for Automatic Evaluation of Sentence Extract Summaries * G.Ravindra +, N.Balakrishnan +, K.R.Ramakrishnan * Supercomputer Education & Research.
©2003 Paula Matuszek CSC 9010: Text Mining Applications Dr. Paula Matuszek (610)
Extracting Keyphrases from Books using Language Modeling Approaches Rohini U AOL India R&D, Bangalore India Bangalore
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
Finding frequent and interesting triples in text Janez Brank, Dunja Mladenić, Marko Grobelnik Jožef Stefan Institute, Ljubljana, Slovenia.
Chapter 6: Analyzing and Interpreting Quantitative Data
Exploring in the Weblog Space by Detecting Informative and Affective Articles Xiaochuan Ni, Gui-Rong Xue, Xiao Ling, Yong Yu Shanghai Jiao-Tong University.
A DYNAMIC APPROACH TO THE SELECTION OF HIGH ORDER N-GRAMS IN PHONOTACTIC LANGUAGE RECOGNITION Mikel Penagarikano, Amparo Varona, Luis Javier Rodriguez-
Chapter 10 Copyright © Allyn & Bacon 2008 This multimedia product and its contents are protected under copyright law. The following are prohibited by law:
Inferential Statistics Introduction. If both variables are categorical, build tables... Convention: Each value of the independent (causal) variable has.
Sociological Research Methods Review. 10 Review Questions 1.Name three types of research 1.Quantitative (Deductive) 2.Qualitative (Inductive) 3.Descriptive.
Comparing Word Relatedness Measures Based on Google n-grams Aminul ISLAM, Evangelos MILIOS, Vlado KEŠELJ Faculty of Computer Science Dalhousie University,
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Learning Taxonomic Relations from Heterogeneous Evidence Philipp Cimiano Aleksander Pivk Lars Schmidt-Thieme Steffen Staab (ECAI 2004)
Query Suggestions in the Absence of Query Logs Sumit Bhatia, Debapriyo Majumdar,Prasenjit Mitra SIGIR’11, July 24–28, 2011, Beijing, China.
Extracting and Ranking Product Features in Opinion Documents Lei Zhang #, Bing Liu #, Suk Hwan Lim *, Eamonn O’Brien-Strain * # University of Illinois.
Lexical Affect Sensing: Are Affect Dictionaries Necessary to Analyze Affect? Alexander Osherenko, Elisabeth André University of Augsburg.
An evolutionary approach for improving the quality of automatic summaries Constantin Orasan Research Group in Computational Linguistics School of Humanities,
GENERATING RELEVANT AND DIVERSE QUERY PHRASE SUGGESTIONS USING TOPICAL N-GRAMS ELENA HIRST.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
How to Really Review Papers CS 8803 AIC. Prvulovic: CORD 2 Paper Reviewing Algorithm Read the paper Think about it Take a look at related work Leave it.
Paradigmatic Modifiability Statistics For the Extraction of
Double and Multiple Sampling Plan
Queensland University of Technology
Sampling Distributions
Terminology problems in literature mining and NLP
Social Knowledge Mining
Finding the Area Between Curves
Testing Hypotheses I Lesson 9.
Chapter 9 Hypothesis Testing: Single Population
Presentation transcript:

You Can’t Beat Frequency (Unless You Use Linguistic Knowledge) – A Qualitative Evaluation of Association Measures for Collocation and Term Extraction Joachim Wermter and Udo Hahn Jena University ACL 2006 Regular Conference Paper

Objective Compare the performance of frequency, t- test, LSM and LPM methods on collocation extraction and domain-specific automatic term recognition

Collocation Extraction Extract idioms “kick the bucket”

Domain-Specific Term Extraction Extract domain-specific phrases “mitochondrial inheritance”

Corpus

LSM A “linguistic knowledge-based” method for collocation extraction proposed by the same authors in another paper Assumes that idioms are less modifiable by supplements –e.g. “kick the beautiful bucket” probability of PNV triple having Supp k : f(x) : frequency of x

LSM Modifiability of a PNV triple Probability of a PNV triple Collocation Score

LPM A “linguistic knowledge-based” method for automatic term recognition proposed by the same authors in another paper Assumes that words in a phrase are less interchangeable –e.g mitochondrion inheritance  money inheritance Modifiability of a phrase: mod k (n-gram) : replace k words sel i : particular replacement

LPM Phrase Score:

Evaluation Criteria Compared to the baseline frequency ranking method, a good ranking function should have the four characteristics: 1.Keep the true positives in the upper portion of the list 2.Keep the true negatives in the lower portion of the list 3.Demote true negatives from the upper portion 4.Promote true positives from the lower portion

Collocation Extraction Results

Automatic Term Recognition Results

Observations CE Criterion 1 –t-test and frequency methods have similar performance –LSM promotes some TPs to top 1/6 ATR Criterion 1 –t-test and frequency methods have similar performance –LPM promotes a few TPs to top 1/6

Observations CE Criterion 2 –LSM promotes a lot more TNs to upper portion than t-test method (bad…) ATR Criterion 2 –Same as above

Observations CE Criterion 3 –LSM demotes a lot more TNs to the lower portion than t-test ATR Criterion 3 –Same as above

Observations CE Criterion 4 –LSM promotes more TPs to upper portion than t-test ATR Criterion 4 –Same as above

Conclusion LSM and LPM methods are better than t- test and frequency methods Pure statistics methods are worse than knowledge-based methods