Using Lexical Knowledge to Evaluate the Novelty of Rules Mined from Text Sugato Basu, Raymond J. Mooney, Krupakar V. Pasupuleti, Joydeep Ghosh Presented.

Slides:



Advertisements
Similar presentations
Relevant characteristics extraction from semantically unstructured data PhD title : Data mining in unstructured data Daniel I. MORARIU, MSc PhD Supervisor:
Advertisements

Correlation Search in Graph Databases Yiping Ke James Cheng Wilfred Ng Presented By Phani Yarlagadda.
Improved TF-IDF Ranker
1 Extended Gloss Overlaps as a Measure of Semantic Relatedness Satanjeev Banerjee Ted Pedersen Carnegie Mellon University University of Minnesota Duluth.
2001/12/181/50 Discovering Robust Knowledge from Databases that Change Author: Chun-Nan Hsu, Craig A. Knoblock Advisor: Dr. Hsu Graduate: Yu-Wei Su.
COLLABORATIVE FILTERING Mustafa Cavdar Neslihan Bulut.
MSS 905 Methods of Missiological Research
Using Structure Indices for Efficient Approximation of Network Properties Matthew J. Rattigan, Marc Maier, and David Jensen University of Massachusetts.
Discovering Complex Matchings across Web Query Interfaces: A Correlation Mining Approach Bin He, Kevin Chen-Chuan Chang, Jiawei Han Presented by Dayi Zhou.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining - revision Martin Russell.
Semantic Video Classification Based on Subtitles and Domain Terminologies Polyxeni Katsiouli, Vassileios Tsetsos, Stathes Hadjiefthymiades P ervasive C.
S IMILARITY M EASURES FOR T EXT D OCUMENT C LUSTERING Anna Huang Department of Computer Science The University of Waikato, Hamilton, New Zealand BY Farah.
C SC 620 Advanced Topics in Natural Language Processing Lecture 10 2/19.
Data Mining – Intro.
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Chapter 5: Information Retrieval and Web Search
Feature Selection for Automatic Taxonomy Induction The Features Input: Two terms Output: A numeric score, or. Lexical-Syntactic Patterns Co-occurrence.
Some Thoughts to Consider 6 What is the difference between Artificial Intelligence and Computer Science? What is the difference between Artificial Intelligence.
Unsupervised Learning. CS583, Bing Liu, UIC 2 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
COMP423.  Query expansion  Two approaches ◦ Relevance feedback ◦ Thesaurus-based  Most Slides copied from ◦
by B. Zadrozny and C. Elkan
The identification of interesting web sites Presented by Xiaoshu Cai.
Introduction. The Role of Statistics in Science Research can be qualitative or quantitative Research can be qualitative or quantitative Where the research.
1 Query Operations Relevance Feedback & Query Expansion.
Quantitative Analysis. Quantitative / Formal Methods objective measurement systems graphical methods statistical procedures.
WORD SENSE DISAMBIGUATION STUDY ON WORD NET ONTOLOGY Akilan Velmurugan Computer Networks – CS 790G.
Chapter 4: Test administration. z scores Standard score expressed in terms of standard deviation units which indicates distance raw score is from mean.
Péter Schönhofen – Ad Hoc Hungarian → English – CLEF Workshop 20 Sep 2007 Performing Cross-Language Retrieval with Wikipedia Participation report for Ad.
Data Mining Knowledge on rough set theory SUSHIL KUMAR SAHU.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Extracting meaningful labels for WEBSOM text archives Advisor.
Introduction to Behavioral Statistics Correlation & Regression.
Unsupervised Word Sense Disambiguation REU, Summer, 2009.
University of Malta CSA3080: Lecture 6 © Chris Staff 1 of 20 CSA3080: Adaptive Hypertext Systems I Dr. Christopher Staff Department.
Data Mining – Intro. Course Overview Spatial Databases Temporal and Spatio-Temporal Databases Multimedia Databases Data Mining.
Statistical Power The power of a test is the probability of detecting a difference or relationship if such a difference or relationship really exists.
TEAM VORTEX -Brandon Gill and Chris Kwon-. -HELLO, MY NAME IS BRANDON GILL. I’ AM A SENIOR AT UKIAH HIGH SCHOOL AND I LOVE SCIENCE! MY FAVORITE SCIENCE.
Characterizing the Uncertainty of Web Data: Models and Experiences Lorenzo Blanco, Valter Crescenzi, Paolo Merialdo, Paolo Papotti Università degli Studi.
Semantic v.s. Positions: Utilizing Balanced Proximity in Language Model Smoothing for Information Retrieval Rui Yan†, ♮, Han Jiang†, ♮, Mirella Lapata‡,
1 Appendix D: Application of Genetic Algorithm in Classification Duong Tuan Anh 5/2014.
Evgeniy Gabrilovich and Shaul Markovitch
Intelligent Database Systems Lab Presenter : Chang,Chun-Chih Authors : CHRISTOS BOURAS, VASSILIS TSOGKAS 2012, KBS A clustering technique for news articles.
So, what’s the “point” to all of this?….
Introduction. The Role of Statistics in Science Research can be qualitative or quantitative Research can be qualitative or quantitative Where the research.
Unit 1 Sections 1-1 & : Introduction What is Statistics?  Statistics – the science of conducting studies to collect, organize, summarize, analyze,
Information Retrieval and Web Search Relevance Feedback. Query Expansion Instructor: Rada Mihalcea.
DeepDive Model Dongfang Xu Ph.D student, School of Information, University of Arizona Dec 13, 2015.
Post-Ranking query suggestion by diversifying search Chao Wang.
Commonsense Reasoning in and over Natural Language Hugo Liu, Push Singh Media Laboratory of MIT The 8 th International Conference on Knowledge- Based Intelligent.
An Applied Ontological Approach to Computational Semantics Sam Zhang.
Natural Language Generation with Markov Chains and Grammar Sam Zhang.
Semantic Grounding of Tag Relatedness in Social Bookmarking Systems Ciro Cattuto, Dominik Benz, Andreas Hotho, Gerd Stumme ISWC 2008 Hyewon Lim January.
PART 2 SPSS (the Statistical Package for the Social Sciences)
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Query expansion COMP423. Menu Query expansion Two approaches Relevance feedback Thesaurus-based Most Slides copied from
Correlation and Regression Q560: Experimental Methods in Cognitive Science Lecture 13.
Similarity Measures for Text Document Clustering
MSS 905 Methods of Missiological Research
Queensland University of Technology
Machine Learning for Online Query Relaxation
Introduction to Behavioral Statistics
The Nature of Probability and Statistics
Statistical Evaluation
WordNet WordNet, WSD.
Statistical Inference for Managers
The Nature of Probability and Statistics
Relevance Feedback & Query Expansion
Descriptive statistics Pearson’s correlation
Dynamic Word Sense Disambiguation with Semantic Similarity
Principles of Experimental Design
Presentation transcript:

Using Lexical Knowledge to Evaluate the Novelty of Rules Mined from Text Sugato Basu, Raymond J. Mooney, Krupakar V. Pasupuleti, Joydeep Ghosh Presented by Joseph Schlecht

Problem Description Modern data-mining techniques discover large number of relationships (rules) –Antecedent  Consequent Few may actually be of interest –CS job hunting: SQL  database How do we find rules that are interesting and novel? Notice this is subjective

Problem Formalization Authors consider text mining –Rules consist of words in natural language Use WordNet and define semantic distance between two words Novelty is defined w.r.t the semantic distance between words in the antecedent and consequent of a rule

Semantic Distance Given words w i and w j, d(w i, w j ) = Dist(P(w i, w j )) + K * Dir(P(w i, w j )) Dist(p) is the distance along path p –Weighted by relation type (15 in WordNet) Dir(p) is the number of directional changes on p –Defined 3 directions according to relation type K is a chosen constant

Weight and Direction Info RelationWeightDirection Synonym, Attribute, Pertainym, Similar 0.5Horizontal Antonym2.5Horizontal Hypernym, (Member|Part|Substance), Meronym 1.5Up Hyponym, (Member|Part|Substance) Holonym, Cause, Entailment 1.5Down

Novelty For each rule, a score of novelty is generated Let A = {set of antecedent words} and C = {set of consequent words} in a given rule For each word w i in A and w j in C –Score(w i, w j )  d(w i, w j ) Score of rule = average of all (w i, w j ) scores

Experiment Measure success by comparing the heuristic’s results of novelty scoring to humans’ Used rules generated by DiscoTEX from 9000 Amazon.com book descriptions Four random samples of 25 rules were made Four groups of humans scored each sample –0.0 (least interesting) to 10.0 (most interesting) One set was used as training for the heuristic (to find K), the other three were used for experiments

Results Human-Human Correlation Heuristic-Human Correlation RawRankRawRank Group Group Group Raw = Pearson’s Raw Score Rank = Spearman’s Ranks Score

Results (cont) Example of rules scored by the heuristic High Score (9.5) romance love heart  midnight Medium Score (5.8) author romance  characters love Low Score (1.9) Astronomy science  space

Discussion Humans rarely agreed with each other Correlation between heuristic and human was similar to human-human correlation –Success, but not too meaningful Provided statistical evidence that correlation is unlikely due to random chance Future tests would use dataset that had higher human-human correlation