Presentation is loading. Please wait.

Presentation is loading. Please wait.

Concept Hierarchy Induction by Philipp Cimiano

Similar presentations


Presentation on theme: "Concept Hierarchy Induction by Philipp Cimiano"— Presentation transcript:

1 Concept Hierarchy Induction by Philipp Cimiano

2 Objective Structure information into categories
Provide a level of generalization to define relationships between data Application: Backbone of any ontology

3 Overview Different approaches of acquiring conceptual hierarchies from text corpus. Various clustering techniques. Evaluation Related Work Conclusion

4 Machine Readable Dictionaries
Entries: ‘a tiger is a mammal’, or ‘mammals such as tigers, lions or elephants’. exploit the regularity of dictionary entries. the head of the first NP - hypernym.

5 Example

6 Exception

7 Exception is-a (corolla, part)………..is a NOT VALID
is-a (republican, member) ……….. is a NOT VALID is-a (corolla, flower)………..is a NOT VALID is-a (republican, political party)………..is a NOT VALID

8 Exception

9 Alshawis solution

10 Results using MRDs Dolan et al. - 87% of the hypernym relations extracted are correct Calzolari cites a precision of > 90% Alshawi - precision of 77%

11 Strengths And Weaknesses
Correct, explicit knowledge Robust basis for ontology learning Weakness- domain independent

12 Lexico-Syntactic patterns
Task: automatically learning hyponym relations from the corpora. 'Such injuries as bruises, wounds and broken bones' hyponym (bruise, injury) hyponym (wound, injury) hyponym (broken bone, injury)

13 Hearst patterns 'Such injuries as bruises, wounds and broken bones'

14 Requirements Occur frequently in many text genres.
Accurately indicate the relation of interest. Be recognizable with little or no pre-encoded knowledge

15 Strengths And Weaknesses
Identified easily and are accurate Weakness: patterns appear rarely is-a relation do not appear in Hearst style pattern

16 Distribution Similarity
'you shall know a word by the company it keeps’ [Firth, 1957]. semantic similarity of words – similarity of the contexts.

17

18 Using distribution similarity

19 Strengths And Weaknesses
reasonable concept hierarchy. Weakness: Cluster tree lacks clear and formal interpretation Does not provide any intentional description of concepts Similarities may be accidental (sparse data)

20 Formal Concept Analysis (FCA)

21

22 FCA output

23 Similarity measures

24 Smoothing

25 Evaluation Semantic cotopy (SC). Taxonomy overlap (TO)

26 Evaluation Measure

27 100% Precision Recall

28 Low Recall

29 Low Precision

30 Results

31 Results

32 Results

33 Results

34 Strengths And Weaknesses
FCA generates formal concepts Provides intentional description Weakness: Size of the lattice can get exponential in the size spurious clusters Finding appropriate labels for the cluster

35 Problems with Unsupervised Approaches to Clustering
Data sparseness leads to spurious syntactic similarities Produced clusters can’t be appropriately labeled Unsupervised approaches are dependant upon calculating similarity of words on the basis of linguistic context

36 Guided Clustering Hypernyms directly used to guide clustering
WordNet Hearst Agglomerative clustering Two terms are only clustered if there is a corresponding common hypernym according to an oracle Hearst—an approach matching lexico-syntactic patterns to find hypernyms

37 Similarity computation calculated by taking the cosine between corresponding context vectors of two terms

38 Similarity Computation
Ten most similar terms of the tourism reference taxonomy

39 The Hypernym Oracle Three sources
WordNet Hearst patterns matched in a corpus Hearst patterns matched in the World Wide Web Record hypernyms and amount of evidence found in support of hypernyms. How the oracle is constructed

40 WordNet Collect hypernyms found in any dominating synset containing term, t Include number of times the hypernym appears in a dominating synset

41 Hearst Patterns (Corpus)
NP = noun phrase Record number of isa-relations found between two terms

42 Hearst Patterns (WWW) Download 100 Google abstracts for each concept and clue: Use clues to make Google queries and to search resulting abstracts for terms that have isa-relationships with the concept Again, record the number of relationships found between two terms (term and concept)

43 Evidence Total Evidence for Hypernyms: time: 4 vacation: 2 period: 2

44 Clustering Algorithm Input a list of terms
Calculate the similarity between each pair of terms and sort from highest to lowest For each potential pair to be clustered consult the oracle.

45 Consulting the Oracle case 1
If term 1 is a hypernym of term 2 or vice-versa: Create appropriate subconcept relationship.

46 Consulting the Oracle case 2
Find the common hypernym for both terms with greatest evidence. If one term has already been classified: If a common hypernym even exists > means ‘hypernym of’ for my purposes t’ = classification h = common hypernym t’ = h h is a hypernym of t’ t’ is a hypernym of h

47 Consulting the Oracle case 3
Neither term has been classified: Each term becomes a subconcept of the common hypernym.

48 Consulting the Oracle case 4
The terms do not share a common hypernym: Set aside the terms for further processing.

49 r-matches For all unprocessed terms, check for r-matches (i.e. ‘credit card’ matches ‘international credit card’) unprocessed terms are those with no similar terms t1 r-matches t2 if t2 literally contains the string t1 in it

50 Further Processing If either term in a pair is already classified as t’, the other term is classified under t’ as well. Otherwise place both terms under the hypernym of either term with the most evidence. Any unclassified terms are added under the root concept. run these algorithms on those pairs that were set aside

51 Example

52 Evaluation Taxonomic overlap (TO) Sibling overlap (SO)
ignore leaf nodes Sibling overlap (SO) measures quality of clusters All of these measures are compared to a handcrafted reference concept hierarchy Ignore leaf nodes, otherwise hierarchies with every concept directly subordinate to root node can be rated very high.

53 Evaluation Tourism domain: Finance domain: Lonely Planet Mecklenburg
Reuters-21578 Caraballo – post-processing the hierarchy with hypernyms

54 Tourism Results—TO These are F-measures
Caraballo’s method = Gold Standard: After agglomerative clustering: label each cluster based on most frequent hypernym (must be a hypernym of at least 2 members of the cluster; otherwise remove cluster)

55 Finance Results—TO These are F-measures
Caraballo’s method = Gold Standard: After agglomerative clustering: label each cluster based on most frequent hypernym (must be a hypernym of at least 2 members of the cluster; otherwise remove cluster)

56 Tourism Results—SO These are F-measures
Caraballo’s method = Gold Standard: After agglomerative clustering: label each cluster based on most frequent hypernym (must be a hypernym of at least 2 members of the cluster; otherwise remove cluster)

57 Finance Results—SO These are F-measures
Caraballo’s method = Gold Standard: After agglomerative clustering: label each cluster based on most frequent hypernym (must be a hypernym of at least 2 members of the cluster; otherwise remove cluster)

58 Human Evaluation Scores 3 : correct 2: Almost correct
1: not completely wrong 0: wrong # = number of non-root taxonomic relations

59 Future Work Take word sense into consideration for the WordNet source.

60 Summary Hypernym guided agglomerative clustering works pretty good.
Better than the “Golden Standard” Good human evaluation Provides labels for clusters No spurious similarities Faster than agglomerative clustering Gold standard = Caraballo faster because similarities are calculated between single elements and not between clusters

61 Learning from Heterogeneous Sources of Evidence
Many ways to learn concept hierarchies Can we combine different paradigms? Any manual attempt to combine strategies would be ad hoc Use supervised learning to combine techniques

62 Determining relationships with machine learning
Example: Determine if a pair of words has an “isa” relationship

63 Feature 1: Matching patterns in a corpus
Given two terms t1 and t2 we record how many times a Hearst-pattern indicating an isa-relation between t1 and t2 is matched in the corpus Normalize by maximum number of Hearst patterns found for t1

64 Example This provided the best F-measure with a single-feature classifier

65 Feature 2: Matching patterns on the web
Use the Google API to count the matches of a certain expression on the Web

66 Feature 3: Downloading webpages
Allows for matching expressions with a more complex linguistic structure Assign functions to each of the Hearst patterns to be matched Use these “clues” to decide what pages to download Download 100 abstracts matching the query “such as conferences”

67 Example

68 Feature 4: WordNet – All senses
Is there a hypernym relationship between t1 and t2? Can be more than one path from the synsets of t1 to the synsets of t2

69 Feature 5: WordNet – First sense
Only consider the first sense of t1

70 Feature 6: “Head”- heuristic
If t1 r-matches t2 we derive the relation isa(t2,t1) e.g. t1 = “conference” t2 = “international conference” isahead(“international conference”,”conference”)

71 Feature 7: Corpus-based subsumption
t1 is a subclass of t2 if all the syntactic contexts in which t1 appears are also shared by t2

72 Feature 8: Document-based subsumption
t1 is a subclass of term t2 if t2 appears in all documents in which t1 appears # of pages where t1 and t2 occur # of pages where t1 occurs

73 Example

74 Naïve Threshold Classifier
Used as a baseline Classify an example as positive if the value of a given feature is above some threshold t For each feature, the threshold has been varied from 0 to 1 in steps of 0.01

75 Baseline Measures

76 Evaluation Classifiers Naïve Bayes Decision Tree Perceptron
Multi-layer perceptron

77 Evaluation Strategies
Undersampling Remove a number of majority class examples (non-isa examples) Oversampling Add additional examples to the minority class Varying the classification threshold Try different threshold values other than 0.5 Introducing a cost matrix Different penalties for different types of misclassification One Class SVMs Only considers positive examples 77

78 Results

79 Results (cont.)

80 Discussion The best results achieved with the one-class SVM (F = 32.96%) More than 10 points above the baseline classifier average (F = 21.28%) and maximum (F = 21%) strategies More than 14 points better than the best single-feature classifier (F = 18.84%) using the isawww feature Second best results obtained with a Multilayer Perceptron using oversampling or undersampling

81 Discussion Gain insight from finding which features were most used by classifiers Used this information to modify features and rerun experiments

82 Summary Using different approaches is useful
Machine learning approaches outperform naïve averaging Unbalanced character of the dataset poses a problem SVMs (which are not affected by the imbalance) produce the best results This approach can show which features are the most reliable as predictors

83 Related Work Taxonomy Construction Taxonomy Refinement
Lexico-syntactic patterns Clustering Linguistic approaches Taxonomy Refinement Taxonomy Extension

84 Lexico-syntactic patterns
Hearst Iswanska et al. – added extra patterns Poesia et al. – anaphoric resolution Ammad et al. – applying to specific domains Etzioni et al. – patterns matched on the www Cederburg and Widdows – precision improved with Latent Semantic Analysis Others working on learning patterns automatically

85 Clustering Hindle Pereira et al. Caraballo group nouns semantically
derive verb-subject and verb-object dependencies from a 6 million word sample of Associated Press news stories Pereira et al. top-down soft clustering algorithm with deterministic annealing words can appear in different clusters (multiple meanings of words) Caraballo bottom-up clustering approach to build a hierarchy of nouns uses conjunctive and appositive constructions for nouns derived from the Wall Street Journal Corpus

86 Clustering (cont.) The ASIUM System The Mo'K Workbench Grefenstette
Gasperin et al. Reinberger et al. Lin et al. CobWeb Crouch et al. Haav Curran et al. Terascale Knowledge Acquisition

87 Linguistic Approaches
Linguistic analysis exploited more directly rather than just for feature extraction OntoLT - use shallow parser to label parts of speech and grammatical relations (e.g. HeadNounToClass-ModToSubClass, which maps a common noun to a concept or class) OntoLearn - analyze multi-word terms compositionally with respect to an existing semantic resource (Word-Net) Morin et al. - tackle the problem of projecting semantic relations between single terms to multiple terms (e.g. project the isa-relation between apple and fruit to an isa-relation between apple juice and fruit juice)

88 Linguistic Approaches
Sanchez and Moreno – download first n hits for a search word and process the neighborhood linguistically to determine candidate modifiers for the search term Sabou - inducing concept hierarchies for the purpose of modeling web services (applies methods not to full text, but to Java-documentation of web services)

89 Taxonomy Refinement Hearst and Schutze Widdows Madche, Pekar and Staab
Alfonseca et aL

90 Taxonomy Extension Agirre et al. Faatz and Steinmetz Turney

91 Conclusions Compared different hierarchical clustering approaches with respect to: effectiveness speed traceability Set-theoretic approaches, as FCA, can outperform similarity-based approaches.

92 Conclusions Presented an algorithm for clustering guided by a hypernym oracle. More efficient than agglomerative clustering.

93 Conclusions Used machine learning techniques to effectively combine different approaches for learning taxonomic relations from text. A learned model indeed outperforms all single approaches. It also outperforms naïve combinations of them.

94 Open Issues Which similarity or weighting measure should be chosen
Which features should be considered to represent a certain term Can features be aggregated to represent a term at a more abstract level How should we model polysemy of terms Can we automatically induce lexico-syntactic patterns (unsupervised!) What other approaches are there for combining different paradigms; and how can we compare these These are all issues very general to the problem of Concept Hierarchy Induction Obviously weight and similarity measure depend largely on the dataset. Cluster features themselves. lexico-syntactic patterns are like Hearst patterns

95 Questions


Download ppt "Concept Hierarchy Induction by Philipp Cimiano"

Similar presentations


Ads by Google