Download presentation
Presentation is loading. Please wait.
Published byWarren Randall Modified over 9 years ago
1
F ROM U NSTRUCTURED I NFORMATION T O L INKED D ATA Axel Ngonga Head of SIMBA@AKSW University of Leipzig IASLOD, August 15/16 th 2012
2
Motivation
3
Where does the LOD Cloud come from? Structured data Triplify, D2R Semi-structured data DBpedia Unstructured data ??? Unstructured data make up 80% of the Web How do we extract Linked Data from unstructured data sources?
4
Overview 1. Problem Definition 2. Named Entity Recognition Algorithms Ensemble Learning 3. Relation Extraction General approaches OpenIE approaches 4. Entity Disambiguation URI Lookup Disambiguation 5. Conclusion NB: Will be mainly concerned with the newest developments.
5
Overview 1. Problem Definition 2. Named Entity Recognition Algorithms Ensemble Learning 3. Relation Extraction General approaches OpenIE approaches 4. Entity Disambiguation URI Lookup Disambiguation 5. Conclusion
6
Problem Definition Simple(?) problem: given a text fragment, retrieve All entities and relations between these entities automatically plus „ground them“ in an ontology Also coined Knowledge Extraction John Petrucci was born in New York. :John_Petrucci :New_York dbo:birthPlace :John_Petrucci dbo:birthPlace :New_York.
7
Problems 1. Finding entities Named Entity Recognition 2. Finding relation instances Relation Extraction 3. Finding URIs URI Disambiguation
8
Overview 1. Problem Definition 2. Named Entity Recognition Algorithms Ensemble Learning 3. Relation Extraction General approaches OpenIE approaches 4. Entity Disambiguation URI Lookup Disambiguation 5. Conclusion
9
Named Entity Recognition Problem definition: Given a set of classes, find all strings that are labels of instances of these classes within a text fragment John Petrucci was born in New York. [John Petrucci, PER] was born in [New York, LOC].
10
Named Entity Recognition Problem definition: Given a set of classes, find all strings that are labels of instances of these classes within a text fragment Common sets of classes CoNLL03: Person, Location, Organization, Miscelleaneous ACE05: Facility, Geo-Political Entity, Location, Organisation, Person, Vehicle, Weapon BioNLP2004: Protein, DNA, RNA, cell line, cell type Several approaches Direct solutions (single algorithms) Ensemble Learning
11
NER: Overview of approaches Dictionary-based Hand-crafted Rules Machine Learning Hidden Markov Model (HMMs) Conditional Random Fields (CRFs) Neural Networks k Nearest Neighbors (kNN) Graph Clustering Ensemble Learning Veto-Based (Bagging, Boosting) Neural Networks
12
NER: Dictionary-based Simple Idea 1. Define mappings between words and classes, e. g., Paris Location 2. Try to match each token from each sentence 3. Return the mapping entities Time-Efficient at runtime × Manuel creation of gazeteers × Low Precision (Paris = Person, Location) × Low Recall (esp. on Persons and Organizations as the number of instances grows)
13
NER: Rule-based Simple Idea 1. Define a set of rule to find entities, e.g., [PERSON] was born in [LOCATION]. 2. Try to match each sentence to one or several rules 3. Return the mapping entities High precision × Manuel creation of rules is very tedious × Low recall (finite number of patterns)
14
NER: Markov Models
15
NER: Hidden Markov Models Extension of Markov Models States are hidden and assigned an output function Only output is seen Transitions are learned from training data How do they work? Input: Discrete sequence of features (e.g., POS Tags, word stems, etc.) Goal: Find the best sequence of states that represent the input Output: hopefully right classification of each token S0S0 S1S1 … SnSn PER _ LOC
16
NER: k Nearest Neighbors Idea Describe each token from a labelled training data set with a set of features (e.g., left and right neigbors) Each new token is described with the same features Assign the class of its k nearest neighbors
17
NER: So far … „Simple approaches“ Apply one algorithm to the NER problem Bound to be limited by assumptions of model Implemented by a large number of tools Alchemy Stanford NER Illinois Tagger Ontos NER Tagger LingPipe …
18
NER: Ensemble Learning Intuition: Each algorithm has its strengths and weaknesses Idea: Use ensemble learning to merge results of different algorithms so as to create a meta-classifier of higher accuracy Dictionary-based approaches Pattern-based approaches Condition Random Fields Support Vector Machines
19
NER: Ensemble Learning Idea: Merge the results of several approaches for improving results Simplest approaches: Voting Weighted voting Input System 1 System 2 System n Merger Output
20
NER: Ensemble Learning When does it work? Accuracy Need for exisiting solutions to be „good“ Merging random results lead to random results Given, current approaches reach 80% F-Score Diversity Need for smallest possible amount of correlation between approaches E.g., merging two HMM-based taggers won‘t help Given, large number of approaches for NER
21
NER:FOX Federated Knowledge Extraction Framework Idea: Apply ensemble learning to NER Classical approach: Voting Does not make use of systematic error Partly difficult to train Use neural networks instead Can make use of systematic errory Easy to train Converge fast http://fox.aksw.org
22
NER: FOX
23
NER: FOX on MUC7
25
NER: FOX on Website Data
27
NER: FOX on Companies and Countries No runtime issues (parallel implementation) NN overhead is small × Overfitting
28
NER: Summary Large number of approaches Dictionaries Hand-Crafted rules Machine Learning Hybrid … Combining approaches leads to better results than single algorithms
29
Overview 1. Problem Definition 2. Named Entity Recognition Algorithms Ensemble Learning 3. Relation Extraction General approaches OpenIE approaches 4. Entity Disambiguation URI Lookup Disambiguation 5. Conclusion
30
RE: Problem Definition Find the relations between NEs if such relations exist. NEs not always given a-priori (open vs. closed RE) bornIn ([John Petrucci, PER], [New York, LOC]). John Petrucci was born in New York. [John Petrucci, PER] was born in [New York, LOC].
31
RE: Approaches Hand-crafted rules Pattern Learning Coupled Learning
32
RE: Pattern-based Hearst patterns [Hearst: COLING‘92] POS-enhanced regular expression matching in natural- language text NP 0 {,} such as {NP 1, NP 2, … (and|or) }{,} NP n NP 0 {,}{NP 1, NP 2, … NP n-1 }{,} or other NP n “The bow lute, such as the Bambara ndang, is plucked and has an individual curved neck for each string.” isA(“Bambara ndang”, “bow lute”) Time-Efficient at runtime × Very low recall × Not adaptable to other relations
33
RE: Pattern-based Noun classification from predicate-argument structures [Hindle: ACL’90] Clustering of nouns by similar verbal phrases Similarity based on co-occurrence frequencies (mutual information)
34
RE: DIPRE DIPRE = Dual Iterative Pattern Relation Extraction Semi-supervised, iterative gathering of facts and patterns Positive & negative examples as seeds for a given target relation e.g. +(Hillary, Bill) ; +(Carla, Nicolas); –(Larry, Google) Various tuning parameters for pruning low-confidence patterns and facts Extended to SnowBall / QXtract (Hillary, Bill) (Carla, Nicolas) X and her husband Y X and Y on their honeymoon X and Y and their children X has been dating with Y X loves Y (Angelina, Brad) (Hillary, Bill) (Victoria, David) (Carla, Nicolas) (Larry, Google) …
35
RE: NELL Never-Ending Language Learner (http://rtw.ml.cmu.edu/)http://rtw.ml.cmu.edu/ Open IE with ontological backbone Closed set of categories & typed relations Seeds/counter seeds (5-10) Open set of predicate arguments (instances) Coupled iterative learners Constantly running over a large Web corpus since January 2010 (200 Mio pages) Periodic human supervision athletePlaysForTeam (Athlete, SportsTeam) athletePlaysForTeam (Alex Rodriguez, Yankees) athletePlaysForTeam (Alexander_Ovechkin, Penguins)
36
RE: NELL Meta-Boostrap Learner (MBL) For i = 1,..,∞ do For each predicate p do For each extractor e do Extract new candidates for p using e with recently promoted instances Filter candidates that violate mutual-exclusion or type constraints Promote candidates that were extracted by all extractors Meta-Boostrap Learner (MBL) For i = 1,..,∞ do For each predicate p do For each extractor e do Extract new candidates for p using e with recently promoted instances Filter candidates that violate mutual-exclusion or type constraints Promote candidates that were extracted by all extractors Coupled Pattern Learner (CPL) For i = 1,..,∞ do – For each predicate p do – Extract new candidates instances/contextual patterns of p using recently promoted instances – Filter candidates that violate constraints – Rank candidate instances/patterns – Promote top candidates for next round Coupled Pattern Learner (CPL) For i = 1,..,∞ do – For each predicate p do – Extract new candidates instances/contextual patterns of p using recently promoted instances – Filter candidates that violate constraints – Rank candidate instances/patterns – Promote top candidates for next round – Coupled output constraints – For f 1 (x 1 ) y 1 and f 2 (x 1 ) y 2 – Restrict output y 1 and y 2 (e.g. f 1 (x) f 2 (x) for functional dependencies, mut.-ex.) – Compositional constraints – For f 1 (x 1 ) y 1 and f 2 (x 1, x 2 ) y 2 – Restrict y 1, y 2 to valid pairs (special case: type checking) – Multi-view agreement – Co-training classifiers f 1 (x 1 ) y and f 2 (x 2 ) y – Constraints employed for experiments – Mutual-exclusiveness predicates – Type checking – Label-agreement Coupled = Concurrent learning of patterns and instances across several learners
37
RE: NELL Conservative strategy Avoid Semantic Drift
38
RE: BOA Bootstrapping Linked Data (http://boa.aksw.org) Core idea: Use instance data in Data Web to discover NL patterns and new instances
39
RE: BOA Follows conservative strategy Only top pattern Frequency threshold Score Threshold Evaluation results
40
RE: Summary Several approaches Hand-crafted rules Machine Learning Hybrid Large number of instances available for many relations Runtime problem Parallel implementations Many new facts can be found × Semantic Drift × Long tail × Entity Disambiguation
41
Overview 1. Problem Definition 2. Named Entity Recognition Algorithms Ensemble Learning 3. Relation Extraction General approaches OpenIE approaches 4. Entity Disambiguation URI Lookup Disambiguation 5. Conclusion
42
ED: Problem Definition Given (a) refence knowledge base(s), a text fragment, a list of NEs (incl. position), and a list a relations, find URIs for each of the NEs and relations Very difficult problem Ambiguity, e.g., Paris = Paris Hilton? Paris (France)? Difficult even for humans, e.g., Paris‘ mayor died yesterday Several solutions Indexing Surface Form Graph-based
43
ED: Problem Definition bornIn ([John Petrucci, PER], [New York, LOC]). John Petrucci was born in New York. [John Petrucci, PER] was born in [New York, LOC]. :John_Petrucci dbo:birthPlace :New_York.
44
ED: Indexing More retrieval than disambiguation Similar to dictionary-based approaches Idea Index all labels in reference knowledge base Given an input label, retrieve all entities with a similar label × Poor recall (unknown surface form, e.g., „Mme Curie“ für „Marie Curie“) × Low precision (Paris = Paris Hilton, Paris (France), …)
45
ED: Type Disambiguation Extension of indexing Index all labels Infer type information Retrieve labels from entities of the given type Same recall as previous approach Higher precision Paris[LOC] != Paris[PER] Still, Paris (France) vs. Paris (Ontario) Need for context
46
ED: Spotlight Known surface forms (http://dbpedia.org/spotlight)http://dbpedia.org/spotlight Based on DBpedia + Wikipedia Uses supplementary knowledge including disambiguation pages, redirects, wikilinks Three main steps Spotting: Finding possible mentions of DBpedia resources, e.g., John Petrucci was born in New York. Candidate Selection: Find possible URIs, e.g., John Petrucci :JohnPetrucci New York :New_York, :New_York_County, … Disambiguation: Map context to vector for each resource New York :New_York
47
ED: YAGO2 Joint Disambiguation Mississippi, one of Bob’s later songs, was first recorded by Sheryl on her album. ♬
48
ED: YAGO2 Mississippi (State) Bob Dylan Songs Sheryl Cruz Sheryl Lee Mississippi (Song) Sheryl Crow Objective: Maximize objective function (e.g., total weight) Constraint: Keep at least one entity per mention Mentions of Entities Entity Candidates sim(cxt(m l ),cxt(e i )) prior(m l,e i ) coh(e i,e j )
49
ED: FOX Generic Approach A-priori score ( ): Popularity of URIs Similarity score ( ): Similarity of resource labels and text Coherence score ( ): Correlation between URIs 49
50
ED:FOX Allows the use of several algorithms HITS Pagerank Apriori Propagation Algorithms … 50
51
ED: Summary Difficult problem even for humans Several approaches Simple search Search with restrictions Known surface forms Graph-based Improved F-Score for DBpedia (70-80%) × Low F-Score for generic knowledge bases × Intrinsically difficult × Still a lot to do
52
Overview 1. Problem Definition 2. Named Entity Recognition Algorithms Ensemble Learning 3. Relation Extraction General approaches OpenIE approaches 4. Entity Disambiguation URI Lookup Disambiguation 5. Conclusion
53
Conclusion Discussed basics of … Knowledge Extraction problem Named Entity Recognition Relation Extraction Entity Disambiguation Still a lot of research necessary Ensemble and active Learning Entity Disambiguation Question Answering …
54
Thank You! Questions?
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.