Presentation is loading. Please wait.

Presentation is loading. Please wait.

Bo Lin Kevin Dela Rosa Rushin Shah.  As part of our research, we are working on a cross- document co-reference resolution system  Co-reference Resolution:

Similar presentations


Presentation on theme: "Bo Lin Kevin Dela Rosa Rushin Shah.  As part of our research, we are working on a cross- document co-reference resolution system  Co-reference Resolution:"— Presentation transcript:

1 Bo Lin Kevin Dela Rosa Rushin Shah

2  As part of our research, we are working on a cross- document co-reference resolution system  Co-reference Resolution: Extract all noun phrases from a document (names, descriptions, pronouns), and cluster them according to the real-world entity they describe.  Each such cluster  a chain  Within-doc: Cluster NPs from a single document  Cross-doc: Cluster NPs from different documents

3  Run a WDC system on a document and extract chains corresponding to real-world entities. For each chain, track all sentences from which its mentions are obtained.  Features over pairs of such chains: ◦ SoftTFIDF similarity between names ◦ All words TFIDF cosine similarity (over sentences) ◦ Named Entity (NE) TFIDF cosine similarity (over sentences) ◦ NE SoftTFIDF cosine similarity (over sentences) ◦ Semantic similarity between the NPs of each chain  Train an SVM that classifies pairs of chains as co-referent or not  Use SVM to cluster all chains from all documents in the corpus.  Store this clustering in a database. Each entity  a list of chains

4 Augment our CDC system with the following:  Attribute Extraction ◦ For each chain, extract attributes such as gender, occupation, nationality, birthdates if they exist ◦ Use attributes to enable SVM to do better co-reference  Relationship Extraction ◦ For each pair of chains, extract a relationship (e.g. part of, role, etc), if it exists ◦ Use relationships for better visualization of clusters

5  Patterns: Take seed examples of (entity, attribute) and learn extraction patterns  Position: Use typical document positions of attributes  Transitive: Use attributes of neighboring entities  Latent: Use document-level topic models to infer attributes  References: Ravinchandran & Hovy ‘02, Garera & Yarowsky ‘09

6  Idea: Attribute values should appear with linguistic clues around it, i.e. it can be defined as a probabilistic language model describing the chance of a word being an attribute given the context. This idea is essentially the same as in KnowItAll or Brin ‘98.  Example: ◦ Marked Text: “ (born ) is a former American, active …” ◦ Generated Context: “(born X)” for X=, “a former American X” for X= …

7  Idea: Certain biographical attributes tend to appear in characteristic positions, often near top of article. Relative position /rank between attributes can be helpful information as well.  Example: “Birthdate” is often times the first date in a biography text, or at least near the beginning of the article, and relatively speaking “Deathdate” almost always occurs sometime afterwards

8  Idea: Intuition is that a named entity is more likely mentioned together with another entity with similar attribute values (the most applicable ones seem to be “occupation”, “religions”)  Example: “Michael Jordan” (the player) is mostly mentioned together with “Wilt Chamberlain”, “Dennis Rodman” and fellow players.

9  Idea: Use “latent wide-document-context” models to detect attributes that may not have been mentioned directly in article  Example: Words such as “songs, album, recorded” can all collectively indicate an occupation of singer or musician

10  Three options: ◦ Extract variety of features and train YFCL ◦ Define Kernels to measure similarity between instances, plug into SVM ◦ Semi-supervised approach. Start with seed examples, learn patterns, iterate. Grow KB (E.g. KnowItAll, TextRunner)  Ideally would prefer semi-supervised, especially since it allows open domain IE, but very time and labor-intensive  Kernel approach more elegant, works better than YFCL  Therefore, we proposed to use Kernels

11  Attributes: NNDB seed pairs, Wikipedia pages  Relations: ACE Phase 2 top level relations  For our CDC system: ◦ John Smith corpus ◦ WePS (person name disambiguation) corpora

12  Data Preparation ◦ Collected attribute extraction data set  Currently our seed pairs have occupation, date-of-birth, and date-of-death values  we plan on collecting birthplace, gender, and nationality ◦ Implemented the program to parse simple Wikipedia pages  Framework ◦ Implemented a pipe-line process which consists of  Loading the list of names and pages  Parsing Wikipedia pages  Sentence segmentation  POS tagging / NE Tagging (If needed)  Modeling and Extraction ◦ Provide capability to support different models for different attributes

13  Implementations ◦ Implemented context/pattern-based model, currently focusing on occupations ◦ Implemented position-based models (absolute & relative), on occupation and date of birth attribute extraction ◦ In progress of implementing transitivity-based model for occupation extraction ◦ In progress of implementing latent wide-document-context models for extracting implicit attributes

14  Issues Actual implementation raises a lot of problems missed in the paper such as whether POS Tagging should be included in the model, the pipeline framework is intended to solve this issue.  Plan ◦ To finish the implementations and compare with two baseline models for each attributes ◦ Extend code to work on chains and consolidate the attributes ◦ Use attributes extracted as features in the CDC system’s SVM classifier to help co-reference resolution

15  Evaluated different kinds of Kernels (subsequence, dependency trees, etc).  Chose subsequence kernels as these perform well and robust to ill-formed documents  As a template, decided to follow the Bunescu and Mooney paper discussed in class  Observation: Although the idea in the paper is easy to understand, implementation is quite complex

16  For SVM, decided to use LibSVM. Bunescu has a modified version that can take custom Kernels, but based on older release  Made updates to LibSVM to reconcile different versions. ◦ Now accepts custom/pre-computed Kernels ◦ Allows richer representation of instances than numerical values (e.g. sentence fragments, named entities, etc.)

17  As we know, hardest part of coding Machine Learning applications isn’t the classifier (plenty of libraries), but feature extraction  Wrote feature extraction code that extracts sentences from XML and produces instances for SVM, where each instance consists of: ◦ The 2 Named Entities of interest ◦ Sentence fragments before, between and after NEs  If a sentence consists of > 2 NEs, make N C 2 copies

18  Implemented general subsequence Kernel algorithm  Currently working on adapting this for the specific case of relationship Kernels (as explained in Bunescu & Mooney)  Once done, will extend this code (which works on sentences), to chains, so it can be plugged into CDC system

19


Download ppt "Bo Lin Kevin Dela Rosa Rushin Shah.  As part of our research, we are working on a cross- document co-reference resolution system  Co-reference Resolution:"

Similar presentations


Ads by Google