Download presentation
Presentation is loading. Please wait.
Published byMaryann Boyd Modified over 8 years ago
1
Authors: Yutaka Matsuo & Mitsuru Ishizuka Designed by CProDM Team
2
Introduction Algorithm implement Evaluation Outline Study Algorithm
3
Introduction Discard stop words Stem Extract frequency Select frequent term Clustering Expected probability Calculate X’ 2 value Output
4
Study Algorithm Preprocessing Goal: - Remove unnecessary words in document. - Get terms which are candidate keywords. Stop word: the function words and, the, and of, or other words with minimal lexical meaning. Stem: remove suffixes from words
5
Discard stop words It might be urged that when playing the “imitation game" the best strategy for the machine may possibly be something other than imitation of the behaviour of a man. This may be, but I think it is unlikely that there is any great effect of this kind. In any case there is no intention to investigate here the theory of the 2 game, and it will be assumed that the best strategy is to try to provide answers that would naturally be given by a man. urged playing “imitation game" best strategy machine possibly imitation behaviour man think unlikely great effect kind. case intention investigate theory 2 game, assumed best strategy try provide answers naturally given man. Study Algorithm Preprocessing
6
urged playing “imitation game" best strategy machine possibly imitation behaviour man think unlikely great effect kind. case intention investigate theory game, assumed best strategy try provide answers naturally given man. Stem urge play “imitation game" best strategi machine possible imitation behaviour man think unlike great effect kind. case intention investigate theory game, assum best strategi try provide answers natural give man. Study Algorithm Preprocessing
7
imitation best strategi man best strategi Extract frequency urge play “imitation game" best strategi machine possible imitation behaviour man think unlike great effect kind. case intention investigate theory game, assum best strategi try provide answers natural give man. Study Algorithm Preprocessing
8
Study Algorithm Term Co-occurrence and Importance the top ten frequent terms (denoted as ) and the probability of occurrence, normalized so that the sum is to be 1
9
Study Algorithm Term Co-occurrence and Importance Two terms in a sentence are considered to co-occur once.
10
co-occurrence probability distribution of some terms and the frequent terms. Study Algorithm Term Co-occurrence and Importance
11
The statistical value of χ2 is defined as P g Unconditional probability of a frequent term g ∈ G (the expected probability) N w The total number of co-occurrence of term w and frequent terms G freq (w, g) Frequency of co-occurrence of term w and term g Study Algorithm Term Co-occurrence and Importance
12
Study Algorithm Term Co-occurrence and Importance
13
Study Algorithm Algorithm improvement P g (the sum of the total number of terms in sentences where g appears) divided by (the total number of terms in the document) N w The total number of terms in the sentences where w appears including w If a term appears in a long sentence, it is likely to co-occur with many terms; if a term appears in a short sentence, it is less likely to co-occur with other terms. We consider the length of each sentence and revise our definitions
14
the following function to measure robustness of bias values Study Algorithm Algorithm improvement
15
To improve extracted keyword quality, we will cluster terms Two major approaches (Hofmann & Puzicha 1998) are: Similarity-based clustering If terms w1 and w2 have similar distribution of co- occurrence with other terms, w1 and w2 are considered to be the same cluster. Pairwise clustering If terms w1 and w2 co-occur frequently, w1 and w2 are considered to be the same cluster. Study Algorithm Algorithm improvement
16
Similarity-based clustering centers upon Red Circles Pairwise clustering focuses on Yellow Circles Study Algorithm Algorithm improvement
17
Where: Similarity-based clustering Cluster a pair of terms whose Jensen-Shannon divergence is and: Study Algorithm Algorithm improvement
18
Cluster a pair of terms whose mutual information is Pairwise clustering Where: Study Algorithm Algorithm improvement
19
Study Algorithm Algorithm improvement
20
Algorithm Implement
21
Discard stop words Stem Extract frequency Algorithm Implement Step 1: Preprocessing
22
Algorithm Implement Step 2: Selection of frequent terms Select the top frequent terms up to 30% of the number of running terms as a standard set of terms Count number of terms in document (Ntotal )
23
Algorithm Implement Step 3: Clustering frequent terms Similarity-base clustering Pairwise clustering
24
Algorithm Implement Step 4: Calculate expected probability Count the number of terms co-occurring with c ∈ C, denoted as n c, to yield the expected probability
25
Algorithm Implement Step 5: Calculate χ’2 value Where: the number of co-occurrence frequency with c ∈ C the total number of terms in the sentences including w
26
Algorithm Implement Step 6: Output keywords
27
Evaluation
28
In this paper, we developed an algorithm to extract keywords from a single document. Main advantages of our method are its simplicity without requiring use of a corpus and its high performance comparable to tfidf algorithm. As more electronic documents become available, we believe our method will be useful in many applications, especially for domain- independent keyword extraction.
29
Thank for your attention
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.