Download presentation
Presentation is loading. Please wait.
1
Classifying Semantic Relations in Bioscience Texts Barbara Rosario Marti Hearst SIMS, UC Berkeley http://biotext.berkeley.edu Supported by NSF DBI-0317510 and a gift from Genentech
2
Problem: Which relations hold between 2 entities? TreatmentDisease Cure? Prevent? Side Effect?
3
Hepatitis Examples Cure These results suggest that con A-induced hepatitis was ameliorated by pretreatment with TJ-135. Prevent A two-dose combined hepatitis A and B vaccine would facilitate immunization programs Vague Effect of interferon on hepatitis B
4
Two tasks Relationship Extraction: Identify the several semantic relations that can occur between the entities disease and treatment in bioscience text Entity extraction: Related problem: identify such entities
5
The Approach Data: MEDLINE abstracts and titles Graphical models Combine in one framework both relation and entity extraction Both static and dynamic models Simple discriminative approach: Neural network Lexical, syntactic and semantic features
6
Outline Related work Data and semantic relations Features Models and results Conclusions
7
Several DIFFERENT Relations between the Same Types of Entities Thus differs from the problem statement of other work on relations Many find one relation which holds between two entities (many based on ACE) Agichtein and Gravano (2000), lexical patterns for location of Zelenko et al. (2002) SVM for person affiliation and organization-location Hasegawa et al. (ACL 2004) Person- Organization -> President “relation” Craven (1999, 2001) HMM for subcellular- location and disorder-association Doesn’t identify the actual relation
8
Related work: Bioscience Many hand-built rules Feldman et al. (2002), Friedman et al. (2001) Pustejovsky et al. (2002) Saric et al.; this conference
9
Data and Relations MEDLINE, abstracts and titles 3662 sentences labeled Relevant: 1724 Irrelevant: 1771 e.g., “Patients were followed up for 6 months” 2 types of Entities, many instances treatment and disease 7 Relationships between these entities The labeled data is available at http://biotext.berkeley.edu
10
Semantic Relationships 810: Cure Intravenous immune globulin for recurrent spontaneous abortion 616: Only Disease Social ties and susceptibility to the common cold 166: Only Treatment Flucticasone propionate is safe in recommended doses 63: Prevent Statins for prevention of stroke
11
Semantic Relationships 36: Vague Phenylbutazone and leukemia 29: Side Effect Malignant mesodermal mixed tumor of the uterus following irradiation 4: Does NOT cure Evidence for double resistance to permethrin and malathion in head lice
12
Features Word Part of speech Phrase constituent Orthographic features ‘is number’, ‘all letters are capitalized’, ‘first letter is capitalized’ … MeSH (semantic features) Replace words, or sequences of words, with generalizations via MeSH categories Peritoneum -> Abdomen
13
Features (cont.): MeSH MeSH Tree Structures 1. Anatomy [A] 2. Organisms [B] 3. Diseases [C] 4. Chemicals and Drugs [D] 5. Analytical, Diagnostic and Therapeutic Techniques and Equipment [E] 6. Psychiatry and Psychology [F] 7. Biological Sciences [G] 8. Physical Sciences [H] 9. Anthropology, Education, Sociology and Social Phenomena [I] 10. Technology and Food and Beverages [J] 11. Humanities [K] 12. Information Science [L] 13. Persons [M] 14. Health Care [N] 15. Geographic Locations [Z]
14
Features (cont.): MeSH 1. Anatomy [A] Body Regions [A01] + Musculoskeletal System [A02] Digestive System [A03] + Respiratory System [A04] + Urogenital System [A05] + Endocrine System [A06] + Cardiovascular System [A07] + Nervous System [A08] + Sense Organs [A09] + Tissues [A10] + Cells [A11] + Fluids and Secretions [A12] + Animal Structures [A13] + Stomatognathic System [A14] (…..) Body Regions [A01] Abdomen [A01.047] Groin [A01.047.365] Inguinal Canal [A01.047.412] Peritoneum [A01.047.596] + Umbilicus [A01.047.849] Axilla [A01.133] Back [A01.176] + Breast [A01.236] + Buttocks [A01.258] Extremities [A01.378] + Head [A01.456] + Neck [A01.598] (….)
15
Models 2 static generative models 3 dynamic generative models 1 discriminative model (neural networks)
16
Static Graphical Models S1: observations dependent on Role but independent from Relation given roles S2: observations dependent on both Relation and Role S1S2
17
Dynamic Graphical Models D1, D2 as in S1, S2 D3: only one observation per state is dependent on both the relation and the role D1 D2 D3
18
Graphical Models Relation node: Semantic relation (cure, prevent, none..) expressed in the sentence
19
Graphical Models Role nodes: 3 choices: treatment, disease, or none
20
Graphical Models Feature nodes (observed): word, POS, MeSH…
21
Graphical Models Different dependencies between the features and the relation nodes D3 D1 S1 D2 S2
22
Graphical Models For Dynamic Model D1: Joint probability distribution over relation, roles and features nodes Parameters estimated with maximum likelihood and absolute discounting smoothing
23
Our D1 Thompson et al. 2003 Frame classification and role labeling for FrameNet sentences Target word must be observed More relations and roles
24
Neural Networks Feed-forward network (MATLAB) Training with conjugate gradient descent One hidden layer (hyperbolic tangent function) Logistic sigmoid function for the output layer representing the relationships Same features Discriminative approach
25
Relation extraction Results in terms of classification accuracy (with and without irrelevant sentences) 2 cases: Roles hidden Roles given Graphical models NN: simple classification problem
26
Relation classification: Results Neural Net always best
27
Relation classification: Results With no smoothing, D1 best Graphical Model
28
Relation classification: Results With Smoothing and No Roles, D2 best GM
29
Relation classification: Results With Smoothing and Roles, D1 best GM
30
Relation classification: Results Dynamic models always outperform Static
31
Relation classification: Confusion Matrix Computed for the model D2, “rel + irrel.”, “only features”
32
Role extraction Results in terms of F-measure Graphical models Junction tree algorithm (BNT) Relation hidden and marginalized over NN Couldn’t run it (features vectors too large) (Graphical models can do role extraction and relationship classification simultaneously)
33
Role Extraction: Results F-measures D1 best when no smoothing
34
Role Extraction: Results F-measures D2 best with smoothing, but doesn’t boost scores as much as in relation classification
35
Features impact: Role Extraction Most important features: 1)Word, 2)MeSH Models D1 D2 All features 0.67 0.71 No word 0.58 0.61 -13.4% -14.1% No MeSH 0.63 0.65 -5.9% -8.4% (rel. + irrel.)
36
Most important features: Roles Accuracy: D1 D2 NN All feat. + roles 91.6 82.0 96.9 All feat. – roles 68.9 74.9 79.6 -24.7% -8.7% -17.8% All feat. + roles – Word 91.6 79.8 96.4 0% -2.8% -0.5% All feat. + roles – MeSH 91.6 84.6 97.3 0% 3.1% 0.4% Features impact: Relation classification (rel. + irrel.)
37
Features impact: Relation classification Most realistic case: Roles not known Most important features: 1) Mesh 2) Word for D1 and NN (but vice versa for D2) Accuracy: D1 D2 NN All feat. – roles 68.9 74.9 79.6 All feat. - roles – Word 66.7 66.1 76.2 -3.3% -11.8% -4.3% All feat. - roles – MeSH 62.7 72.5 74.1 -9.1% -3.2% -6.9% (rel. + irrel.)
38
Conclusions Classification of subtle semantic relations in bioscience text Discriminative model (neural network) achieves high classification accuracy Graphical models for the simultaneous extraction of entities and relationships Importance of lexical hierarchy Future work: A new collection of disease/treatment data Different entities/relations Unsupervised learning to discover relation types
39
Thank you! Barbara Rosario Marti Hearst SIMS, UC Berkeley http://biotext.berkeley.edu
40
Additional slides
41
Smoothing: absolute discounting Lower the probability of seen events by subtracting a constant from their count (ML estimate: ) The remaining probability is evenly divided by the unseen events
42
F-measures for role extraction in function of smoothing factors
43
Relation accuracies in function of smoothing factors
44
Role Extraction: Results Static models better than Dynamic for Note: No Neural Networks
45
Features impact: Role Extraction Most important features: 1)Word, 2)MeSH Models D1 D2 Average All features 0.67 0.71 No word 0.58 0.61 -13.4% -14.1% -13.7% No MeSH 0.63 0.65 -5.9% -8.4% -7.2% (rel. + irrel.)
46
Features impact: Role extraction Most important features: 1) Word, 2) MeSH F-measures: D1 D2 Average All features 0.72 0.73 No word 0.65 0.66 -9.7% -9.6% -9.6% No MeSH 0.69 0.69 -4.2% -5.5% -4.8% (only rel.)
47
Features impact: Role extraction Most important features: 1) Word, 2) MeSH F-measures: D1 D2 All features 0.72 0.73 No word 0.65 0.66 -9.7% -9.6% No MeSH 0.69 0.69 -4.2% -5.5% (only rel.)
48
Most important features: Roles Accuracy: D1 D2 NN Avg. All feat. + roles 91.6 82.0 96.9 All feat. – roles 68.9 74.9 79.6 -24.7% -8.7% -17.8% -17.1% All feat. + roles – Word 91.6 79.8 96.4 0% -2.8% -0.5% -1.1% All feat. + roles – MeSH 91.6 84.6 97.3 0% 3.1% 0.4% 1.1% Features impact: Relation classification (rel. + irrel.)
49
Features impact: Relation classification Most realistic case: Roles not known Most important features: 1) Mesh 2) Word for D1 and NN (but vice versa for D2) Accuracy: D1 D2 NN Avg. All feat. – roles 68.9 74.9 79.6 All feat. - roles – Word 66.7 66.1 76.2 -3.3% -11.8% -4.3% -6.4% All feat. - roles – MeSH 62.7 72.5 74.1 -9.1% -3.2% -6.9% -6.4% (rel. + irrel.)
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.