QuASI: Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley.

Slides:



Advertisements
Similar presentations
1 Probability and the Web Ken Baclawski Northeastern University VIStology, Inc.
Advertisements

CILC2011 A framework for structured knowledge extraction and representation from natural language via deep sentence analysis Stefania Costantini Niva Florio.
Learning Accurate, Compact, and Interpretable Tree Annotation Recent Advances in Parsing Technology WS 2011/2012 Saarland University in Saarbrücken Miloš.
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
QuASI: Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley.
GRAMMAR & PARSING (Syntactic Analysis) NLP- WEEK 4.
For Monday Read Chapter 23, sections 3-4 Homework –Chapter 23, exercises 1, 6, 14, 19 –Do them in order. Do NOT read ahead.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
Probabilistic Parsing: Enhancements Ling 571 Deep Processing Techniques for NLP January 26, 2011.
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
Search and Retrieval: More on Term Weighting and Document Ranking Prof. Marti Hearst SIMS 202, Lecture 22.
Automating Discovery from Biomedical Texts Marti Hearst & Barbara Rosario UC Berkeley Agyinc Visit August 16, 2000.
Are Linguists Dinosaurs? 1.Statistical language processors seem to be doing away with the need for linguists. –Why do we need linguists when a machine.
1/13 Parsing III Probabilistic Parsing and Conclusions.
1 Noun compounds (NCs) Any sequence of nouns that itself functions as a noun asthma hospitalizations asthma hospitalization rates health care personnel.
QuASI: Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley.
1 Classification of Semantic Relations in Noun Compounds using MeSH Marti Hearst, Barbara Rosario SIMS, UC Berkeley.
Classifying Semantic Relations in Bioscience Texts Barbara Rosario Marti Hearst SIMS, UC Berkeley Supported by NSF DBI
1/17 Probabilistic Parsing … and some other approaches.
XML on Semantic Web. Outline The Semantic Web Ontology XML Probabilistic DTD References.
1 Classification of Semantic Relations in Noun Compounds via a Domain-Specific Lexical Hierarchy Barbara Rosario, Marti Hearst SIMS, UC Berkeley.
تمرين شماره 1 درس NLP سيلابس درس NLP در دانشگاه هاي ديگر ___________________________ راحله مکي استاد درس: دکتر عبدالله زاده پاييز 85.
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
Machine Learning in Natural Language Processing Noriko Tomuro November 16, 2006.
Probabilistic Parsing Ling 571 Fei Xia Week 5: 10/25-10/27/05.
Semantic Interpretation of Medical Text Barbara Rosario, SIMS Steve Tu, UC Berkeley Advisor: Marti Hearst, SIMS.
1 The BioText Project Myers Seminar Sept 22, 2003 Marti Hearst Associate Professor SIMS, UC Berkeley Projected sponsored by NSF DBI , ARDA AQUAINT,
1 Basic Parsing with Context Free Grammars Chapter 13 September/October 2012 Lecture 6.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
BİL711 Natural Language Processing1 Statistical Parse Disambiguation Problem: –How do we disambiguate among a set of parses of a given sentence? –We want.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
For Friday Finish chapter 23 Homework: –Chapter 22, exercise 9.
The Descent of Hierarchy, and Selection in Relational Semantics* Barbara Rosario, Marti Hearst, Charles Fillmore UC Berkeley *with apologies to Charles.
Using Text Mining and Natural Language Processing for Health Care Claims Processing Cihan ÜNAL
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
Annotating Words using WordNet Semantic Glosses Julian Szymański Department of Computer Systems Architecture, Faculty of Electronics, Telecommunications.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
Introduction to CL & NLP CMSC April 1, 2003.
A Cascaded Finite-State Parser for German Michael Schiehlen Institut für Maschinelle Sprachverarbeitung Universität Stuttgart
인공지능 연구실 황명진 FSNLP Introduction. 2 The beginning Linguistic science 의 4 부분 –Cognitive side of how human acquire, produce, and understand.
1 CSI 5180: Topics in AI: Natural Language Processing, A Statistical Approach Instructor: Nathalie Japkowicz Objectives of.
Capturing patterns of linguistic interaction in a parsed corpus A methodological case study Sean Wallis Survey of English Usage University College London.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Coarse-to-Fine Efficient Viterbi Parsing Nathan Bodenstab OGI RPE Presentation May 8, 2006.
What you have learned and how you can use it : Grammars and Lexicons Parts I-III.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Supertagging CMSC Natural Language Processing January 31, 2006.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
Commonsense Reasoning in and over Natural Language Hugo Liu, Push Singh Media Laboratory of MIT The 8 th International Conference on Knowledge- Based Intelligent.
Levels of Linguistic Analysis
QuASI: Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley.
C H A P T E R T W O Linking Syntax And Semantics Programming Languages – Principles and Paradigms by Allen Tucker, Robert Noonan.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
PARSING David Kauchak CS159 – Fall Admin Assignment 3 Quiz #1  High: 36  Average: 33 (92%)  Median: 33.5 (93%)
Dependency Parsing Niranjan Balasubramanian March 24 th 2016 Credits: Many slides from: Michael Collins, Mausam, Chris Manning, COLNG 2014 Dependency Parsing.
Natural Language Processing Vasile Rus
CSC 594 Topics in AI – Natural Language Processing
PRESENTED BY: PEAR A BHUIYAN
QuASI: Question Answering using Statistics, Semantics, and Inference
Basic Parsing with Context Free Grammars Chapter 13
Machine Learning in Natural Language Processing
Levels of Linguistic Analysis
The Descent of Hierarchy, and Selection in Relational Semantics*
Parsing Unrestricted Text
David Kauchak CS159 – Spring 2019
Marti Hearst Associate Professor SIMS, UC Berkeley
CS249: Neural Language Model
Presentation transcript:

QuASI: Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley / ICSI / Stanford Univ

Outline Project Overview Three topics: Assigning semantic relations via lexical hierarchies From sentences to meanings via syntax From text analysis to inference using conceptual schemas

Main Goals Support Question-Answering and NLP in general by: Deepening our understanding of concepts that underlie all languages Creating empirical approaches to identifying semantic relations from free text Developing probabilistic inferencing algorithms

Two Main Thrusts Text-based: Use empirical corpus-based techniques to extract simple semantic relations Combine these relations to perform simple inferences “statistical semantic grammar” Concept-based: Determine language-universal conceptual principles Determine how inferences are made among these

Assigning Semantic Relations Using a Lexical Hierarchy

Noun Compounds (NCs) Any sequence of nouns that itself functions as a noun asthma hospitalizations asthma hospitalization rates health care personnel hand wash Technical text is rich with NCs Open-labeled long-term study of the subcutaneous sumatriptan efficacy and tolerability in acute migraine treatment.

NCs: 3 computational tasks (Lauer & Dras ’94) Identification Syntactic analysis (attachments) [Baseline [headache frequency]] [[Tension headache] patient] Semantic analysis Headache treatment treatment for headache Corticosteroid treatment treatment that uses corticosteroid

The lexical Hierarchy: MeSH Tree Structures 1. Anatomy [A] 2. Organisms [B] 3. Diseases [C] 4. Chemicals and Drugs [D] 5. Analytical, Diagnostic and Therapeutic Techniques and Equipment [E] 6. Psychiatry and Psychology [F] 7. Biological Sciences [G] 8. Physical Sciences [H] 9. Anthropology, Education, Sociology and Social Phenomena [I] 10. Technology and Food and Beverages [J] 11. Humanities [K] 12. Information Science [L] 13. Persons [M] 14. Health Care [N] 15. Geographic Locations [Z]

MeSH Tree Structures 1. Anatomy [A] Body Regions [A01] + Musculoskeletal System [A02] Digestive System [A03] + Respiratory System [A04] + Urogenital System [A05] + Endocrine System [A06] + Cardiovascular System [A07] + Nervous System [A08] + Sense Organs [A09] + Tissues [A10] + Cells [A11] + Fluids and Secretions [A12] + Animal Structures [A13] + Stomatognathic System [A14] (…..) Body Regions [A01] Abdomen [A01.047] Groin [A ] Inguinal Canal [A ] Peritoneum [A ] + Umbilicus [A ] Axilla [A01.133] Back [A01.176] + Breast [A01.236] + Buttocks [A01.258] Extremities [A01.378] + Head [A01.456] + Neck [A01.598] (….)

Mapping Nouns to MeSH Concepts headache recurrence C C headache pain C G breast cancer cells A C04 A11

Descent of Hierarchy Idea: Use the top levels of the lexical hierarchy to identify semantic relations Hypothesis: A particular semantic relations holds between all 2-word NCs that can be categorized by a category pair.

Linguistic Motivation Can cast NC into head-modifier relation, and assume head noun has an argument and qualia structure. (used-in): kitchen knife (made-of): steel knife (instrument-for): carving knife (used-on): putty knife (used-by): butcther’s knife

Distribution of Category Pairs

Classification decisions A01 N02 A01 N A01 N A01 N A01 N03 (W) J01 A01 (W) A02 F01 C04 H01 C04 H C04 H C04 H01.770

Levels of the classification decision Anatomy: 250 CPs 187 (75%) remain first level 56 (22%) descend one level 7 (3%) descend two levels Natural Science (H01): 21 CPs 1 (4%) remain first level 8 (39%) descend one level 12 (57%) descend two level Neoplasm (C04) 3 CPs: 3 (100%) descend one level

Evaluation Test decisions on “testing” set Count how many NCs that fall in the groups defined in the classification “rules” are similar with each other Accuracy: Anatomy: 91% accurate Natural Science: 79% Diseases: 100% Total: 89.6% via intra-category averaging 90.8% via extra-category averaging

Future work Analyze full spectrum of hierarchy NCs with > 2 terms [[growth hormone] deficiency] (purpose + defect) Other syntactic structures Non-biomedical words Other ontologies (e.g.,WordNet)?

From sentences to meanings via syntax A* Parsing & Stochastic HPSG

1. A* Parsing Goal: develop parsers that are Accurate – produce good parses Exact – find the models’ best parses Fast – seconds to parse long sentences Exhaustive Parsing – Slow but Exact e.g., chart Parsing, [Earley 70, Kay 80] Approximate Parsing – Fast but Inexact Beam Parsing, [Collins 97, Charniak 01] Best-First Parsing [Charniak et al. 98, etc.] Technology exists to get any two, but not all three of these goals

A* Search for Parsing Problem with uniform- cost parse search Even unlikely small edges have high score. We end up processing every small edge! Solution: A* Estimates Small edges have to fit into a full parse. The smaller the edge, the more the full parse will cost! Score =  Score =  +    

The Estimate Trade- off The more we specify, the better estimate of  we get… Fix outside size: Score = Add left tag: Score = Add right tag: Score = Entire context gives the exact best parse. Score = -18.1

A* Savings: Penn Treebank (SX-F filters more than Caraballo and Charniak 1998 while guaranteeing optimality, but less than Charniak et al. 1998)

2. Stochastic HPSG / Redwoods Treebank The Redwoods treebank is being built at Stanford as a resource for for deep NLP Provides full HPSG (Head-driven Phrase Structure Grammar) analyses, including semantic logical forms Current corpus is spoken dialog data (Verbmobil) parsed by robust broad coverage HPSG grammar Information at different levels of detail can be extracted from the treebank Precise deep grammatical analyses can be combined with probabilistic models Procedures are being developed for automatically updating the treebank

Basic Representation Levels Derivation tree of lexical items and constructions Phrase Structure Tree (S (NP “I”)(VP (V “am”) (ADJ_P “sorry”))) Underspecified MRS meaning representation <e1:BOOL:INDICATIVE*:PRESENT*:STRICT_NONPRF, {h2: pron_rel(x3:-*:STD_1SG:1SG:GENDER), h4: def_rel(x3,h5,h6,v7:BOOL),h8:_sorry_rel(e1,x3,v9:BOOL,v10:BOOL),h11: prpstn_rel(h12)}, {h5 QEQ h2, h12 QEQ h8}> Full HPSG signs for sentences are available IHCOMP “am” “I” BE_C_AM “sorry” SORRY_A1 SUBJH

Initial exploration: PCFG Results Accuracy Complete match (parse selection) accuracy Models with increasing parent node annotation PC FG Log lin ear Generative PCFG vs. loglinear modeling

In progress work: Using semantic forms Example: “I need to meet with you again”. I need to [[meet with you] again] (preferred) (ii) I [need to [meet with you] again] People use semantic information to disambiguate Building random field models over relations _need2_rel _meet_v_rel pron_rel with_rel again_rel pron_rel _need2_rel pron_rel with_rel again_relpron_rel _meet_v_rel

Concept-based Analysis From text analysis to inference using conceptual schemas

Inference and Conceptual Schemas Hypothesis: Linguistic input is converted into a mental simulation based on bodily-grounded structures. Components: Semantic schemas image schemas and executing schemas are abstractions over neurally grounded perceptual and motor representations Linguistic units lexical and phrasal construction representations invoke schemas, in part through metaphor Inference links these structures and provides parameters for a simulation engine

Conceptual Schemas Much is known about conceptual schemas, particularly image schemas However, this understanding has not yet been formalized We will develop such a formalism They have also not been checked extensively against other languages We will examine Chinese, Russian, and other languages in addition to English

Schema Formalism SCHEMA SUBCASE OF EVOKES AS ROLES : CONSTRAINTS :: :: |

A Simple Example SCHEMA hypotenuse SUBCASE OF line-segment EVOKES right-triangle AS rt ROLES Comment inherited from line-segment CONSTRAINTS SELF rt.long-side

Source-Path-Goal SCHEMA: spg ROLES: source: Place path: Directed Curve goal: Place trajector: Entity

Translational Motion SCHEMA translational motion SUBCASE OF motion EVOKES spg AS s ROLES mover s.trajector source s.source goal s.goal CONSTRAINTS before:: mover.location source after:: mover.location goal

Extending Inferential Capabilities Given the formalization of the conceptual schemas How to use them for inferencing? Earlier pilot systems Used metaphor and Bayesian belief networks Successfully construed certain inferences But don’t scale New approach Probabilistic relational models Support an open ontology

A Common Representation Representation should support Uncertainty, probability Conflicts, contradictions Current plan Probabilistic Relational Models (Koller et al.) DAML + OIL

An Open Ontology for Conceptual Relations Build a formal markup language for conceptual schemas We propose to use DAML+OIL as the base. Advantages of the approach Common framework for extending and reuse Closer ties to other efforts within AQUAINT as well as the larger research community on the Semantic Web. Some Issues Expressiveness of DAML+OIL Representing Probabilistic Information Extension to MetaNet, capture abstract concepts

DAML-I: An Image Schema Markup Language A basic type of schema <daml:subPropertyOf rdf:resource="&conc-rel;#role"/

Putting it all Together We have proposed two different types of semantics Universal conceptual schemas Semantic relations In Phase I they will remain separate However, we are exploring using PRMs as a common representational format In later Phases they will be combined