Computational Lexical Semantics Lecture 8: Selectional Restrictions Linguistic Institute 2005 University of Chicago.

Slides:



Advertisements
Similar presentations
Semantics (Representing Meaning)
Advertisements

N-gram model limitations Important question was asked in class: what do we do about N-grams which were not in our training corpus? Answer given: we distribute.
A UTOMATICALLY A CQUIRING A S EMANTIC N ETWORK OF R ELATED C ONCEPTS Date: 2011/11/14 Source: Sean Szumlanski et. al (CIKM’10) Advisor: Jia-ling, Koh Speaker:
The Interaction of Lexical and Syntactic Ambiguity by Maryellen C. MacDonald presented by Joshua Johanson.
Non Monotonic Reasoning for Disambiguation and Parsing Arthi Murugesan.
Syntax Lecture 9: Verb Types 2.
Learning When (and When Not) to Omit Objects in English: The Role of Verb Semantic Selectivity Tamara Nicol Medina IRCS, University of Pennsylvania Collaborators:
Statistical NLP: Lecture 3
SI485i : NLP Set 11 Distributional Similarity slides adapted from Dan Jurafsky and Bill MacCartney.
Introduction and Jurafsky Model Resource: A Probabilistic Model of Lexical and Syntactic Access and Disambiguation, Jurafsky 1996.
LEDIR : An Unsupervised Algorithm for Learning Directionality of Inference Rules Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From EMNLP.
Automatic Metaphor Interpretation as a Paraphrasing Task Ekaterina Shutova Computer Lab, University of Cambridge NAACL 2010.
Introduction to Semantics To be able to reason about the meanings of utterances, we need to have ways of representing the meanings of utterances. A formal.
November 20, 2003 Chapter 16 Lexical Semantics. Words have structured meanings Lexeme – a pairing of a form with a sense Orthographic form – the way the.
1 SIMS 290-2: Applied Natural Language Processing Marti Hearst Sept 22, 2004.
Distributional clustering of English words Authors: Fernando Pereira, Naftali Tishby, Lillian Lee Presenter: Marian Olteanu.
1 Reading time evidence for enriched composition McElree et al. (2001) Rianne Oostwoud-Wijdenes & Maartje Schulpen.
Using Information Content to Evaluate Semantic Similarity in a Taxonomy Presenter: Cosmin Adrian Bejan Philip Resnik Sun Microsystems Laboratories.
The Game of Algebra or The Other Side of Arithmetic The Game of Algebra or The Other Side of Arithmetic © 2007 Herbert I. Gross by Herbert I. Gross & Richard.
Chemical Reactions & Chemical Equations
Albert Gatt Corpora and Statistical Methods Lecture 5.
The time-course of prediction in incremental sentence processing: Evidence from anticipatory eye movements Yuki Kamide, Gerry T.M. Altman, and Sarah L.
Albert Gatt LIN 3098 Corpus Linguistics. In this lecture Some more on corpora and grammar Construction Grammar as a theoretical framework Collostructional.
Syntax Lecture 8: Verb Types 1. Introduction We have seen: – The subject starts off close to the verb, but moves to specifier of IP – The verb starts.
1 Statistical NLP: Lecture 10 Lexical Acquisition.
Probabilistic Parsing Reading: Chap 14, Jurafsky & Martin This slide set was adapted from J. Martin, U. Colorado Instructor: Paul Tarau, based on Rada.
“How much context do you need?” An experiment about context size in Interactive Cross-language Question Answering B. Navarro, L. Moreno-Monteagudo, E.
中文信息处理 Chinese NLP Lecture 11.
CAS LX 502 8b. Formal semantics A fragment of English.
Multiplying Whole Numbers © Math As A Second Language All Rights Reserved next #5 Taking the Fear out of Math 9 × 9 81 Single Digit Multiplication.
Improving Subcategorization Acquisition using Word Sense Disambiguation Anna Korhonen and Judith Preiss University of Cambridge, Computer Laboratory 15.
Lexical Semantics Chapter 16
Unaddition (Subtraction)
Lecture 22 Word Similarity Topics word similarity Thesaurus based word similarity Intro. Distributional based word similarityReadings: NLTK book Chapter.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
Modelling Human Thematic Fit Judgments IGK Colloquium 3/2/2005 Ulrike Padó.
11 Chapter 19 Lexical Semantics. 2 Lexical Ambiguity Most words in natural languages have multiple possible meanings. –“pen” (noun) The dog is in the.
Introduction to Human Biology. What do you think is involved in a healthy lifestyle?
Annotation for Hindi PropBank. Outline Introduction to the project Basic linguistic concepts – Verb & Argument – Making information explicit – Null arguments.
Artificial Intelligence: Natural Language
1 Statistical NLP: Lecture 7 Collocations. 2 Introduction 4 Collocations are characterized by limited compositionality. 4 Large overlap between the concepts.
Albert Gatt LIN3021 Formal Semantics Lecture 4. In this lecture Compositionality in Natural Langauge revisited: The role of types The typed lambda calculus.
Lecture 21 Computational Lexical Semantics Topics Features in NLTK III Computational Lexical Semantics Semantic Web USCReadings: NLTK book Chapter 10 Text.
Presupposition is what the speaker assumes to be the case prior to making an utterance. Entailment, which is not a pragmatic concept, is what logically.
Technical Science Scientific Tools and Methods Tables and Graphs.
Estimating N-gram Probabilities Language Modeling.
VERBS. VERB A word that expresses an action or state of being.
Dec 11, Human Parsing Do people use probabilities for parsing?! Sentence processing Study of Human Parsing.
FILTERED RANKING FOR BOOTSTRAPPING IN EVENT EXTRACTION Shasha Liao Ralph York University.
1 Introduction to Computational Linguistics Eleni Miltsakaki AUTH Fall 2005-Lecture 3.
Knowledge Structure Vijay Meena ( ) Gaurav Meena ( )
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
Strategies for Essay Tests. Preparing for the test Know what is expected of you. What content will be covered? How many questions will be on the test?
Chapter 8 Lexical Acquisition February 19, 2007 Additional Notes to Manning’s slides.
Using Tag Questions Using Tag Questions. She's very beautiful, isn't she?
A Database of Narrative Schemas A 2010 paper by Nathaniel Chambers and Dan Jurafsky Presentation by Julia Kelly.
Lec. 10.  In this section we explain which constituents of a sentence are minimally required, and why. We first provide an informal discussion and then.
Lecture 6: More On Wh-movement
PRESUPPOSITION AND ENTAILMENT
Sentences as Arguments
What Is a Test of Significance?
Syntax Lecture 9: Verb Types 1.
Statistical NLP: Lecture 3
Syntax Lecture 10: Verb Types 2.
CSC 594 Topics in AI – Applied Natural Language Processing
WordNet: A Lexical Database for English
THE NATURE OF SCIENCE.
Technical Science Scientific Tools and Methods
Syntax Lecture 12: Extended VP.
Statistical NLP: Lecture 10
Presentation transcript:

Computational Lexical Semantics Lecture 8: Selectional Restrictions Linguistic Institute 2005 University of Chicago

Selectional Restrictions Introduction

Dan Jurafsky Selectional Restrictions Consider the two interpretations of: I want to eat someplace nearby. a)sensible: Eat is intransitive and “someplace nearby” is a location adjunct b)Speaker is Godzilla Eat is transitive and “someplace nearby” is a direct object How do we know speaker didn’t mean b) ? Because the T HEME of eating tends to be something edible 3

Dan Jurafsky Selectional restrictions are associated with senses The restaurant serves green-lipped mussels. T HEME is some kind of food Which airlines serve Denver? T HEME is an appropriate location 4

Dan Jurafsky Selectional restrictions vary in specificity I often ask the musicians to imagine a tennis game. To diagonalize a matrix is to find its eigenvalues. Radon is an odorless gas that can’t be detected by human senses. 5

Dan Jurafsky Representing selectional restrictions 6 Instead of representing “eat” as: Just add: And “eat a hamburger” becomes But this assumes we have a large knowledge base of facts about edible things and hamburgers and whatnot.

Dan Jurafsky Let’s use WordNet synsets to specify selectional restrictions The THEME of eat must be WordNet synset {food, nutrient} “any substance that can be metabolized by an animal to give energy and build tissue” Similarly THEME of imagine: synset {entity} THEME of lift: synset {physical entity} THEME of diagonalize: synset {matrix} This allows imagine a hamburger and lift a hamburger, Correctly rules out diagonalize a hamburger. 7

Selectional Restrictions Selectional Preferences

Dan Jurafsky Selectional Preferences In early implementations, selectional restrictions were strict constraints (Katz and Fodor 1963) Eat [+FOOD] But it was quickly realized selectional constraints are really preferences (Wilks 1975) But it fell apart in 1931, perhaps because people realized you can’t eat gold for lunch if you’re hungry. In his two championship trials, Mr. Kulkarni ate glass on an empty stomach, accompanied only by water and tea. 9

Dan Jurafsky Selectional Association (Resnik 1993) Selectional preference strength: amount of information that a predicate tells us about the semantic class of its arguments. eat tells us a lot about the semantic class of its direct objects be doesn’t tell us much The selectional preference strength difference in information between two distributions: P(c) the distribution of expected semantic classes for any direct object P(c|v) the distribution of expected semantic classes for this verb The greater the difference, the more the verb is constraining its object 10

Dan Jurafsky Selectional preference strength Relative entropy, or the Kullback-Leibler divergence is the difference between two distributions Selectional preference: How much information (in bits) the verb expresses about the semantic class of its argument Selectional Association of a verb with a class: The relative contribution of the class to the general preference of the verb 11

Dan Jurafsky Computing Selectional Association A probabilistic measure of the strength of association between a predicate and a semantic class of its argument Parse a corpus Count all the times each predicate appears with each argument word Assume each word is a partial observation of all the WordNet concepts associated with that word Some high and low associations: 12

Dan Jurafsky Results from similar models 13 Ó Séaghdha and Korhonen (2012)

Dan Jurafsky Instead of using classes, a simpler model of selectional association Model just the association of predicate v with a noun n (one noun, as opposed to the whole semantic class in WordNet) Parse a huge corpus Count how often a noun n occurs in relation r with verb v: log count(n,v,r) Or the probability: 14

Dan Jurafsky Evaluation from Bergsma, Lin, Goebel 15

Selectional Restrictions Conclusion

Dan Jurafsky Summary: Selectional Restrictions Two classes of models of the semantic type constraint that a predicate places on its argument: Represent the constraint between predicate and WordNet class Represent the constraint between predicate and a word One fun recent use case: detecting metonomy (type coercion) Coherent with selectional restrictions: The spokesman denied the statement (PROPOSITION). The child threw the stone (PHYSICAL OBJECT) Coercion: The president denied the attack (EVENT → PROPOSITION). The White House (LOCATION → HUMAN) denied the statement. 17 Pustejovsky et al (2010)