LIN3021 Formal Semantics Lecture 2

Slides:



Advertisements
Similar presentations
Artificial Intelligence
Advertisements

1 Knowledge Representation Introduction KR and Logic.
Comparative Succinctness of KR Formalisms Paolo Liberatore.
Inference and Reasoning. Basic Idea Given a set of statements, does a new statement logically follow from this. For example If an animal has wings and.
Syntax Lecture 10: Auxiliaries. Types of auxiliary verb Modal auxiliaries belong to the category of inflection – They are in complementary distribution.
Basic Structures: Sets, Functions, Sequences, Sums, and Matrices
L41 Lecture 2: Predicates and Quantifiers.. L42 Agenda Predicates and Quantifiers –Existential Quantifier  –Universal Quantifier 
Albert Gatt LIN1180/LIN5082 Semantics Lecture 2. Goals of this lecture Semantics -- LIN 1180 To introduce some of the central concepts that semanticists.
CAS LX 502 8a. Formal semantics Truth and meaning The basis of formal semantics: knowing the meaning of a sentence is knowing under what conditions.
Albert Gatt LIN3021 Formal Semantics Lecture 5. In this lecture Modification: How adjectives modify nouns The problem of vagueness Different types of.
Knowledge Representation Methods
Albert Gatt LIN1180 – Semantics Lecture 10. Part 1 (from last week) Theories of presupposition: the semantics- pragmatics interface.
Linguistic Theory Lecture 8 Meaning and Grammar. A brief history In classical and traditional grammar not much distinction was made between grammar and.
Copyright © Cengage Learning. All rights reserved. CHAPTER 1 SPEAKING MATHEMATICALLY SPEAKING MATHEMATICALLY.
Comp 205: Comparative Programming Languages Semantics of Imperative Programming Languages denotational semantics operational semantics logical semantics.
1 Introduction to Computability Theory Lecture12: Reductions Prof. Amos Israeli.
CAS LX 502 Semantics 1b. The Truth Ch. 1.
CAS LX a. A notational holiday. Sets A set is a collection of entities of any kind. They can be finite: {√2, John Saeed, 1984}. They can be infinite:
Let remember from the previous lesson what is Knowledge representation
CS 330 Programming Languages 09 / 16 / 2008 Instructor: Michael Eckmann.
EE1J2 - Slide 1 EE1J2 – Discrete Maths Lecture 6 Limitations of propositional logic Introduction to predicate logic Symbols, terms and formulae, Parse.
LING 364: Introduction to Formal Semantics Lecture 5 January 26th.
Meaning and Language Part 1.
CAS LX 502 Semantics 2b. A formalism for meaning 2.5, 3.2, 3.6.
Lecture 1 Introduction: Linguistic Theory and Theories
Copyright © Cengage Learning. All rights reserved.
What makes communication by language possible? Striking fact (a) If someone utters a sentence and you know which proposition her utterance expresses, then.
Linguistic Theory Lecture 3 Movement. A brief history of movement Movements as ‘special rules’ proposed to capture facts that phrase structure rules cannot.
What makes communication by language possible? Striking fact (a) If someone utters a sentence and you know which proposition her utterance expresses, then.
February 2009Introduction to Semantics1 Logic, Representation and Inference Introduction to Semantics What is semantics for? Role of FOL Montague Approach.
The Game of Algebra or The Other Side of Arithmetic The Game of Algebra or The Other Side of Arithmetic © 2007 Herbert I. Gross by Herbert I. Gross & Richard.
Albert Gatt LIN 3098 Corpus Linguistics. In this lecture Some more on corpora and grammar Construction Grammar as a theoretical framework Collostructional.
CAS LX 502 Semantics 3a. A formalism for meaning (cont ’ d) 3.2, 3.6.
LIN1180/LIN5082 Semantics Lecture 3
CAS LX 502 8b. Formal semantics A fragment of English.
November 2003CSA4050: Semantics I1 CSA4050: Advanced Topics in NLP Semantics I What is semantics for? Role of FOL Montague Approach.
1 LIN 1310B Introduction to Linguistics Prof: Nikolay Slavkov TA: Qinghua Tang CLASS 13, Feb 16, 2007.
Declarative vs Procedural Programming  Procedural programming requires that – the programmer tell the computer what to do. That is, how to get the output.
CMPF144 FUNDAMENTALS OF COMPUTING THEORY Module 5: Classical Logic.
Pattern-directed inference systems
Formal Semantics Chapter Twenty-ThreeModern Programming Languages, 2nd ed.1.
LOGIC AND ONTOLOGY Both logic and ontology are important areas of philosophy covering large, diverse, and active research projects. These two areas overlap.
Unit 5 : PREDICATES.
Copyright © Cengage Learning. All rights reserved. CHAPTER 3 THE LOGIC OF QUANTIFIED STATEMENTS THE LOGIC OF QUANTIFIED STATEMENTS.
Propositional Calculus CS 270: Mathematical Foundations of Computer Science Jeremy Johnson.
Semantic Construction lecture 2. Semantic Construction Is there a systematic way of constructing semantic representation from a sentence of English? This.
Albert Gatt LIN3021 Formal Semantics Lecture 4. In this lecture Compositionality in Natural Langauge revisited: The role of types The typed lambda calculus.
LECTURE 2: SEMANTICS IN LINGUISTICS
Key Concepts Representation Inference Semantics Discourse Pragmatics Computation.
Programming Languages and Design Lecture 3 Semantic Specifications of Programming Languages Instructor: Li Ma Department of Computer Science Texas Southern.
1 CA 208 Logic PQ PQPQPQPQPQPQPQPQ
Rules, Movement, Ambiguity
Theory and Applications
BBI 3420 Critical Reading and Thinking Critical Reading Strategies: Identifying Arguments.
Naïve Set Theory. Basic Definitions Naïve set theory is the non-axiomatic treatment of set theory. In the axiomatic treatment, which we will only allude.
1 LIN 1310B Introduction to Linguistics Prof: Nikolay Slavkov TA: Qinghua Tang CLASS 16, March 6, 2007.
Topic and the Representation of Discourse Content
LDK R Logics for Data and Knowledge Representation Propositional Logic Originally by Alessandro Agostini and Fausto Giunchiglia Modified by Fausto Giunchiglia,
Albert Gatt LIN3021 Formal Semantics Lecture 3. Aims This lecture is divided into two parts: 1. We make our first attempts at formalising the notion of.
1 Lecture 3 The Languages of K, T, B and S4. 2 Last time we extended the language PC to the language S5 by adding two new symbols ‘□’ (for ‘It is necessary.
PREDICATES AND QUANTIFIERS COSC-1321 Discrete Structures 1.
PHILOSOPHY OF LANGUAGE Some topics and historical issues of the 20 th century.
Meaning and Language Part 1. Plan We will talk about two different types of meaning, corresponding to two different types of objects: –Lexical Semantics:
Chapter 2 1. Chapter Summary Sets (This Slide) The Language of Sets - Sec 2.1 – Lecture 8 Set Operations and Set Identities - Sec 2.2 – Lecture 9 Functions.
Lec. 10.  In this section we explain which constituents of a sentence are minimally required, and why. We first provide an informal discussion and then.
English Syntax Week 12. NP movement Text 9.2 & 9.3.
Part I: Basics and Constituency
Logics for Data and Knowledge Representation
Logics for Data and Knowledge Representation
Validity and Soundness, Again
Presentation transcript:

LIN3021 Formal Semantics Lecture 2 Albert Gatt

Aims Our aim today will be to: Revisit the principle of Compositionality and try to put it on a concrete footing. Explore the relationship between syntax and semantics in constructing the meaning of a phrase. Revisit some formal machinery to help us achieve this. The main linguistic phenomenon we’ll focus on today is predication

From worlds to models (a reminder of some things covered in LIN1032)

A bit of notation In what follows, we’ll be using: Italics for linguistic expressions Double brackets to denote the meaning of these expressions:

Propositions as functions What we’ve said so far about propositions could be formalised as follows: A proposition is a function from possible worlds to truth values. In other words, understanding a proposition means being able to check, for any conceivable possible world, whether the proposition is true in that world or not.

Propositions as functions (II) This situation could be modelled like this: In other words, we can view a proposition as a function from possible worlds to truth values. T Proposition F

Model theory The kind of semantics we’ll be doing is often called model- theoretic. That’s because we assume that interpretation is carried out relative to a model of the world. A model is: A structured domain of the relevant entities which allows us to interpret all the expressions of our meta-language. For now, think of it as a “small partial model of the world” A natural language expression will be translated into the (logical) meta-language and interpreted according to the model.

Models and formal systems Models are used to make precise the semantics of many formal systems (as well as Natural Languages). Given a formal theory, a model specifies the minimal structure (abstract or concrete) needed in order to interpret all the primitive statements in the theory. (In formal systems, we also require that the model is such that all the true statements of the theory come out true in the model.)

Models and natural language Components of a model: a universe of individuals U an interpretation function I which assigns semantic values to our constants. the truth values {T,F} as usual (for propositions) Formally: M = <U,I> “Model M is made up of U and I”

Constructing a world Suppose our example world contains exactly 4 individuals. U = {Isabel Osmond, Emma Bovary, Alexander Portnoy, Beowulf}

Assigning referents to constants To each individual constant, there corresponds some individual in the world, as determined by the interpretation function I: [[a]]M = Isabel Osmond [[b]]M = Emma Bovary [[c]]M = Alexander Portnoy [[d]]M = Beowulf

Predicates 3-place predicates are sets of triples (3-place relations): We also have a fixed set of predicates in our meta-language. These correspond to natural language expressions. Their interpretation (extension) needs to be fixed for our model too: 1-place predicates are sets of individuals [[P]]M  U 2-place predicates are sets of ordered pairs (relations) [[Q]]M  U x U 3-place predicates are sets of triples (3-place relations): [[R]]M  U x U x U … and so on

Assigning extensions to predicates Suppose we have 2 1-place predicates in our language: tall and clever Let us fix their extension like this: [[tall]]M = {Emma Bovary, Beowulf} [[clever]]M = {Alexander Portnoy, Emma Bovary, Isabel Osmond} Thus, in our world, we know the truth of the following propositions: [[tall(a)]]M = FALSE (since Isabel isn’t tall) [[clever(c)]]M = TRUE (since Alexander is clever)

More formally… A sentence of the form P(t), where P is a predicate and t is an individual term (constant or variable) is true in some model M = <U,I> iff: the object assigned to t by I is in the extension of P (i.e. the object that t points to is in the set of things of which P is true) i.e. t  [[P]]M How would you extend this to a sentence of the form P(t1, t2), using a 2-place predicate?

For two-place predicates A sentence of the form P(t1, t2), where P is a predicate and t1, t2 are individual terms (constants or variables) is true in some model M iff: the ordered pair of the objects assigned to t1 and t2 can be found among the set of ordered pairs assigned to P in that interpretation

Propositions We evaluate propositions against our model, and determine whether they’re true or false. If β is a proposition, then [[β]]M is either the value TRUE or the value FALSE, i.e.: [[β]]M  {TRUE, FALSE}

Complex formulas Once we have constructed our interpretation (assigning extensions to predicates and values to constants), complex formulas involving connectives can easily be interpreted. We just compute the truth of a proposition based on the connectives and the truth of the components. Recall that connectives have truth tables associated with them. For propositions that contain connectives, the truth tables describe the function from worlds to truth values.

Truth tables and models For a complex proposition (e.g. pq), our truth table tells us in which worlds this proposition would be true or false, namely: pq is true in any world where p is false and q is false, or p is true and q is true, etc

Complex formulas: example Model: [[a]]M = I. Osmond [[b]]M = E. Bovary [[c]]M = A. Portnoy [[d]]M = Beowulf [[tall]]M = {E. Bovary, Beowulf} [[clever]]M = {A. Portnoy, E. Bovary, I. Osmond} Formulas: [[tall(a) Λ clever(b)]]M “Isabel is tall and Emma is clever” FALSE [[clever(a) ν tall(a)]]M Isabel is tall or clever TRUE [[¬tall(a)]]M Isabel is not tall

Compositionality

Frege’s principle The meaning of an expression is a function of the meaning of its component parts and the manner in which they are combined. You’ve seen this before. Why is it important?

Productivity and combinatoriality We could identify the meaning of words with some kind of dictionary. This is much more difficult to do with sentences. Sentence (and phrase) formation is a productive process, so that the range of possible sentences (resp. phrases) in a language is potentially infinite. Clearly, in order to comprehend sentences we’ve never heard before, we need some kind of equally productive interpretative process.

Compositional semantics Our main task is not going to be to figure out the meaning of a particular word, or phrase. The main task is to figure out how a word of phrase contributes to the meaning of a larger word or phrase.   We are effectively following the advice of philosopher David Lewis (1970): In order to say what a meaning is, we may first find what a meaning does, and then find something that does that.

Compositional semantics and truth We’ve said quite a bit about truth conditions already. Here’s another piece of advice, this time from Max Cresswell: For two sentences α and β, if in some situation α is true and β is false, then α and β must have different meanings. If we combine this with Lewis’s advice, we arrive at the conclusion that if two distinct expressions give rise to different truth conditions, all other things being equal, then they must have different meanings. Keep this in mind – we’ll come back to it later.

An initial case study Predication and reference

Predication Consider the following sentence: Paul sleeps Note that it consists of a subject NP (Paul) and a predicate (sleeps). There are two things we need to consider here: The meaning of names like Paul The status of predicates like sleeps We’re going to think of predicates as ‘incomplete propositions’, that is, propositions that have some missing information. Once this information is supplied, they give us a complete proposition.

Names and reference Let’s think of names as referring to individuals. Then: [[Paul]] = some individual called Paul (in the world/model under consideration) under some theories, a name means something like a description (e.g. [[Paul]] = the guy I met yesterday) we’ll just assume that there is a direct link between the name and the individual. [[Paul]] =

Predicates and saturation What contribution does sleeps make to the sentence Paul sleeps? Let’s think of the proposition [[Paul sleeps]] as follows: if the model contains a situation where Paul is indeed asleep, we have a true sentence; else we have a false sentence We can represent the proposition graphically as follows:   Intuitively, this picture is a composite of the meanings [[Paul]] and [[sleeps]].

Predicates and saturation To represent the contribution of the predicate, we can remove the contribution of Paul, leaving whatever it is that sleeps means. The ‘hole’ is meant to represent the missing reference. Essentially, this says: if you plug something into sleeps, and it’s the right type of object, you have a proposition. So predicates are unsaturated propositions. Equivalently, in predicate logic, we call them open propositions or propositional functions.

Predicates and saturation When a property (sleeps) is predicated of an individual (Paul), we are saturating the predicate. Therefore, predication is saturation. Formally, we typically represent predicates using the language of predicate logic: sleep(x ) The variable here corresponds to the unsaturated ‘hole’ in our graphic example. Combining the predicate with the individual that is referred to by Paul (that is, [[Paul]]), we get: sleep(p) (using p to stand for the individual [[Paul]] Here’s a something to think about: Our graphic represents the unsaturated predicate a bit like a function: given the input of the right type, it becomes a proposition. But a proposition is a function which returns a truth value (in a specific world/situation) So predicates can be thought of as functions from entities to propositions.

The syntax-semantics correspondence We’ve seen how predicates can combine with referring expressions like names, to form complete propositions.  In order to give a completely compositional account, we need to bring syntax into the picture S NP N Paul VP V sleeps

The syntax-semantics correspondence There are at least two ways in which we can operationalise the principle of compositionality: rule by rule theories: Assume that for each syntactic rule, there must be a corresponding semantic rule. Syntax and semantics work hand in hand. interpretive theories: semantics applies to the output of syntax. Syntax and semantics need not work completely in lockstep. The main difference between the two is that rule by rule theories require that every syntactic building block has a corresponding semantic operation. In interpretive theories, we don’t require this.

An illustration Rule by rule Interpretive (From the lexicon) [[Paul]]=p (From the lexicon) [[sleep]]=sleep(x ) Assign Paul category N; [[N]] = p Assign sleeps category V; [[V]]=sleep(x ) [[NP]] = [[N]] (no other daughters) [[VP]] = [[V]] (no other daughters) [[S]] = [[NP]] + [[VP]] (From the lexicon) [[Paul]]=p (From the lexicon) [[sleep]]=sleep(x ) [[N]] = p [[V]] = sleep(x ) [[NP]] = [[N]] (no other daughters) [[VP]] = [[V]] (no other daughters) [[S]] = [[NP]] + [[VP]]

Implications of the distinction Consider the question: Who did Paul kill? Suppose we adopt a generative framework for our syntactic analysis. This would assume that the question had the following underlying form: whoi did Paul kill ti which essentially views who as a placeholder for something which was originally in the object position. A rule-by-rule theory would need to state a rule corresponding to the movement of who. An interpretive theory wouldn’t: it would just need to identify the link between who and the verb object position.

The two approaches in context  The rule-to-rule framework was explicitly proposed by Montague, who went on to formalise several fragments of English. Montague attempted to state precise semantic rules for every syntactic rule.   Interpretive frameworks are more typical of theories influenced by Chomsky’s Government and Binding theory (and later frameworks). Under this view, semantic interpretation takes place after syntax, at the level of LF (‘logical form’).

Simplifying the semantic rules So far, we’ve stated semantic rules directly in terms of the syntax (using categories like N and VP). We’d like to simplify our rules as much as possible. Here are some generalisations: If a node x has a single daughter y, then [[x]]=[[y]]; If a node x has two daughters y and z, and [[y]] is an individual and [[z]] is a property, then saturate the meaning of z with the meaning of y and assign the resulting meaning to x; Note: The second rule above implicitly refers to the semantic type of the meanings we’re talking about (are they individuals? are they properties?). This does away with the need to specify syntactic categories (N, V etc). More on this later!

Some implications

Why bother with syntax and semantics? Recall the two pieces of advice we started out with: In order to say what a meaning is, we may first find what a meaning does, and then find something that does that. For two sentences α and β, if in some situation α is true and β is false, then α and β must have different meanings. Now, consider the following example fragments (from ): almost half full almost half empty Do they mean the same thing?

Half full and half empty We might consider half full and half empty to be synonymous (recall the possible-worlds definition of synonymy). But combined with almost, things look slightly different: the sentences don’t have the same truth conditions. So by Cresswell’s principle, the sentences have different meanings. Suppose we analyse these as: almost + DEGREE ADJ  Since almost is common to both, then by Lewis’s advice, the difference in meaning must reside in half full and half empty.

Half full and half empty Now, consider two possible syntactic structures for these expressions: If we adopt the first structure, then we’ll have to conclude that half full and half empty mean different things (perhaps as a result of combining full and empty with the degree adjective). If we adopt the second structure, then all we need is for full and empty to have different meanings (the rest is constant). AdjP AdjP almost AdjP DegP ADJ DEGREE ADJ almost DEGREE

Other kinds of predicates

Other predicates So far, we’ve considered one kind of predicate, namely, intransitive verbs (e.g. sleep). Let’s consider a few others: adjectives: Paul is human predicate nominals: Paul is a man relative clauses: Paul is who I’m telling you about transitive verbs: Mary killed Paul

Adjectives and predicate nominals Paul is tall Paul is a man   Questions: What does the word is mean here? What about the word a?

Adjectives and predicate nomimals Compare: Pawlu twil Pawlu raġel In many languages (such as Maltese), the copula (be) is omitted. Does this suggest that is is meaningless (has a purely syntactic function) or is deleted at the semantic level?

The meaning of be We can think of is as a kind of ‘super-property’, a property that is saturated by another property. Under this view, we can think of the predicate adjective or nominal on a par with the intransitive verb sleep, except that (in English) they need to be slotted into another property: [[tall]] = tall(x) [[man]] = man(x)

The meaning of a Observe that a has roughly (but not completely) the same distribution as other determiners: the/a/every/some dog bit me But not always: John is a (*the/some/every) man For now, let’s think of a as having more or less no function in these examples.

Transitive verbs The difference between instransitive verbs like sleep and transitive ones like kill is that the latter need more than one argument: sleep(x) kill(x,y)

Relative clauses In a sentence such as John is who Mary killed the phrase who Mary killed is also a predicate (i.e. an unsaturated proposition, that yields a complete proposition when combining with John) We need to deal with the internal structure of this complex predicate.

Relative clauses: a rough estimation [[who Mary killed]] = the proposition that Mary killed someone underlined part represents whatever’s missing and needs to saturate the proposition. We start with Mary killed e: the placeholder ei temporarily saturates kill We add who who indicates that the logical object of kill is no longer in place this can be interpreted as an instruction to (re-)unsaturate kill, before combining it with John who Mary killed ei Again, we’re assuming that who has moved to the front of the sentence. The underlying proposition is Mary killed e.

An interim summary Names refer. In model-theoretic terms, names denote individuals in the world. Predicates (adjectives, verbs, nominals) are properties. They are saturated by terms denoting individuals to yield propositions. A compositional analysis looks at the role of each in the syntactic context in which it occurs. Different manners of combination can yield different interpretations.