Spoken Dialogue Systems: System Overview

Slides:



Advertisements
Similar presentations
Conversational Implicature (Based on Paltridge, chapter 3)
Advertisements

Conversations  Conversation are cooperative events:  Without cooperation, interaction would be chaotic. Would be no reason to communicate  Grice's.
Topic 10: conversational implicature Introduction to Semantics.
The Cooperative Principle
1 MODULE 2 Meaning and discourse in English COOPERATION, POLITENESS AND FACE Lecture 14.
CS 224S / LINGUIST 285 Spoken Language Processing Dan Jurafsky Stanford University Spring 2014 Lecture 9: Human conversation, frame-based dialogue systems.
People & Speech Interfaces CS 260 Wednesday, October 4, 2006.
Speech and Language Processing
ITCS 6010 Spoken Language Systems: Architecture. Elements of a Spoken Language System Endpointing Feature extraction Recognition Natural language understanding.
6/28/20151 Spoken Dialogue Systems: Human and Machine Julia Hirschberg CS 4706.
Pragmatics.
Semantics 3rd class Chapter 5.
Game Theory and Grice’ Theory of Implicatures Anton Benz.
Theories of Discourse and Dialogue. Discourse Any set of connected sentences This set of sentences gives context to the discourse Some language phenomena.
1 Spoken Dialogue Systems Dialogue and Conversational Agents (Part I) Chapter 19: Draft of May 18, 2005 Speech and Language Processing: An Introduction.
Department of English Introduction To Linguistics Level Four Dr. Mohamed Younis.
Research Methods in T&I Studies I Cooperative Principle and Culture-Specific Maxims.
Pragmatics.
1 Natural Language Processing Lecture Notes 14 Chapter 19.
Dr. Katie Welch LING  Heretofore, we have talked about the form of language  But, this is only half the story.  We must also consider the.
MLS 570 Critical Thinking Reading Notes Fogelin: Ch. 1 Fall Term 2006 North Central College Dr. Sally Fowler.
Presentation about pragmatic concepts Implicatures Presuppositions
Welcome Back, Folks! We’re travelling to a littele bit far-end of Language in Use Studies EAA remains your faithful companion.
UNIT 2 - IMPLICATURE.
Optimal answers and their implicatures A game-theoretic approach Anton Benz April 18 th, 2006.
Cooperation and Implicature (Conversational Implicature) When people talk with each other, they try to converse smoothly and successfully. Cooperation.
Implicature. I. Definition The term “Implicature” accounts for what a speaker can imply, suggest or mean, as distinct from what the speaker literally.
Aristotel‘s concept to language studies was to study true or false sentences - propositions; Thomas Reid described utterances of promising, warning, forgiving.
SEMANTICS ??? aardvark SEMANTICS ??? aardvark. SEMANTICS: word and sentence meaning. PRAGMATICS: speaker meaning. The semiotic triangle:
Natural Language Understanding
Utilization of Sentences Lec. 3
COMMUNICATION OF MEANING
CS 224S / LINGUIST 285 Spoken Language Processing
MODULE 2 Meaning and discourse in English
Telephone etiquette Ref: Pinner D & Pinner D 2003 Communication Skills Pearson Longman, New Zealand Ch. 15: pp
COOPERATIVE PRINCIPLE:
Information and Advice
COOPERATION and IMPLICATURE
GRICE’S CONVERSATIONAL MAXIMS
How can we become good learners?
Competency – 3 Engages in active listening an response appropriately
Grice’s Maxims LO: to understand the co-operative principle and how we can use it within our own analysis.
Discourse and Pragmatics
Spoken Dialogue Systems
Issues in Spoken Dialogue Systems
Spoken Dialogue Systems: Human and Machine
Spoken Dialogue Systems: Managing Interaction
Lecture 4: October 7, 2004 Dan Jurafsky
Automatic Speech Recognition
Introduction to Linguistics
Lecture 3: October 5, 2004 Dan Jurafsky
Symbolic Systems 100 Spring 2005
Lecture 29 Word Sense Disambiguation-Revision Dialogues
Managing Dialogue Julia Hirschberg CS /28/2018.
Why conversation works.
The Cooperative Principle
COOPERATIVE PRINCIPLE.
Nofsinger. R., Everyday Conversation, Sage, 1991
Problem Solving Skill Area 305.1
Lecture 30 Dialogues November 3, /16/2019.
Toward a new taxonomy for pragmatic inference: Q-based & R-based implicature Laurence R. Horn (1984)
How to Avoid Redundant Object-References
The study of meaning in context
The Cooperative Principle
Pragmatics Predmetni nastavnik: doc. dr Valentna Boskovic Markovic
Gricean Cooperative Principle (Maxim) and Implicature
Introduction to pragmatics
Nofsinger. R., Everyday Conversation, Sage, 1991
Spoken Language Processing
Presentation transcript:

Spoken Dialogue Systems: System Overview Julia Hirschberg CS 4706 2/18/2019

Today Conversational Agents Cooperative Question-Answering ASR NLU Generation Cooperative Question-Answering Dialogue Representations 2/18/2019

Conversational Agents AKA: Interactive Voice Response Systems Dialogue Systems Spoken Dialogue Systems Applications: Travel arrangements (Amtrak, United airlines) Telephone call routing Tutoring Communicating with robots Anything with limited screen/keyboard 2/18/2019

A travel dialog: Communicator 2/18/2019

Call routing: ATT HMIHY 2/18/2019

A tutorial dialogue: ITSPOKE 2/18/2019

Conversational Structure Telephone conversations Stage 1: Enter a conversation Stage 2: Identification Stage 3: Establish joint willingness to converse Stage 4: First topic is raised, usually by caller 2/18/2019

Why is this customer confused? Customer: (rings) Operator: Directory Enquiries, for which town please? Customer: Could you give me the phone number of um: Mrs. um: Smithson? Operator: Yes, which town is this at please? Customer: Huddleston. Operator: Yes. And the name again? Customer: Mrs. Smithson 2/18/2019

Why is this customer confused? A: And, what day in May did you want to travel? C: OK, uh, I need to be there for a meeting that’s from the 12th to the 15th. Note that client did not answer question. Meaning of client’s sentence: Meeting Start-of-meeting: 12th End-of-meeting: 15th Doesn’t say anything about flying!!!!! How does agent infer client is informing him/her of travel dates? 2/18/2019

Will this client be confused? A: … there’s 3 non-stops today. True if in fact 7 non-stops today. But agent means: 3 and only 3. How can client infer that agent means: only 3 2/18/2019

Grice: Conversational Implicature Implicature: a particular class of licensed inferences Grice (1975): how do conversational participants draw inferences beyond what is actually asserted? Cooperative Principle "Make your contribution such as it is required, at the stage at which it occurs, by the accepted purpose or direction of the talk exchange in which you are engaged." A tacit agreement by speakers and listeners to cooperate in communication 2/18/2019

4 Gricean Maxims Relevance: Be relevant Quantity: Do not make your contribution more or less informative than required Quality: try to make your contribution one that is true (don’t say things that are false or for which you lack adequate evidence) Manner: Avoid ambiguity and obscurity; be brief and orderly 2/18/2019

Relevance A: Is Regina here? B: Her car is outside. Implication: yes Hearer thinks: why would he mention the car? It must be relevant. How could it be relevant? It could since if her car is here she is probably here. Client: I need to be there for a meeting that’s from the 12th to the 15th Hearer thinks: Speaker is following maxims, would only have mentioned meeting if it was relevant. How could meeting be relevant? If client meant me to understand that he had to depart in time for the mtg. 2/18/2019

Quantity A:How much money do you have on you? B: I have 5 dollars Implication: not 6 dollars Similarly, 3 non stops can’t mean 7 non-stops (hearer thinks: if speaker meant 7 non-stops she would have said 7 non-stops A: Did you do the reading for today’s class? B: I intended to Implication: No B’s answer would be true if B intended to do the reading AND did the reading, but would then violate maxim 2/18/2019

Dialogue System Architecture 2/18/2019

Speech recognition Input: acoustic waveform Output: string of words Basic components: a recognizer for phones, small sound units like [k] or [ae]. a pronunciation dictionary like cat = [k ae t] a grammar telling us what words are likely to follow what words A search algorithm to find the best string of words 2/18/2019

Natural Language Understanding Or “NLU” Or “Computational semantics” There are many ways to represent the meaning of sentences For speech dialogue systems, most common is “Frame and slot semantics”. 2/18/2019

An example of a frame Show me morning flights from Boston to SF on Tuesday. SHOW: FLIGHTS: ORIGIN: CITY: Boston DATE: Tuesday TIME: morning DEST: CITY: San Francisco 2/18/2019

How to generate this semantics? Many methods Simplest: “semantic grammars” E.g. a simple Context Free Grammar LHS of rules is a semantic category: LIST -> show me | I want | can I see|… DEPARTTIME -> (after|around|before) HOUR | morning | afternoon | evening HOUR -> one|two|three…|twelve (am|pm) FLIGHTS -> (a) flight|flights ORIGIN -> from CITY DESTINATION -> to CITY CITY -> Boston | San Francisco | Denver | Washington 2/18/2019

Semantics for a sentence LIST FLIGHTS ORIGIN Show me flights from Boston DESTINATION DEPARTDATE to San Francisco on Tuesday DEPARTTIME morning 2/18/2019

Generation and TTS Generation component TTS component Chooses concepts to express to user Plans how to express these concepts in words Assigns appropriate prosody to the words TTS component Takes words and prosodic annotations Synthesizes a waveform 2/18/2019

Generation Component Content Planner Language Generation Decides what content to express to user (ask a question, present an answer, etc) Often merged with dialogue manager Language Generation Chooses syntactic structures and words to express meaning. Simplest method Pre-specify basic sentences and fill in the blanks as appropriate “Template-based generation” Can have variables: What time do you want to leave CITY-ORIG? Will you return to CITY-ORIG from CITY-DEST? 2/18/2019

More sophisticated language generation component Natural Language Generation Approach: Dialogue manager builds representation of meaning of utterance to be expressed Passes this to a “generator” Generators have three components Sentence planner Surface realizer Prosody assigner 2/18/2019

Architecture of a generator for a dialogue system (after Walker and Rambow 2002) 2/18/2019

HCI constraints on generation for dialogue: Coherence Discourse markers and pronouns (Coherence”): Please say the date.… Please say the start time.… Please say the duration… Please say the subject… First, tell me the date.… Next, I’ll need the time it starts.… Thanks. <pause> Now, how long is it supposed to last?… Last of all, I just need a brief description Bad! Good! 2/18/2019

HCI Constraints on Dialogue Generation Coherence: Tapered Prompts Prompts which get incrementally shorter: System: Now, what’s the first company to add to your watch list? Caller: Cisco System: What’s the next company name? (Or, you can say, “Finished”) Caller: IBM System: Tell me the next company name, or say, “Finished.” Caller: Intel System: Next one? Caller: America Online. System: Next? Caller: … 2/18/2019

Next Class Dialogue Management 2/18/2019