The NESPOLE Interchange Format (IF) Lori Levin, Emanuele Pianta, Donna Gates, Kay Peterson, Dorcas Wallace, Herve Blanchon, Roldano Cattoni, Jean-Philippe.

Slides:



Advertisements
Similar presentations
Follow the Samples Act Out Put in Use Data Bank Talking Face to Face
Advertisements

TEACHING GRAMMAR Bui Thi Thao Truong Thuy Duong.
Present Perfect Dragana Filipovic.
Grammar Review 4. Reported Speech.
Complex Infinitives to wash: (simple) infinitive to be washing:
Non-Native Users in the Let s Go!! Spoken Dialogue System: Dealing with Linguistic Mismatch Antoine Raux & Maxine Eskenazi Language Technologies Institute.
Home This training presentation is designed to introduce the Residency Management Suite to new users. This presentation covers the following topics: Login.
Materials and Lesson Planning
VERB TENSE: PRESENT PERFECT SIMPLE
Present Perfect Present X X ?.
1 MODULE 2 Meaning and discourse in English COOPERATION, POLITENESS AND FACE Lecture 14.
EL111 Unit Eleven Ms. Khadeeja Rabah .
“What do you want me to do now?”
9/10/2004Use Case Workshop 1 CSC480 Software Engineering Workshop 1 Requirements Modeling.
1 GetThere User Training Booking & Managing Online Travel.
Substitute FAQs SubFinder Overview. FAQs Do I have to have touch-tone service to use SubFinder? No, but you do need a telephone that can be switched from.
Part II Questions & Responses. Information Questions (1)Who is your sister in this picture? (A) She is generous. (B) There are three girls in this picture.
Present Perfect and Simple Past (II)
1 of 5 This document is for informational purposes only. MICROSOFT MAKES NO WARRANTIES, EXPRESS OR IMPLIED, IN THIS DOCUMENT. © 2006 Microsoft Corporation.
Advanced Technical Writing Lecture 8 Memorandums 29 June 2008.
Glennysha Jurado-Moran
Machine Translation Challenges and Language Divergences Alon Lavie Language Technologies Institute Carnegie Mellon University : Machine Translation.
MOOD Indicative, Imperative, Subjunctive
Lesson 4 Making Telephone Calls Business English Conversation & Listening Instructor: Hsin-Hsin Cindy Lee, PhD.
Verb Tense Review.
Reported Speech teach u how 2 do it!!. How do you recognise Direct Speech? You have “…..” Or the name of the person speaking is given At the end, or at.
Reported Speech Roll No Presented By:- Class: Ixth “A”
Prehľad časov v anglickom jazyku gramatická príručka pre študentov religionistiky FF PU Vypracoval: Tomáš Buchala.
1 English for the Hotel Business Unit 7: Customer Care LA2401.
Glennysha Jurado-Moran
A First Program Using C#
REPORTED SPEECH Unit 11 – English 12 Instructor: Nguyễn Ngọc Vũ
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Spoken dialog for e-learning supported by domain ontologies Dario Bianchi, Monica Mordonini and Agostino Poggi Dipartimento di Ingegneria dell’Informazione.
SIG IL 2000 Evaluation of a Practical Interlingua for Task-Oriented Dialogue Lori Levin, Donna Gates, Alon Lavie, Fabio Pianesi, Dorcas Wallace, Taro Watanabe,
THE PASSIVE 2nd of Bachillerato. USES THE PASSIVE VOICE IS VERY COMMON IN ENGLISH AND IT IS USED TO: -TO FOCUS THE ATTENTION ON THE ACTION RATHER THAN.
Practice Examples 1-4. Def: Semantics is the study of Meaning in Language  Definite conclusions Can be arrived at concerning meaning.  Careful thinking.
Speech-to-Speech MT in the JANUS System Lori Levin and Alon Lavie Language Technologies Institute Carnegie Mellon University.
Business Communication 1. Guidelines 2 1. Use common courtesy in your request – ask rather than demand. 2. Include all the information the recipients.
Advanced Technical Writing Lecture 4 Memorandums.
English for the Hotel Business
Department of Psychology Experiment Management System Experimenter Tutorial Stony Brook University Subject Pool Office
Unit 9 Can you come to my party? Language Goals: Make invitations Decline invitations accept invitations Talk about obligations.
Ideas for 100K Word Data Set for Human and Machine Learning Lori Levin Alon Lavie Jaime Carbonell Language Technologies Institute Carnegie Mellon University.
MT with an Interlingua Lori Levin April 13, 2009.
Designing a Machine Translation Project Lori Levin and Alon Lavie Language Technologies Institute Carnegie Mellon University CATANAL Planning Meeting Barrow,
Into English 3 Unit 1 12th Grade Miss. Peres. N C M O A I T I M N U C O  Make up as many words as you can from these letters.  Count  Mountain  Can.
Rapid Development in new languages Limited training data (6hrs) provided by NECTEC from 34 speakers, + 8 spks for development and test Romanization of.
Week 4. Working with the person sitting next to you, answer the following questions. Remember to use full sentences! » What might you buy if you had more.
Unit 7 Grammar Forms & Functions 3
Aviation English Week Two Unit One Planning your trip.
Recent Advances in Speech Translation Systems ESSLLI-2002 Tutorial Course August 12-16, 2002 Course Organizers: Alon Lavie – Carnegie Mellon University.
Thursday, September 16, Announcements Movie night, Friday September 25, at 6:00 pm. Bring your family and friends. There will be vocabulary review.
Dialogue Modeling 2. Indirect Requests “Can I have a cup of coffee?”  One approach to dealing with these kinds of requests is by plan-based inference.
Parsing into the Interlingua Using Phrase-Level Grammars and Trainable Classifiers Alon Lavie, Chad Langley, Lori Levin, Dorcas Wallace,Donna Gates and.
Word and the Writing Process. To create a document 1.On the Start menu, point to Programs, and then click Microsoft Word. A new document opens in Normal.
FUTURE PERFECT TENSE Grammar. Future simple Simple Future has two different forms in English: "will" and "be going to." Although the two forms can sometimes.
Carnegie Mellon IRST-itc Balancing Expressiveness and Simplicity in an Interlingua for Task Based Dialogue Lori Levin, Donna Gates, Dorcas Wallace, Kay.
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
Tuesday, April 6, 2016 Level 3 Week 7. Announcements CASAS #2 tonight Units 8-12 test tomorrow Review for the midterm on Monday Midterm on Tuesday.
Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers Chad Langley, Alon Lavie, Lori Levin, Dorcas Wallace, Donna Gates and Kay.
This I Believe Essay Writer’s Workshop: Introductions, Juicy Details, & Conclusions 8 th ELA St. Joseph School.
Hotel English 《饭店英语》制作人:舒飞霞 任务 2 :酒店预订之网络预订、前台预订 知识技能模块一:前厅服务 Unit2 Making Reservations by Internet & Talk.
LingWear Language Technology for the Information Warrior Alex Waibel, Lori Levin Alon Lavie, Robert Frederking Carnegie Mellon University.
PET WRITING & LISTENING.
World Travel, Inc. Concur Online User Guide
How to Be Successful in English 3000
GRAMMAR TASK INFORMATION
CS305, HW1, Spring 2008 Evaluation Assignment
Presentation transcript:

The NESPOLE Interchange Format (IF) Lori Levin, Emanuele Pianta, Donna Gates, Kay Peterson, Dorcas Wallace, Herve Blanchon, Roldano Cattoni, Jean-Philippe Gibaud, Chad Langley, Alon Lavie, Nadia Mana, Fabio Pianesi

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Coverage: how do we measure coverage of the domain –Reliability: so that an analyzer written by one person in Italy can work with a generator written by someone he has never met in Korea. –Scalability: move to broader semantic domains without a constant increase in the amount of work.

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Coverage –Reliability –Scalability

What is an interlingua? Representation of meaning or speaker intention. Sentences that are equivalent for the translation task have the same interlingua representation. The room costs 100 Euros per night. The room is 100 Euros per night. The price of the room is 100 Euros per night.

Mi chiamo Alex WaibelMy name is Alex Waibel. Give-information+personal-data (name=alex_waibel) [ s [ vp accusative_pronoun “chiamare” proper_name]] [ s [ np [possessive_pronoun “name”] ] [ vp “be” proper_name]] Direct Transfer Interlingua Vaquois MT Triangle

Other Approaches to Machine Translation Direct: –Very little analysis of the source language. Transfer: –Analysis of the source language. –The structure of the source language input may not be the same as the structure of the target language sentence. –Transfer rules relate source language structures to target language structures.

Note Some transfer systems may produce a more detailed meaning representation than some interlingua systems. The difference is whether translation equivalents in the source and target languages are related by a single canonical representation.

Multilingual Translation with an Interlingua Japanese Arabic Chinese (input sentence) San1 tian1 qian2, wo3 kai1 shi3 jue2 de2 tong4 English French German Italian Korean Arabic Chinese (paraphrase) wo3 yi3 jin1 tong4 le4 san1 tian1 English (output sentence) The pain started three days ago. French German Italian Japanese Korean Analyzers Generators Spanish Catalan Interlingua give-information+onset+body-state (body-state-spec=pain, time=(interval=3d, relative=before))

Multilingual translation with transfer Transfer-rules-1: Arabic-Catalan Transfer-rules-2: Catalan-Arabic Transfer-rules-3: Arabic-Chinese Transfer-rules-4: Chinese-Arabic Transfer-rules-5: Arabic-English Transfer-rules-6: English-Arabic Etc.

Advantages of Interlingua Add a new language easily –get all-ways translation to all previous languages by adding one grammar for analysis and one grammar for generation Mono-lingual development teams. Paraphrase –Generate a new source language sentence from the interlingua so that the user can confirm the meaning

Disadvantages of Interlingua “Meaning” is arbitrarily deep. –What level of detail do you stop at? If it is too simple, meaning will be lost in translation. If it is too complex, analysis and generation will be too difficult. Should be applicable to all languages. Human development time.

Interlingual MT Systems University of Maryland – Lexical Conceptual Structure (Dorr) Carnegie Mellon –Kantoo (Mitamura and Nyberg) –Nespole/C-STAR (Waibel, Levin, Lavie) UNL (Universal Networking Language) Microcosmos (Nirenburg) Verbmobil – Domain actions (Block)

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Reliability –Coverage

A Travel Dialogue Translated from Italian A: Albergo Gabbia D’Oro. Good evening. B: My name is Anna Maria DeGasperi. I’m calling from Rome. I wish to book two single rooms. A: Yes. B: From Monday to Friday the 18th, I’m sorry, to Monday the 21st. A: Friday the 18th of June. B: The 18th of July. I’m sorry. A: Friday the 18th of July to, you were saying, Sunday. B: No. Through Monday the 21st.

A Travel Dialogue (Continued) B: So with departure on Tuesday the 22nd. A: Then leaving on the 22nd. Yes. We have two singles certainly. B: Yes. A: Would you like breakfast? B: Is it possible to have all meals? A: No. We serve meals only in the evening. B: Ok. If you can do breakfast and dinner. A: Ok. B: Do you need a deposit?

A Travel Dialogue (Continued) A: You can give me your credit card number. B: Ok. Just a moment. Ok. My name is Anna Maria DeGaperi. The card is A: Good. B: Expiration A: Good. Thank you. We need a confirmation on the 18th of July before 6pm. B: Goodbye. A: Thanks. Goodbye. B: Thanks. Goodbye.

A Non-Task-Oriented Dialogue (We can’t translate this.) A: Are you cooking? B: My father is cooking. I’m cleaning. I just finished cleaning the bathroom. A: Look. What do you know about Monica? B: I don’t know anything. Look. I don’t know anything. A: You don’t know anything? I wrote her three weeks ago, but if she hasn’t received the letter, they would have returned it. I hope she received it. B: Because Celia told me that the address that Monica had given us was wrong. She said that if I was going to write to her, well, …. From the Spanish CallHome corpus: unlimited conversation between family members.

The Ideal MT System… Fully automatic High quality Domain independent (any topic) …. isn’t within the current state-of-the-art.

Design Principles of the Interchange Format Suitable for task oriented dialogue Based on speaker’s intent, not literal meaning –Can you pass the salt is represented only as a request for the hearer to perform an action, not as a question about the hearer’s ability. Abstract away from the peculiarities of any particular language –resolve translation mismatches. Instructions: Delete sample document icon and replace with working document icons as follows: Create document in Word. Return to PowerPoint. From Insert Menu, select Object… Click “Create from File” Locate File name in “File” box Make sure “Display as Icon” is checked. Click OK Select icon From Slide Show Menu, Select Action Settings. Click “Object Action” and select “Edit” Click OK

Translation Mismatches Sentences that are translation-equivalents in two languages do not have the same syntactic structure or predicate-argument structure. (Unitrans; Eurotra) –I like to swim. –I swam across the river. –Sue met with Sam/Sue met Sam.

Design Principles (continued) Domain independent framework with domain-specific parts Simple and reliable enough to use: –at multiple research sites with high intercoder agreement. –with widely varying type of parsers and generators. Allow robust language engines –Underspeicification must be possible. –Fragments must be represented.

Speech Acts: Speaker intention vs literal meaning Can you pass the salt? Literal meaning: The speaker asks for information about the hearer’s ability. Speaker intention: The speaker requests the hearer to perform an action.

Remember this term: Domain Action

Domain Actions: Extended, Domain-Specific Speech Acts give-information+existence+body-state It hurts. give-information+onset+body-object The rash started three days ago. request-information+availability+room Are there any rooms available? request-information+personal-data What is your name?

Domain Actions: Extended, Domain-Specific Speech Acts In domain. –I sprained my ankle yesterday. –When did the headache start? Out of domain –Yesterday I slipped in the driveway on my way to the garage. –The headache started after my boss noticed that I deleted the file.

Formulaic Utterances Good night. tisbaH cala xEr waking up on good Romanization of Arabic from CallHome Egypt

Same intention, different syntax rigly bitiwgacny my leg hurts candy wagac fE rigly I have pain in my leg rigly bitiClimny my leg hurts fE wagac fE rigly there is pain in my leg rigly bitinqaH calya my leg bothers on me Romanization of Arabic from CallHome Egypt.

Language Neutrality Comes from representing speaker intention rather than literal meaning for formulaic and task-oriented sentences. How about … suggestion Why don’t you… suggestion Could you tell me… request info. I was wondering… request info.

Domain Action Interlingua and Lexical Semantic Interlingua  and how will you be paying for this  Domain Action representation:  a:request-information+payment (method=question)  Lexical Semantic representation: predicate: pay time: future agent: hearer product: distance: proximate, type: demonstrative manner: question

Complementary Approaches Domain actions – limited to task oriented sentences Lexical Semantics– less appropriate for formulaic speech acts that should not be translated literally

Components of the Interchange Format a: speaker a: (agent) give-information speech act give-information +availability+room concept* +availability+room (room-type=(single & double), argument* (room-type=(single & double), time=md12) Instructions: Delete sample document icon and replace with working document icons as follows: Create document in Word. Return to PowerPoint. From Insert Menu, select Object… Click “Create from File” Locate File name in “File” box Make sure “Display as Icon” is checked. Click OK Select icon From Slide Show Menu, Select Action Settings. Click “Object Action” and select “Edit” Click OK

Components of IF as of February speech acts give-information –domain independent, –20 are dialog managing 108 concepts availability, accommodation –mostly domain dependent 304 arguments room-type, time –domain dependent and independent 7,652 values single, double, 12th

Examples  no that’s not necessary  c:negate  yes I am  c:affirm  my name is alex waibel  c:give-information+personal-data (person-name=(given- name=alex, family-name=waibel))  and how will you be paying for this  a:request-information+payment (method=question)  I have a mastercard  c:give-information+payment (method=mastercard) Instructions: Delete sample document icon and replace with working document icons as follows: Create document in Word. Return to PowerPoint. From Insert Menu, select Object… Click “Create from File” Locate File name in “File” box Make sure “Display as Icon” is checked. Click OK Select icon From Slide Show Menu, Select Action Settings. Click “Object Action” and select “Edit” Click OK

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Reliability –Coverage –Scalability

Conventional Speec Acts thank you. c:thank can I help you ? a:offer+help (who=i, to-whom=you) my name is Chad c:give-information+personal-data (person-name=(given-name=chad))

Fragments: ellipsis and in a restaurant. a:give-information+concept (conjunction=discourse, location=(restaurant, identifiability=no)) which town? c:request-information+concept (concept-spec=(town, identifiability=question))

Fragments: abandoned You should a: suggest+concept (who=you) What should I c: request-suggestion+concept (who=I)

Coordination of Sentences I want to go to France and I would prefer to leave today. c:give-information+disposition+trip (destination=(object-name=france), disposition=(who=i, desire)) c:give-information+disposition+departure (conjunction=discourse, time=(relative- time=today), disposition=(who=i, preference))

Coordination of sentences, reduced I want to leave Pittsburgh at 2 and return from Rome at 5. c:give-information+disposition+departure (conjunction=discourse, origin=(object- name=pittsburgh), disposition=(who=i, desire), time=(clock=(hours=2))) c:give-information+trip (conjunction=discourse, factuality=unspecified, trip- spec=return, origin=(object-name=rome), time=(clock=(hours=5)))

Conjunctive Set I like festivals and plays. c:give-information+disposition+event (... event-spec=(operator=conjunct, [(festival, quantity=plural), (play, quantity=plural)]))

Conjunction of modifiers I prefer red and blue cars. c:give- information+disposition+vehicle (... vehicle-spec=(car, quantity=plural, color=(operator=conjunct, [red, blue])))

Disjunctive Sets I prefer hotels or cabins. c:give- information+disposition+accommodation (... accommodation- spec=(operator=disjunct, [(hotel, quantity=plural), (cabin, quantity=plural)]))

Contrastive Set I like hotels but not cabins. c:give- information+disposition+accommodat ion (... accommodation- spec=(operator=contrast, [(hotel, quantity=plural), (polarity=negative, cabin, quantity=plural)]))

Attitudes: often a source of mismatches Disposition Eventuality Evidentiality Feasibility Knowledge Obligation Main verbs in English that occur in other languages as affixes, adverbs, or other construtions that are not clearly bi-clausal.

Disposition and I would like to arrive around September ninth. c:give-information+disposition+arrival (disposition=(who=i, desire), /* attitude */ conjunction=discourse, /* rhetorical information */ time=(exactness=approximate, month=9, md=9)) /* time */

Disposition I would like to stay in a hotel. –Disposition=desire I hate mushroom picking. –Disposition=dislike I am waiting to see the circle. –Disposition=expectation But wouldn’t matter. –Disposition=indifferent When do you plan on arriving in Pittsburgh? –Disposition=intention

Eventuality It is possible I may be arriving earlier. give-information+eventuality+arrival (eventuality=possible) I’m sure that they will arrive tomorrow. Maybe there is something beautiful to see. It is not impossible.

Evidentiality: Source of information Apparently there are many castles. Give-information+evidentiality+attraction I heard there are many castles. I noticed there is a winter package available. I’ve been told I must leave before ten.

Feasibility You can rent skis at the resort. Give-information+feasibility+rent+equipment (feasibility=feasible….)

Knowledge I didn’t know that Trento has lakes. Give-information+negation+knowledge+contain+attraction (knowledge=(who=I, polarity=negative), contain=(lake, quantity=plural), attraction-spec=name-trento) I know the location of the hotel.

Obligation You must make a reservation. Give-information+obligation+reservation (obligation=required….) You may cancel at any time. We require that you cancel 24 hours in advance.

Negation: with limited facilities for representing scope Of conventional speech act: I didn’t hear. Negate-dialogue-hear Of main predication: I did not make a reservation. Give-information+negation+reserve+accommodation (polarity=negative…) Of attitude: I don’t know if it’s all right. Give-information+negation+knowledge+feature+object (knowledge=(who=I, polarity=negative), object-spec=pronoun, feature=(modifier=acceptable))

Negation I didn’t promise I would not come. –Negate-promise+negation+action Of a concept: There is no downhill skiiing? Request-information+existence+activity (activity-spec=(polarity=negative, downhill_skiing)

Relative Clauses Broken into two IF’s. I want the hotel that you suggested. Give-information+disposition+accommodation (disposition=(desire, who=I), accommodation-spec=(hotel, identifiability=yes)) Give-information+recommendation+object (object-spec=relative, who=you, e-time=previous) Sentence internal relative clauses (e.g., modifying the subject) are not handled very well. –The only hotel that I can show you is a four star hotel. No long-distance gaps. –They are rare anyway.

Some simple relative clauses aren’t broken The hotel that is in Cavalese give-information+concept (accommodation-spec=(hotel, identifiability=yes, location=name- cavalese))

Yes-No Questions Conventional speech act: Do you hear me? Dialog-request-hear Does the flight leave at 2:00? Tell me if the flight leaves at 2:00. request-information+departure (transportation-spec=(flight, identifiability=yes), time=(clock=(hours=2)))

Wh-questions Who is traveling? request-information+trip (who=question) When are you traveling? What date are you traveling? How quiet is the hotel? Where are you traveling to? How are you traveling? What are you doing? No long distance gaps.

Rhetorical Relations Therefore I arrived late. give-information+arrival (cause=discourse, who=I…) I arrived late because of the snow give-information+arrival (who=I, cause=snow, e-time=previous, time=late)

Rhetorical Relations because I was tired. give-information+feature+person (rhetorical=cause, …) Other relations: after, before, besides, co- occurrence, concessive, condition, contrastive, dependency, purpose, related- to, restrictive-result, result, while

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Reliability –Coverage –Scalability

The Interchange Format Database olang I lang I Prv IRST “telefono per prenotare delle stanze per quattro colleghi” olang I lang E Prv IRST “I’m calling to book some rooms for four colleagues” IF Prv IRST c:request-action+reservation +features+room (for-whom=(associate, quantity=4)) comments: dial-oo5-spkB-roca d.u.sdu olang X lang Y Prv Z sdu in language Y on one line d.u.sdu olang X lang Z Prv Z sdu in language Z on one line d.u.sdu IF Prv Z IF on-one-line d.u. sdu comments: your comments d.u. sdu comments: go here

NESPOLE! Database Annotated turns (end 2001): –English: 815 (235 distinct DAs) –German: 2,873 (367) –Italian: 1,286 (233) –French: 234 (94) Total distinct DAs: 610 Annotated turns (end 2002): 30/40 % more

Tools and Resources IF specifications (available on the web) IF discussion board C-STAR and NESPOLE! Data Bases IF Checker (web interface) IF test suite IF emacs mode

The C-STAR Interchange Format Database English Dialogues English Sentences Korean Dialogues Korean Sentences Italian Dialogues Italian Sentences Japanese Dialogues Japanese Utterances Distinct Dialogue Acts (310 agent, 244 client)

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Reliability –Coverage –Scalability

Comparison of two interlinguas I would like to make a reservation for the fourth through the seventh of July. IF-1 (C-STAR II, ) c:request-action+reservation+temporal+hotel (time=(start-time=md4, end-time=(md7,july))) IF-2 (NESPOLE, ) c:give-information+disposition+reservation +accommodation (disposition=(who=I, desire), reservation-spec=(reservation, identifiability=no), accommodation-spec=hotel, object-time=(start-time=(md=4), end-time=(md=7, month=7, incl-excl=inclusive)))

Comparison of four databases (travel domain, role playing, spontaneous speech) DB-1: C-STAR II English database tagged with IF-1 –2278 sentences DB-2: C-STAR II English database tagged with IF-2 – 2564 sentences DB-3: NESPOLE English database tagged with IF-2 – 1446 sentences –Only about 50% of the vocabulary overlaps with the C-STAR database. DB-4: Combined database tagged with IF-2 –4010 sentences Same data, different interlingua Significantly larger domain

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Reliability –Coverage –Scalability

Measuring Coverage No-tag rate: –Can a human expert assign an interlingua representation to each sentence? –C-STAR II no-tag rate: 7.3% –NESPOLE no-tag rate: 2.4% 300 more sentences were covered in the C-STAR English database End-to-end translation performance: Measures recognizer, analyzer, and generator performance in combination with interlingua coverage.

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Reliability –Coverage –Scalability

Example of failure of reliability Input: 3:00, right? Interlingua: verify (time=3:00) Poor choice of speech act name: does it mean that the speaker is confirming the time or requesting verification from the user? Output: 3:00 is right.

Measuring Reliability: Cross-site evaluations Compare performance of: –Analyzer  interlingua  generator –Where the analyzer and generator are built at the same site (or by the same person) –Where the analyzer and generator are built at different sites (or by different people who may not know each other) C-STAR II interlingua: comparable end-to-end performance within sites and across sites. –around 60% acceptable translations from speech recognizer output. NESPOLE interlingua: cross-site end-to-end performance is lower (but not clearly because of the IF).

Intercoder agreement: average of percent agreeent pairwise Speech actDomain ActionArguments IF-1: Site 1 and Site2 (exp.) 82%66%86% IF-2: Site 1 and Site 2 (4 experts) 92%75%87 % IF-2: Within Site 1 (3 experts) 94%88%90% IF-2: Site 1 vs Site 2 (3 experts and 1 experts) 89%62%83% IF-2: Site 1 and Site 2 (experts and novices) 88%63%86% IF-2: Within Site 2 (expert and novices) 89 %64 %87% IF-2: Within Site 2 (novices) 91%61%83%

Workshop on InterlinguaReliability SIG-IL Association for Machine Translation in the Americas October 8, 2002 Tiburon, California Intent to participate in coding experiment (dependency representation) ( to

Outline Approaches to MT: Interlingua, Transfer, Direct. The NESPOLE! Interlingua. –Overview and motivation –Linguistic coverage Tools and resources. Evaluating an interlingua –Reliability –Coverage –Scalability

Comparison of four databases (travel domain, role playing, spontaneous speech) DB-1: C-STAR II English database tagged with IF-1 –2278 sentences DB-2: C-STAR II English database tagged with IF-2 – 2564 sentences DB-3: NESPOLE English database tagged with IF-2 – 1446 sentences –Only about 50% of the vocabulary overlaps with the C-STAR database. DB-4: Combined database tagged with IF-2 –4010 sentences Same data, different interlingua Significantly larger domain

Measuring Scalability: Coverage Rate What percent of the database is covered by the top n most frequent domain actions? Coverage of 50 most frequent domain actions C-STAR client66.7% NESPOLE client66.5% Combined client62.9% C-STAR agent67.3% NESPOLE agent71.4% Combined agent64.0%

Measuring Scalability: Number of domain actions as a function of database size Sample size from 100 to 3000 sentences in increments of 25. Average number of unique domain actions over ten random samples for each sample size. Each sample includes a random selection of frequent and infrequent domain actions.

Comparison of four databases (travel domain, role playing, spontaneous speech) English database 1 tagged with interlingua 1: 2278 sentences English database 1 tagged with interlingua 2: 2564 sentences English database 2 tagged with interlingua 2: 1446 sentences –Only about 50% of the vocabulary overlaps with the English database 1. Combined databases tagged with interlingua 2: 4010 sentences Same data, different interlingua Significantly larger domain

Conclusions An interlingua based on domain actions is suitable for task-oriented dialogue: –Reliable –Good coverage –Scalable without explosion of domain actions It is possible to evaluate an interlingua for –Realiability –Expressivity –Scalability

How to have success with an interlingua in a multi-site project Keep it simple. Periodically check for intercoder agreement. Good documentation Discussion board for developers Know your language typology.