Develop a fast semantic decoder for dialogue systems Capability to parse 10 – 100 ASR hypotheses in real time Robust to speech recognition noise Semantic.

Slides:



Advertisements
Similar presentations
Machine Learning Approaches to the Analysis of Large Corpora : A Survey Xunlei Rose Hu and Eric Atwell University of Leeds.
Advertisements

Hidden Information State System A Statistical Spoken Dialogue System M. Gašić, F. Jurčíček, S. Keizer, F. Mairesse, B. Thomson, K. Yu and S. Young Cambridge.
Dialogue Policy Optimisation
Probabilistic Adaptive Real-Time Learning And Natural Conversational Engine Seventh Framework Programme FP7-ICT
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
15.0 Utterance Verification and Keyword/Key Phrase Spotting References: 1. “Speech Recognition and Utterance Verification Based on a Generalized Confidence.
Event Extraction Using Distant Supervision Kevin Reschke, Martin Jankowiak, Mihai Surdeanu, Christopher D. Manning, Daniel Jurafsky 30 May 2014 Language.
1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong.
Gaussian Processes for Fast Policy Optimisation of POMDP-based Dialogue Managers M. Gašić, F. Jurčíček, S. Keizer, F. Mairesse, B. Thomson, K. Yu, S. Young.
1 I256: Applied Natural Language Processing Marti Hearst Sept 27, 2006.
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
Course Summary LING 575 Fei Xia 03/06/07. Outline Introduction to MT: 1 Major approaches –SMT: 3 –Transfer-based MT: 2 –Hybrid systems: 2 Other topics.
Probabilistic Parsing Ling 571 Fei Xia Week 5: 10/25-10/27/05.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
ANTLR.
STRUCTURED PERCEPTRON Alice Lai and Shi Zhi. Presentation Outline Introduction to Structured Perceptron ILP-CRF Model Averaged Perceptron Latent Variable.
Spoken Language Understanding for Conversational Dialog Systems
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
1 The Hidden Vector State Language Model Vidura Senevitratne, Steve Young Cambridge University Engineering Department.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
Lecture 12: 22/6/1435 Natural language processing Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Some Thoughts on HPC in Natural Language Engineering Steven Bird University of Melbourne & University of Pennsylvania.
A Natural Language Interface for Crime-related Spatial Queries Chengyang Zhang, Yan Huang, Rada Mihalcea, Hector Cuellar Department of Computer Science.
The Impact of Grammar Enhancement on Semantic Resources Induction Luca Dini Giampaolo Mazzini
A Semantic Approach to IE Pattern Induction Mark Stevenson and Mark Greenwood Natural Language Processing Group University of Sheffield, UK.
An Extended GHKM Algorithm for Inducing λ-SCFG Peng Li Tsinghua University.
Evaluation of SDS Svetlana Stoyanchev 3/2/2015. Goal of dialogue evaluation Assess system performance Challenges of evaluation of SDS systems – SDS developer.
Part-Of-Speech Tagging using Neural Networks Ankur Parikh LTRC IIIT Hyderabad
Discriminative Models for Spoken Language Understanding Ye-Yi Wang, Alex Acero Microsoft Research, Redmond, Washington USA ICSLP 2006.
Develop a fast semantic decoder Robust to speech recognition noise Trainable on different domains: Tourist information (TownInfo) Air travel information.
Presented by: Fang-Hui Chu Boosting HMM acoustic models in large vocabulary speech recognition Carsten Meyer, Hauke Schramm Philips Research Laboratories,
Interpreter CS 124 Reference: Gamma et al (“Gang-of-4”), Design Patterns Some material taken from:
DIALOG SYSTEMS FOR AUTOMOTIVE ENVIRONMENTS Presenter: Joseph Picone Inst. for Signal and Info. Processing Dept. Electrical and Computer Eng. Mississippi.
Markov Logic and Deep Networks Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
Lecture 12 Applications and demos. Building applications Previous lectures have discussed stages in processing: algorithms have addressed aspects of language.
A Semantic Approach to IE Pattern Induction Mark Stevenson and Mark A. Greenwood Natural Language Processing Group University of Sheffield, UK.
16.0 Spoken Dialogues References: , Chapter 17 of Huang 2. “Conversational Interfaces: Advances and Challenges”, Proceedings of the IEEE,
Still Talking to Machines (Cognitively Speaking) Machine Intelligence Laboratory Information Engineering Division Cambridge University Engineering Department.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
DIALOG SYSTEMS FOR AUTOMOTIVE ENVIRONMENTS Presenter: Joseph Picone Inst. for Signal and Info. Processing Dept. Electrical and Computer Eng. Mississippi.
MROrder: Flexible Job Ordering Optimization for Online MapReduce Workloads School of Computer Engineering Nanyang Technological University 30 th Aug 2013.
Speech Communication Lab, State University of New York at Binghamton Dimensionality Reduction Methods for HMM Phonetic Recognition Hongbing Hu, Stephen.
Dynamic Slot Allocation Technique for MapReduce Clusters School of Computer Engineering Nanyang Technological University 25 th Sept 2013 Shanjiang Tang,
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Lti Shaping Spoken Input in User-Initiative Systems Stefanie Tomko and Roni Rosenfeld Language Technologies Institute School of Computer Science Carnegie.
Disambiguating inventor names using deep neural networks Steve Petrie T’Mir Julius.
CS 4705 Lecture 17 Semantic Analysis: Robust Semantics.
金聲玉振 Taiwan Univ. & Academia Sinica 1 Spoken Dialogue in Information Retrieval Jia-lin Shen Oct. 22, 1998.
Develop a fast semantic decoder for dialogue systems Capability to parse 10 – 100 ASR hypothesis in real time Robust to speech recognition noise Trainable.
Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore
Exploiting Named Entity Taggers in a Second Language Thamar Solorio Computer Science Department National Institute of Astrophysics, Optics and Electronics.
Integrating Multiple Knowledge Sources For Improved Speech Understanding Sherif Abdou, Michael Scordilis Department of Electrical and Computer Engineering,
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Department of Computer Science The University of Texas at Austin USA Joint Entity and Relation Extraction using Card-Pyramid Parsing Rohit J. Kate Raymond.
Part-of-Speech Tagging CSCI-GA.2590 – Lecture 4 Ralph Grishman NYU.
Gaussian Mixture Language Models for Speech Recognition Mohamed Afify, Olivier Siohan and Ruhi Sarikaya.
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Automatic Speech Recognition
Van-Khanh Tran and Le-Minh Nguyen
Semantic Parsing for Question Answering
Learning to Transform Natural to Formal Languages
Data Recombination for Neural Semantic Parsing
Learning to Parse Database Queries Using Inductive Logic Programming
PROJ2: Building an ASR System
Unsupervised Pretraining for Semantic Parsing
Natural Language to SQL(nl2sql)
Parsing Unrestricted Text
PRESENTATION: GROUP # 5 Roll No: 14,17,25,36,37 TOPIC: STATISTICAL PARSING AND HIDDEN MARKOV MODEL.
Presentation transcript:

Develop a fast semantic decoder for dialogue systems Capability to parse 10 – 100 ASR hypotheses in real time Robust to speech recognition noise Semantic parsing maps natural language to formal language In current dialogue systems, most of the mapping is fairly simple (cities, times,... ) Transformation based – error driven learning was shown to be efficient and fast on this type of task find all flights between Toronto and San Diego that arrive on Saturday 1.Parser assigns initial semantics to input sentence 2.Rules are sequentially applied, whenever a trigger matches the sentence and the hypothesised semantics Compare TBL Semantic Parser with: A handcrafted Phoenix grammar The Hidden Vector State model (He & Young, 2006) Probabilistic Combinatory Categorial Grammar Induction (Zettlemoyer & Collins, 2007) Markov Logic networks (Meza-Ruiz et al., 2008) Semantic Tuple Classifiers (Mairesse et al., 2008) Evaluation metrics used are precision, recall and F-measure of dialogue act items (e.g. food=Chinese, toloc.city=New York) Conclusion: TBL Semantic Parser is robust to noise and competitive with the state of the art. The code is available at Transformation-based Learning for Semantic Parsing F. Jurcicek, M. Gasic, S. Keizer, F. Mairesse, B. Thomson, K. Yu, S. Young Cambridge University Engineering Department | {fj228, mg436, sk561, f.mairesse, brmt2, ky219, Goals & Motivation Example of parsing 3.Rules can correct previous errors On average 1 million rules is tested in each iteration Training takes about 24 hours on an Intel Pentium 2.8 GHz Inferred list of rules is very small for domains such as ATIS or TownInfo Average number of semantic concepts in a sentence is 5 Python implementation of the parser needs 6ms per sentence Acknowledgements This research was partly funded by the UK EPSRC under grant agreement EP/F013930/1 and by the EU FP7 Programme under grant agreement (CLASSiC project: ). Evaluation Semantic decoderItem precision Item recall Item F ATIS dataset with transcribed utterances: Semantic Tuple Classifiers Hidden Vector State90.3 PCCG Induction Markov Logic Networks TBL Semantic Parser TownInfo dataset with transcribed utterances: Semantic Tuple Classifiers Phoenix grammar TBL Semantic Parser TownInfo dataset with ASR output: Semantic Tuple Classifiers Phoenix grammar TBL Semantic Parser what are the lowest airfares from Washington to Boston GOAL= airfare airfare.type= lowest from.city= Washington to.city= Boston Open source Transformations TriggersTransformations “Seattle"add the slot "to.city=Seattle" “connecting"replace the slot “to.city=*" by “stop.city=*“ “from * Francisco”delete the slot “toloc.city=San Francisco" GOAL= flight from.city=Toronto to.city=San Diego departure.day=Saturday #TriggerTransformation 1“between Toronto"add the slot “from.city=Toronto“ 2“and San Diego"add the slot “to.city=Sand Diego“ 3“Saturday"add the slot “departure.day=Saturday" #TriggerTransformation 4“arrive"replace the slot “departure.day=*" by “arrival.day=*" GOAL= flight from.city=Toronto to.city=San Diego arrive.day=Saturday Efficiency Dataset # of inferred rulesconcepts# of rules/concepts ATIS TownInfo Learning Test the rules Select the best rule Apply the best rule Exit, if no improvement Generate applicable rules Initial semantics