Evolvable dialogue systems

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Advertisements

Bayesian Belief Propagation
Dialogue Modelling Milica Gašić Dialogue Systems Group.
Dialogue Policy Optimisation
Statistical Dialogue Modelling Milica Gašić Dialogue Systems Group.
On-line dialogue policy optimisation Milica Gašić Dialogue Systems Group.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Yiannis Demiris and Anthony Dearden By James Gilbert.
Modular Neural Networks CPSC 533 Franco Lee Ian Ko.
Gaussian Processes for Fast Policy Optimisation of POMDP-based Dialogue Managers M. Gašić, F. Jurčíček, S. Keizer, F. Mairesse, B. Thomson, K. Yu, S. Young.
MSRC Summer School - 30/06/2009 Cambridge – UK Hybrids of generative and discriminative methods for machine learning.
Classification for High Dimensional Problems Using Bayesian Neural Networks and Dirichlet Diffusion Trees Radford M. Neal and Jianguo Zhang the winners.
Goals of Adaptive Signal Processing Design algorithms that learn from training data Algorithms must have good properties: attain good solutions, simple.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Introduction to Machine Learning Approach Lecture 5.
Helsinki University of Technology Adaptive Informatics Research Centre Finland Variational Bayesian Approach for Nonlinear Identification and Control Matti.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario Silva, Isabel Trancoso Learning Word Representations from Scarce Data By: Aadil Hayat (13002)
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
Machine Learning in Spoken Language Processing Lecture 21 Spoken Language Processing Prof. Andrew Rosenberg.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Learning Automata based Approach to Model Dialogue Strategy in Spoken Dialogue System: A Performance Evaluation G.Kumaravelan Pondicherry University, Karaikal.
Still Talking to Machines (Cognitively Speaking) Machine Intelligence Laboratory Information Engineering Division Cambridge University Engineering Department.
Learning Agents MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
Reinforcement Learning AI – Week 22 Sub-symbolic AI Two: An Introduction to Reinforcement Learning Lee McCluskey, room 3/10
Transfer Learning in Sequential Decision Problems: A Hierarchical Bayesian Approach Aaron Wilson, Alan Fern, Prasad Tadepalli School of EECS Oregon State.
Univariate Gaussian Case (Cont.)
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
Reinforcement Learning for Mapping Instructions to Actions S.R.K. Branavan, Harr Chen, Luke S. Zettlemoyer, Regina Barzilay Computer Science and Artificial.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
A Method to Approximate the Bayesian Posterior Distribution in Singular Learning Machines Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Univariate Gaussian Case (Cont.)
Online Multiscale Dynamic Topic Models
Machine Learning overview Chapter 18, 21
Deep Learning Amin Sobhani.
Machine Learning overview Chapter 18, 21
Eick: Introduction Machine Learning
Chapter 6. Data Collection in a Wizard-of-Oz Experiment in Reinforcement Learning for Adaptive Dialogue Systems by: Rieser & Lemon. Course: Autonomous.
Van-Khanh Tran and Le-Minh Nguyen
Done Done Course Overview What is AI? What are the Major Challenges?
Chapter 6: Temporal Difference Learning
Chapter 11: Learning Introduction
Chapter 5. The Bootstrapping Approach to Developing Reinforcement Learning-based Strategies in Reinforcement Learning for Adaptive Dialogue Systems, V.
CH. 1: Introduction 1.1 What is Machine Learning Example:
COMP61011 : Machine Learning Ensemble Models
Integrating Learning of Dialog Strategies and Semantic Parsing
Deep reinforcement learning for dialogue policy optimisation
Computational models for imaging analyses
Overview of Machine Learning
Reinforcement Learning
Robust Full Bayesian Learning for Neural Networks
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Chapter 6: Temporal Difference Learning
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Deep Reinforcement Learning
LECTURE 09: BAYESIAN LEARNING
LECTURE 07: BAYESIAN ESTIMATION
CHAPTER I. of EVOLUTIONARY ROBOTICS Stefano Nolfi and Dario Floreano
Christoph F. Eick: A Gentle Introduction to Machine Learning
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Sofia Pediaditaki and Mahesh Marina University of Edinburgh
Toward a Great Class Project: Discussion of Stoianov & Zorzi’s Numerosity Model Psych 209 – 2019 Feb 14, 2019.
Machine Learning overview Chapter 18, 21
Instructor: Vincent Conitzer
Presentation transcript:

Evolvable dialogue systems Milica Gašić Dialogue Systems Group

What constitutes a spoken dialogue system? Understanding the user Deciding what to say back Conducting a conversation beyond simple voice commands or question answering

What breaks a spoken dialogue system? Speech recognition errors Not keeping track of what happened previously Need to hand-craft a large number of rules Poor decisions User's request is not supported ...

Machine learning Allows learning from data Automates the development process Uses sophisticated statistical models

Machine learning for dialogue modelling

Data Abundant data / unlabeled data / out-of-domain data Limited data / labeled data / in-domain data Noisy data

Machine learning methods Supervised learning Unsupervised learning Reinforcement learning

Architecture Tree-base architecture (VoiceXML) Modular architecture End-2-end neural network architecture

Prediction What constitutes a good dialogue? How to evaluate dialogue systems? Which measures to use?

Why Machine Learning for dialogue modelling? Cheaper development / avoids handcrafting Robust to errors / models that propagate uncertainty Improves with the time of use / support adaptation According to Darwin's Origin of Species: “It is not the most intellectual of the species that survives; it is not the strongest that survives; but the species that survives is the one that is able best to adapt and adjust to the changing environment in which it finds itself.”

Language Understanding Dialogue management Can I have Korean food? Speech Recognition Language Understanding inform(food=Korean) Belief tracker Policy Optimisation Dialogue Management Knowledge Base A system that allows human computer interaction using speech as the primary medium. Instant and human-like acquisition of information. Speech Synthesis Language Generation inform(name=“Little Seoul”, food=Korean) Little Seoul is a nice Korean restaurant

What is adaptation? Testing in different conditions to training Using a small amount of in-domain data to retrain a system In dialogue modelling Different noise conditions Different user Different domain

Problem: Dialogue systems adaptation – new domain Data: Small amount of in-domain data Model: Method: reinforcement learning, Gaussian processes, Bayesian committee machine, multi-agent learning Architecture: distributed, committee model Predictions: Faster convergence

Machine Learning Theory: Reinforcement learning observations ot Dialogue system reward rt belief states bt action at

Machine Learning Theory: Reinforcement learning observations ot Dialogue system reward rt belief states bt action at Policy

Reinforcement learning in practice

Machine Learning Theory: Gaussian process

Machine Learning Theory: Gaussian process

Machine Learning Theory: Gaussian process

Machine Learning Theory: Gaussian process

Machine Learning Theory: Gaussian process

Q-function as a Gaussian process

Gaussian process Reinforcement learning observations ot Dialogue system reward rt Estimate Q-function action at

Other suitable models Desirable properties Uncertainty estimate Incorporation of the prior knowledge Other possible estimators include Bayesian Neural Networks Need sufficient training data The unsertanty estimates can be obtained using dropout, vatiational inference aproach, recent developments for batch gradient based MCMC. Dropout – remove some hidden untis randomly Variational approach – use another probility distibution to approximate the posterior over parameters Batch versions of MCMC – We use MCMC to approximate the posterior but not the whole training set but in batch

Problem 1: Build dialogue systems from limited data Limited training data available for new domain(s) Solution: Use distributed architecture to train a model from disperse data Essential ingredients: Gaussian process prior

Distributed dialogue management DH +DR MV MV MV MR MH DR DH restaurant restaurant restaurant restaurant restaurant DH DR DL restaurant DH DR DL MR DH DR DL MR DH DR DL MR DH DR DL MR DH DR DL MR MR hotel MH hotel ML laptops MH hotel ML laptops MH hotel ML laptops MH hotel ML laptops MH hotel ML laptops MH ML laptops Committee Model Committee Model Committee Model Committee Model Committee Model Committee Model

Results Gašić et al, Distributed dialogue policies for multi-domain statistical dialogue management, ICASSP, 2015

Problem 2: Domain knowledge is not shared Estimates of Q-function in different domains do not share knowledge Solution: Reuse knowledge obtained in one domain for another domain Essential ingredients: Bayesian committee machine

Committee of dialogue managers DR MR DH MH ML DL Committee Model restaurant restaurant restaurant restaurant restaurant DH DR DL restaurant DH DR DL MR DH DR DL MR DH DR DL MR DH DR DL MR DH DR DL MR MR hotel MH hotel ML laptops MH hotel ML laptops MH hotel ML laptops MH hotel ML laptops MH hotel ML laptops MH ML laptops Committee Model Committee Model Committee Model Committee Model Committee Model Committee Model

Results

Results Gašić et al, Policy Committee for adaptation in multi-domain spoken dialogue systems, ASRU, 2015

Problem 3: Learning from out-of-domain data Estimates of Q-function do not learn from out-of-domain data Solution: Perform learning using both in-domain and out-of-domain data Essential ingredients: Multi-agent learning

Multi-agent learning DR MR DH MH ML DL Committee Model

Multi-agent results Gašić et al, Multi-agent learning in multi-domain spoken dialogue systems, NIPS SLUI workshop, 2015 , 2015

Summary: Adaptation Learning from limited data for a new domain Model for new domain uses knowledge from models for other domains Model for new domain learns from out-of-domain dialogues

Evolution roadmap I feel sad… I’ve got a cold what do I do? Dialogue depth (complexity) Tell me a joke. Single domain systems Extended systems Multi-domain systems Open domain systems What is influenza? Dialogue breadth (coverage)

Properties of natural conversation Less task focused, but more open-ended Beyond speech recognition – incorporates sentiment, gesture and emotion Long-term conversation Incremental operation …

Conclusions Machine learning enables building evolving dialogue systems Gaussian processes in combination with reinforcement learning allow domain adaptation From here we can Model more natural conversation And address new challenging tasks

Dialogue Systems Group Steve Young Lina Rojas Barahona Stefan Ultes Tsung-Hsien Wen Pei-Hao Su Nikola Mrkšić Pawel Budzianowski