Van-Khanh Tran and Le-Minh Nguyen

Slides:



Advertisements
Similar presentations
SOMM: Self Organizing Markov Map for Gesture Recognition Pattern Recognition 2010 Spring Seung-Hyun Lee G. Caridakis et al., Pattern Recognition, Vol.
Advertisements

Application of RNNs to Language Processing Andrey Malinin, Shixiang Gu CUED Division F Speech Group.
Develop a fast semantic decoder for dialogue systems Capability to parse 10 – 100 ASR hypotheses in real time Robust to speech recognition noise Semantic.
Develop a fast semantic decoder Robust to speech recognition noise Trainable on different domains: Tourist information (TownInfo) Air travel information.
 Ontology Induction (Chen et al., 2013 & 2014) Frame-semantic parsing on ASR results (Das et al., 2013) frame  slot candidate lexical unit  slot filler.
Learning to Answer Questions from Image Using Convolutional Neural Network Lin Ma, Zhengdong Lu, and Hang Li Huawei Noah’s Ark Lab, Hong Kong
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
A Sentence Interaction Network for Modeling Dependence between Sentences Biao Liu, Minlie Huang Tsinghua University.
Attention Model in NLP Jichuan ZENG.
R-NET: Machine Reading Comprehension With Self-Matching Networks
Evolvable dialogue systems
Best viewed with Computer Modern fonts installed
IEEE BIBM 2016 Xu Min, Wanwen Zeng, Ning Chen, Ting Chen*, Rui Jiang*
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Deep Learning Amin Sobhani.
Wu et. al., arXiv - sept 2016 Presenter: Lütfi Kerem Şenel
Recurrent Neural Networks for Natural Language Processing
Neural Machine Translation by Jointly Learning to Align and Translate
Attention Is All You Need
A Hierarchical Model of Reviews for Aspect-based Sentiment Analysis
ICS 491 Big Data Analytics Fall 2017 Deep Learning
Intelligent Information System Lab
Synthesis of X-ray Projections via Deep Learning
Neural networks (3) Regularization Autoencoder
Deep learning and applications to Natural language processing
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Image Question Answering
Deep reinforcement learning for dialogue policy optimisation
Master’s Thesis defense Ming Du Advisor: Dr. Yi Shang
Deep Learning based Machine Translation
Recursive Structure.
Attention-based Caption Description Mun Jonghwan.
Grid Long Short-Term Memory
Advanced Artificial Intelligence
Paraphrase Generation Using Deep Learning
Image Captions With Deep Learning Yulia Kogan & Ron Shiff
Final Presentation: Neural Network Doc Summarization
Understanding LSTM Networks
Code Completion with Neural Attention and Pointer Networks
The Big Health Data–Intelligent Machine Paradox
Introduction to Machine Reading Comprehension
Other Classification Models: Recurrent Neural Network (RNN)
Introduction to Natural Language Processing
Machine Translation(MT)
Dialogue State Tracking & Dialogue Corpus Survey
Unsupervised Pretraining for Semantic Parsing
Natural Language to SQL(nl2sql)
Report by: 陆纪圆.
Attention.
实习生汇报 ——北邮 张安迪.
Neural networks (3) Regularization Autoencoder
Designing architectures by hand is hard
Please enjoy.
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Attention for translation
Learn to Comment Mentor: Mahdi M. Kalayeh
Jointly Generating Captions to Aid Visual Question Answering
Presented by: Anurag Paul
Neural Machine Translation using CNN
The experiments based on Recurrent Neural Networks
Question Answering System
Baseline Model CSV Files Pandas DataFrame Sentence Lists
Recurrent Neural Networks
Bidirectional LSTM-CRF Models for Sequence Tagging
LHC beam mode classification
Neural Machine Translation by Jointly Learning to Align and Translate
Mahdi Kalayeh David Hill
A Neural Network for Car-Passenger matching in Ride Hailing Services.
The experiment based on hier-attention
Presentation transcript:

Natural Language Generation for Spoken Dialogue System using RNN Encoder-Decoder Networks Van-Khanh Tran and Le-Minh Nguyen Japan Advanced Institute of Science and Technology, JAIST 1-1 Asahidai, Nomi, Ishikawa, 923-1292, JAPAN

The Neural Language Generator Content Introduction The Neural Language Generator Attention-based RNN Encoder-Decoder RALSTM cell Experiments Conclusion

The Neural Language Generator Content Introduction The Neural Language Generator Attention-based RNN Encoder-Decoder RALSTM cell Experiments Conclusion

NLG Task Mapping a MR (meaning representation) to a natural language utterance. Dialogue Act inform(name=Bar Crudo, food=Mexican) Realizations Bar Crudo is a Mexican restaurant. Bar Crudo serves Mexican food

Natural Language Generator NLG processes Bar Crudo serves Mexican food Lexicalization Step SLOT_NAME serves SLOT_FOOD food RNN-based LSTM-based GRU-based Encoder-Decoder based … Natural Language Generator SLOT_NAME serves SLOT_FOOD food Delexicalization Step Bar Crudo serves Mexican food inform(name=Bar Crudo, food=Mexican)

The Neural Language Generator Content Introduction The Neural Language Generator Attention-based RNN Encoder-Decoder RALSTM cell Experiments Conclusion

The Neural Language Generator Wen et. al, Toward multi-domain language generation using recurrent neural networks, 2016.

The Neural Language Generator Content Introduction The Neural Language Generator Attention-based RNN Encoder-Decoder RALSTM cell Experiments Conclusion

Attention-based RNN Encoder-Decoder wt … e1 e2 eT inform name= Bar Crudo price-range= moderate food= Mexican Aligner dt RNN Separate parameterization of Slot-Value pair

Attention-based RNN Encoder-Decoder wt … e1 e2 eT inform name= Bar Crudo price-range= moderate food= Mexican Aligner dt RALSTM st-1 st st+1 Dialog Act 1-hot representation Separate parameterization of Slot-Value pair

DA feature vector s controlling

RALSTM Cell ct wt ht wt dt dt it ot ht-1 ht-1 wt wt dt dt ft ht-1 ht-1 tanh wt wt ct dt tanh dt ft ht-1 ht-1 xt Refinement Cell tanh rt wt dt ht-1

RALSTM Cell ct st-1 st ha ht xt at xt ht xt dt dt it ot ht-1 ht-1 xt tanh Adjustment Cell xt LSTM Cell ht xt dt dt it ot ht-1 ht-1 tanh xt xt ct dt tanh dt ft ht-1 ht-1 xt Refinement Cell tanh rt wt dt ht-1

The Neural Language Generator Content Introduction The Neural Language Generator Attention-based RNN Encoder-Decoder RALSTM cell Experiments Conclusion

Experiments Datasets collected by Wen et. al 2015a,b, 2016 Finding a restaurant Finding a hotel Buying a Laptop Buying a TV Training: BPTT, SGD with Early stopping, L2 reg, Hidden size 80, Keep Dropout 70%. Evaluation metrics: BLEU Slot error rate ERR

Generated Outputs

The Neural Language Generator Content Introduction The Neural Language Generator Attention-based RNN Encoder-Decoder RALSTM cell Experiments Conclusion

Conclusion Proposed RALSTM architecture Training NLG N2N using Attentional RNN Encoder- Decoder Networks Evaluation metrics

Thanks for your attention! Question?

References Tsung-Hsien Wen, Milica Gasˇic ́, Dongho Kim, Nikola Mrksˇic ́, Pei-Hao Su, David Vandyke, and Steve Young. 2015a. Stochastic Language Generation in Dialogue using Recurrent Neural Networks with Convolutional Sentence Reranking. In Proceedings SIGDIAL. Association for Computational Linguistics. Tsung-Hsien Wen, Milica Gasic, Nikola Mrksic, Lina M Rojas-Barahona, Pei-Hao Su, David Vandyke, and Steve Young. 2016a. Multi-domain neural network language generation for spoken dia- logue systems. arXiv preprint arXiv:1603.01232 . Tsung-Hsien Wen, Milica Gasˇic, Nikola Mrksˇic, Lina M Rojas-Barahona, Pei-Hao Su, David Vandyke, and Steve Young. 2016b. Toward multi- domain language generation using recurrent neural networks . Tsung-Hsien Wen, Milica Gasˇic ́, Nikola Mrksˇic ́, Pei- Hao Su, David Vandyke, and Steve Young. 2015b. Semantically conditioned lstm-based natural lan- guage generation for spoken dialogue systems. In Proceedings of EMNLP. Association for Computtional Linguistics. Tsung-Hsien Wen, David Vandyke, Nikola Mrksic, Milica Gasic, Lina M Rojas-Barahona, Pei-Hao Su, Stefan Ultes, and Steve Young. 2016c. A network- based end-to-end trainable task-oriented dialogue system. arXiv preprint arXiv:1604.04562 .