Ask and Answer Questions

Slides:



Advertisements
Similar presentations
Date: 2014/05/06 Author: Michael Schuhmacher, Simon Paolo Ponzetto Source: WSDM’14 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Knowledge-based Graph Document.
Advertisements

Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Training dependency parsers by jointly optimizing multiple objectives Keith HallRyan McDonaldJason Katz- BrownMichael Ringgaard.
Retrieval Models for Question and Answer Archives Xiaobing Xue, Jiwoon Jeon, W. Bruce Croft Computer Science Department University of Massachusetts, Google,
Unsupervised Constraint Driven Learning for Transliteration Discovery M. Chang, D. Goldwasser, D. Roth, and Y. Tu.
Date: 2012/08/21 Source: Zhong Zeng, Zhifeng Bao, Tok Wang Ling, Mong Li Lee (KEYS’12) Speaker: Er-Gang Liu Advisor: Dr. Jia-ling Koh 1.
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
DeepWalk: Online Learning of Social Representations
R-NET: Machine Reading Comprehension With Self-Matching Networks
Neural Machine Translation
Customized of Social Media Contents using Focused Topic Hierarchy
Bridging Domains Using World Wide Knowledge for Transfer Learning
End-To-End Memory Networks
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Open question answering over curated and extracted knowledge bases
Constrained Clustering -Semi Supervised Clustering-
Neural Machine Translation by Jointly Learning to Align and Translate
Show and Tell: A Neural Image Caption Generator (CVPR 2015)
Giuseppe Attardi Dipartimento di Informatica Università di Pisa
Are End-to-end Systems the Ultimate Solutions for NLP?
Deep learning and applications to Natural language processing
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
On the Generative Discovery of Structured Medical Knowledge
Summarizing answers in non-factoid community Question-answering
Materials & Methods Introduction Abstract Results Conclusion
Speaker: Jim-an tsai advisor: professor jia-lin koh
Speaker: Jim-an tsai advisor: professor jia-lin koh
Paraphrase Generation Using Deep Learning
Weakly Learning to Match Experts in Online Community
Your topic here Your group list and ID
Final Presentation: Neural Network Doc Summarization
Speaker: Jim-an tsai advisor: professor jia-lin koh
Introduction to Machine Reading Comprehension
Advisor: Jia-Ling Koh Presenter: Yin-Hsiang Liao Source: ACL 2018
Sourse: Www 2017 Advisor: Jia-Ling Koh Speaker: Hsiu-Yi,Chu
Dialogue State Tracking & Dialogue Corpus Survey
Unsupervised Pretraining for Semantic Parsing
Materials & Methods Introduction Abstract Results Conclusion
Enriching Taxonomies With Functional Domain Knowledge
Cheng-Kuan Wei1 , Cheng-Tao Chung1 , Hung-Yi Lee2 and Lin-Shan Lee2
Word2Vec.
Materials & Methods Introduction Abstract Results Conclusion
HeteroMed: Heterogeneous Information Network for Medical Diagnosis
Word Embedding 모든 단어를 vector로 표시 Word vector Word embedding Word
A Data-Driven Question Generation Model for Educational Content
(max 5 slides and 5 minutes) (-5 point for each extra slide or minute)
Attention for translation
Speaker: Ming-Chieh, Chiang
Deep Interest Network for Click-Through Rate Prediction
Neural Joint Model for Transition-based Chinese Syntactic Analysis
Jointly Generating Captions to Aid Visual Question Answering
Presented by: Anurag Paul
Keshav Balasubramanian
Materials & Methods Introduction Abstract Results Conclusion
The experiments based on Recurrent Neural Networks
Baseline Model CSV Files Pandas DataFrame Sentence Lists
Heterogeneous Graph Attention Network
Attention Is All You Need
Week 3 Presentation Ngoc Ta Aidean Sharghi.
Bidirectional LSTM-CRF Models for Sequence Tagging
Improving Cross-lingual Entity Alignment via Optimal Transport
Da-Rong Liu, Kuan-Yu Chen, Hung-Yi Lee, Lin-shan Lee
Materials & Methods Introduction Abstract Results Conclusion
Neural Machine Translation by Jointly Learning to Align and Translate
Learning Dual Retrieval Module for Semi-supervised Relation Extraction
(Type Answer Here) (Type Answer Here) (Type Answer Here)
The experiment based on hier-attention
Faithful Multimodal Explanation for VQA
Presentation transcript:

Ask and Answer Questions Self-Training for Jointly Learning to Ask and Answer Questions Advisor: Jia-lin Koh Speaker: Yin-hsiang Liao Source: NAACL 2018 Date: Dec11/ 2018

Outline Introduction Method Experiment Conclusion

Introduction Question answering and question generation QA: well-studied problem in NLP which focus on answering questions using structured or unstructured source of knowledge. QG: Focusing on generating question from given knowledge.

Introduction Separate tasks  A joint learning task QA: Answering a question q based on a passage p. A sentence selection task here. QG: Transforming a sentence in the passage into a question

Outline Introduction Method Experiment Conclusion

Method Overview Train initial QA/QG model Get Q-A pairs from unlabeled text Update the models

Get Q-A pairs from unlabeled text Method Train initial QA/QG model Get Q-A pairs from unlabeled text Update the models QA: select an answer sentence from the passage Using two bi-LSTMs as encoders for passages and questions. Transforming a passage to a contextual vector

Get Q-A pairs from unlabeled text Method Train initial QA/QG model Get Q-A pairs from unlabeled text Update the models QA: select an answer sentence from the passage Final answer: Loss function:

Get Q-A pairs from unlabeled text Method Train initial QA/QG model Get Q-A pairs from unlabeled text Update the models QG: generating question from a passage Using S2S model, a bi-LSTM as encoder and LSTM as decoder. Effective Approaches to Attention-based Neural Machine Translation

Method Train initial QA/QG model Get Q-A pairs from unlabeled text Update the models After training QA, QG models from labeled corpus, the QG model are used to create more questions from the unlabeled text corpus and the QA model is used to answer these questions. Therefore, we get new Q-A pairs.

Method The problem is how to selection them? Train initial QA/QG model Get Q-A pairs from unlabeled text Update the models The problem is how to selection them? Question selection oracles introduce the most confident pairs. Oracles are: QG QA QG + QA (max min conf.)

Method Curriculum learning Train initial QA/QG model Get Q-A pairs from unlabeled text Update the models Curriculum learning Easy question first, larder latter. The challenge is how to define ‘easiness’ Here are some heuristics: P.632

Method Train initial QA/QG model Get Q-A pairs from unlabeled text Update the models

Outline Introduction Method Experiment Conclusion

Experiment Word embedding size d = 100, pretrained GloVe word embeddings. Datasets: SQUAD, MS MARCO, WikiQA and TrecQA.

Experiment Oracle matters for QA

Experiment Oracle matters for QG

Outline Introduction Method Experiment Conclusion

Conclusion Self-training algorithm for jointly learning to answer and ask questions. The proposed model led improvement for both QA, QG tasks.