Introduction to Machine Reading Comprehension

Slides:



Advertisements
Similar presentations
Passage Retrieval & Re-ranking Ling573 NLP Systems and Applications May 5, 2011.
Advertisements

Search and Retrieval: More on Term Weighting and Document Ranking Prof. Marti Hearst SIMS 202, Lecture 22.
Distributed Representations of Sentences and Documents
Scalable Text Mining with Sparse Generative Models
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Modelling Language Evolution Lecture 2: Learning Syntax Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
Mianwei Zhou, Kevin Chen-Chuan Chang University of Illinois at Urbana-Champaign Unifying Learning to Rank and Domain Adaptation -- Enabling Cross-Task.
Carnegie Mellon School of Computer Science Copyright © 2001, Carnegie Mellon. All Rights Reserved. JAVELIN Project Briefing 1 AQUAINT Phase I Kickoff December.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
1/21 Automatic Discovery of Intentions in Text and its Application to Question Answering (ACL 2005 Student Research Workshop )
Reasoning, Attention, Memory (RAM) NIPS Workshop 2015
A Survey on Automatic Text Summarization Dipanjan Das André F. T. Martins Tolga Çekiç
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
11 Thoughts on STS regarding Machine Reading Ralph Weischedel 12 March 2012.
Fill-in-The-Blank Using Sum Product Network
Ensembling Diverse Approaches to Question Answering
Attention Model in NLP Jichuan ZENG.
R-NET: Machine Reading Comprehension With Self-Matching Networks
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
End-To-End Memory Networks
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
A Deep Memory Network for Chinese Zero Pronoun Resolution
Dhruv Batra Georgia Tech
Zero-Shot Relation Extraction via Reading Comprehension
Topics Question answering at Bing
Jiani ZHANG Tuesday, October 4, 2016
Unifying QA, Dialog, VQA and Visual Dialog
Van-Khanh Tran and Le-Minh Nguyen
Adversarial Learning for Neural Dialogue Generation
Attention Is All You Need
Visualizing and Understanding Neural Models in NLP
CSCI 5922 Neural Networks and Deep Learning: Image Captioning
Are End-to-end Systems the Ultimate Solutions for NLP?
mengye ren, ryan kiros, richard s. zemel
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Know What You Don’t Know: Unanswerable Questions for SQuAD
Image Question Answering
Deep Learning based Machine Translation
convolutional neural networkS
Distributed Representation of Words, Sentences and Paragraphs
Toward improved document classification and retrieval
Final Presentation: Neural Network Doc Summarization
Improving DevOps and QA efficiency using machine learning and NLP methods Omer Sagi May 2018.
Memory Networks for Question Answering
MEgo2Vec: Embedding Matched Ego Networks for User Alignment Across Social Networks Jing Zhang+, Bo Chen+, Xianming Wang+, Fengmei Jin+, Hong Chen+, Cuiping.
CS548 Fall 2018 Text Mining Showcase
Unsupervised Pretraining for Semantic Parsing
ConvAI2 Competition: FUTURE WORK
Report by: 陆纪圆.
MTBI Personality Predictor using ML
Rich Context Competition
Textual Video Prediction
Word Embedding 모든 단어를 vector로 표시 Word vector Word embedding Word
Attention for translation
-- Ray Mooney, Association for Computational Linguistics (ACL) 2014
How Much is 131 Million Dollars
Presented by: Anurag Paul
Building Dynamic Knowledge Graphs From Text Using Machine Reading Comprehension Rajarshi Das, Tsendsuren Munkhdalai, Xingdi Yuan, Adam Trischler, Andrew.
Ask and Answer Questions
CoQA: A Conversational Question Answering Challenge
Extracting Why Text Segment from Web Based on Grammar-gram
Baseline Model CSV Files Pandas DataFrame Sentence Lists
Sequence-to-Sequence Models
Knowledge Representation and Reasoning
Bidirectional LSTM-CRF Models for Sequence Tagging
Visual Grounding.
A Neural Passage Model for Ad-hoc Document Retrieval
Presentation transcript:

Introduction to Machine Reading Comprehension Runqi Yang runqiyang@gmail.com

What is reading comprehension? Provided with one or more pieces of evidence, give the answer to a question. 2/15/2019 Introduction to Machine Reading Comprehension

Difference from other QA settings General QA: use everything rules/templates, knowledge bases/dictionaries, raw text, ... Reading comprehension: a single knowledge source forces the model to “comprehend” 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension 4 Phases Small real data/large synthetic data: bAbI, ... Case study: MemNN Cloze-style Reading Comprehension: CNN/Daily Mail, CBT, ... Case study: AoA Reader Single evidence, a continuous span as the answer: SQuAD, ... Case study: BiDAF Multi evidence, no restriction to answers: MS MACRO, Trivia QA, ... Case study: DrQA 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension bAbI Dataset Synthetic data 20 tasks Single Supporting Fact Two Supporting Facts Counting Simple Negation Time Reasoning ...... Weston, Jason, et al. "Towards AI-complete question answering: A set of prerequisite toy tasks."  (2015). 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension Case study: MemNN 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension CNN/Daily Mail & CBT Real data, automatic generated CNN/Daily Mail: fill in the blank with entity in summary Children’s Book Test: select to fill in the blank in the next sentence Large in scale CNN/Daily Mail: 1.4M Q-A pairs Children’s Book Test: 688K Q-A pairs Not natural 2/15/2019 Introduction to Machine Reading Comprehension

CNN/Daily Mail-example 2/15/2019 Introduction to Machine Reading Comprehension

Children’s Book-example 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension Case study: AoA Reader 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension SQuAD “ImageNet in NLP” Hidden test set with leaderboard From Wikipedia, human-labeled Human generated question, a continuous span as the answer Evaluation: EM & F1 ≥3 answers in dev & test set Human performance: EM 82%, F1 91% Rajpurkar, Pranav, et al. "Squad: 100,000+ questions for machine comprehension of text." (2016). 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension Case study: BiDAF 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension MS MACRO More diverse passages 100K queries, the same as SQuAD #documents: 200K+ vs 536 Natural QA Questions from user log Answers not limited to span Evaluation: Rouge-L & Bleu-1 human performace: 47, 46 Nguyen, Tri, et al. "MS MACRO: A human generated machine reading comprehension dataset." (2016). 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension Case study: DrQA 2/15/2019 Introduction to Machine Reading Comprehension

Case study: DrQA-Retriever Baseline Wikipedia Search API: TF-IDF weighted BOW vectors Improvement Hashed Bigram 2/15/2019 Introduction to Machine Reading Comprehension

Case study: DrQA-Reader (1) Paragraph Embedding GloVe vectors exact match POS, NER, normalized TF aligned question embedding: project-dot-softmax-weighted average Paragraph Encoding 3-layer BiLSTM Question Encoding 3-layer BiLSTM + attention pooling 2/15/2019 Introduction to Machine Reading Comprehension

Case study: DrQA-Reader (2) Answer Prediction bilinear score (unnormalized exponential): find the answer span: take argmax over all considered paragraph spans for the final prediction. 2/15/2019 Introduction to Machine Reading Comprehension

Implementation of DrQA https://github.com/hitvoice/DrQA 2/15/2019 Introduction to Machine Reading Comprehension

Introduction to Machine Reading Comprehension Reference Weston, Jason, et al. "Towards ai-complete question answering: A set of prerequisite toy tasks." arXiv preprint arXiv:1502.05698 (2015). Hermann, Karl Moritz, et al. "Teaching machines to read and comprehend." Advances in Neural Information Processing Systems. 2015. Hill, Felix, et al. "The Goldilocks Principle: Reading Children's Books with Explicit Memory Representations." arXiv preprint arXiv:1511.02301(2015). Nguyen, Tri, et al. "Ms marco: A human generated machine reading comprehension dataset." arXiv preprint arXiv:1611.09268 (2016). Rajpurkar, Pranav, et al. "Squad: 100,000+ questions for machine comprehension of text." arXiv preprint arXiv:1606.05250 (2016). Sukhbaatar, Sainbayar, Jason Weston, and Rob Fergus. "End-to-end memory networks." Advances in neural information processing systems. 2015. Cui, Yiming, et al. "Attention-over-attention neural networks for reading comprehension." arXiv preprint arXiv:1607.04423 (2016). Seo, Minjoon, et al. "Bidirectional attention flow for machine comprehension." arXiv preprint arXiv:1611.01603 (2016). Chen, Danqi, et al. "Reading Wikipedia to Answer Open-Domain Questions." arXiv preprint arXiv:1704.00051 (2017). 2/15/2019 Introduction to Machine Reading Comprehension