Reasoning, Attention, Memory (RAM) NIPS Workshop 2015

Slides:



Advertisements
Similar presentations
Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
Advertisements

Advanced topics.
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
Deep Learning in NLP Word representation and how to use it for Parsing
Deep Learning and Neural Nets Spring 2015
Deep Learning.
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Deep Learning for Speech and Language Yoshua Bengio, U. Montreal NIPS’2009 Workshop on Deep Learning for Speech Recognition and Related Applications December.
Steps Toward an AGI Roadmap Włodek Duch ( Google: W. Duch) AGI, Memphis, 1-2 March 2007 Roadmaps: A Ten Year Roadmap to Machines with Common Sense (Push.
I Robot.
CS10: The Beauty and Joy of Computing Artificial Intelligence Anna Rafferty (Slides adapted from Dan Garcia) 18 November 2013.
1 The main topics in AI Artificial intelligence can be considered under a number of headings: –Search (includes Game Playing). –Representing Knowledge.
Neural Net Language Models
Kai Sheng-Tai, Richard Socher, Christopher D. Manning
ECE 6504: Deep Learning for Perception Dhruv Batra Virginia Tech Topics: –Recurrent Neural Networks (RNNs) –BackProp Through Time (BPTT) –Vanishing / Exploding.
Haitham Elmarakeby.  Speech recognition
A Roadmap towards Machine Intelligence
Considerations for Evaluating Models of Language Understanding and Reasoning Gabriel Recchia University of Cambridge.
Deep Visual Analogy-Making
Three challenges for computational models of cognition
1 Artificial Intelligence & Prolog Programming CSL 302.
Memory Networks for Language Understanding
Learning to Answer Questions from Image Using Convolutional Neural Network Lin Ma, Zhengdong Lu, and Hang Li Huawei Noah’s Ark Lab, Hong Kong
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Variational Autoencoders Theory and Extensions
Reading Documents with Memory Networks
Memory Networks for Language Understanding
Conditional Generation by RNN & Attention
Brief Intro to Machine Learning CS539
Neural Architectures with Memory
Attention Model in NLP Jichuan ZENG.
Lecture 6 Smaller Network: RNN
R-NET: Machine Reading Comprehension With Self-Matching Networks
(End-to-End Methods for) Dialogue, Interaction and Learning
Unsupervised Learning of Video Representations using LSTMs
Hierarchical Question-Image Co-Attention for Visual Question Answering
End-To-End Memory Networks
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Natural Language and Text Processing Laboratory
Chapter 11: Artificial Intelligence
Jiani ZHANG Tuesday, October 4, 2016
Unifying QA, Dialog, VQA and Visual Dialog
Neural Machine Translation by Jointly Learning to Align and Translate
Show and Tell: A Neural Image Caption Generator (CVPR 2015)
References [1] - Y. LeCun, L. Bottou, Y. Bengio and P. Haffner, Gradient-Based Learning Applied to Document Recognition, Proceedings of the IEEE, 86(11): ,
Intelligent Information System Lab
Different Units Ramakrishna Vedantam.
Deep learning and applications to Natural language processing
Machine Comphrension Heng Ji (Many slides from Antoine Bordes)
Please, Pay Attention, Neural Attention
Trends and Future of NLP
Basic Intro Tutorial on Machine Learning and Data Mining
Deep Learning based Machine Translation
convolutional neural networkS
Distributed Representation of Words, Sentences and Paragraphs
Advanced Recurrent Architectures
convolutional neural networkS
Memory Networks for Question Answering
Understanding LSTM Networks
Seminar Topics and Projects
Introduction to Machine Reading Comprehension
Introduction to Natural Language Processing
ConvAI2 Competition The task
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Question Answering System
Deep learning: Recurrent Neural Networks CV192
Bidirectional LSTM-CRF Models for Sequence Tagging
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Reasoning, Attention, Memory (RAM) NIPS Workshop 2015 Organizers: Jason Weston, Sumit Chopra and Antoine Bordes

Recent Excitement !! RAM is not a new subject… In this intro we will summarize some of the recent trends.

Learning of Basic Algorithms (e.g. addition, multiplication, sorting) Methods include adding stacks and addressable memory to RNNs “Neural Net Architectures for Temporal Sequence Processing” M. Mozer. “Neural Turing Machines” A. Graves, G. Wayne, I. Danihelka. “Inferring Algorithmic Patterns with Stack Augmented Recurrent Nets” A. Joulin, T. Mikolov. “Learning to Transduce with Unbounded Memory” E. Grefenstette et al. “Neural Programmer-Interpreters” S. Reed, N. de Freitas. “Reinforcement Learning Turing Machine” W. Zaremba and I. Sutskever. “Learning Simple Algorithms from Examples” W. Zaremba, T. Mikolov, A. Joulin, R. Fergus. Also at RAM “The Neural GPU and the Neural RAM machine” I. Sutskever. “Structured Memory for Neural Turing Machines” W. Zhang, Y. Yu, B. Zhou. “Evolving Neural Turing Machines” R. Greve, E. Jacobsen, S. Risi.

Reasoning with Synthetic Language New ways of performing/evaluating ML for AI “A Roadmap towards Machine Intelligence” T. Mikolov, A. Joulin, M. Baroni. “Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks” J. Weston, A. Bordes, S. Chopra, A.. Rush, B. van Merriënboer, A. Joulin, T. Mikolov. New models attempt to solve bAbI tasks, some at RAM “End-To-End Memory Networks” S. Sukhbaatar, A. Szlam, J. Weston, R. Fergus. “Towards Neural Network-based Reasoning” B. Peng, Z. Lu, H. Li, K. Wong. “Dynamic Memory Networks for Natural Language Processing” A. Kumar, O. Irsoy, P. Ondruska, M. Iyyer, J. Bradbury, I. Gulrajani, R. Socher. Also at RAM “Considerations for Evaluating Models of Language Understanding and Reasoning” G. Recchia. “Chess Q&A : Question Answering on Chess Games” V. Cirik, L. Morency, E. Hovy.

New NLP Datasets for RAM Understanding text (news, children’s books) “Teaching Machines to Read and Comprehend” K. Hermann, T. Kočiský, E. Grefenstette, L. Espeholt, W. Kay, M. Suleyman, P. Blunsom. “The Goldilocks Principle: Reading Children's Books with Explicit Memory Representations” F. Hill, A. Bordes, S. Chopra, J. Weston. Conducting Dialog “Hierarchical Neural Network Generative Models for Movie Dialogues” I. Serban, A. Sordoni, Y. Bengio, A. Courville, J. Pineau. “A Neural Network Approach to Context-Sensitive Generation of Conversational Responses” Sordoni et al. “Neural Responding Machine for Short-Text Conversation” L. Shang, Z. Lu, H.Li. “Evaluating Prerequisite Qualities for Learning End-to-End Dialog Systems” J. Dodge, A. Gane, X. Zhang, A. Bordes, S. Chopra, A. Miller, A. Szlam, J. Weston. General Question Answering “Large-scale Simple Question Answering with Memory Networks” A. Bordes, N. Usunier, S. Chopra, J. Weston.

Classic NLP Tasks for RAM Classic Language Modeling “Long short-term memory” S. Hochreiter, J. Schmidhuber. “Learning Longer Memory in Recurrent Neural Networks” T. Mikolov, A. Joulin, S. Chopra, M. Mathieu, M. Ranzato. Machine translation “Sequence to Sequence Learning with Neural Networks” I. Sutskever, O. Vinyals, Q. Le. “Neural Machine Translation by Jointly Learning to Align and Translate” D. Bahdanau, K. Cho, Y. Bengio. Parsing “Grammar as a Foreign Language” O. Vinyals, L. Kaiser, T. Koo, S. Petrov, I. Sutskever, G. Hinton. Entailment “Reasoning about Entailment with Neural Attention” T. Rocktäschel, E. Grefenstette, K. Hermann, T. Kočiský, P. Blunsom. Summarization “A Neural Attention Model for Abstractive Sentence Summarization” A. M. Rush, S. Chopra, J. Weston.

Vision Tasks for RAM Image generation/classification “Learning to Combine Foveal Glimpses with a Third-Order Boltzmann Machine” H. Larochelle and G. Hinton. “Learning Where to Attend with Deep Architectures for Image Tracking” Denil et. al. “Recurrent Models of Visual Attention” V. Mnih, N. Hees, A. Graves and K. Kavukcuoglu. “Draw: A Recurrent Neural Network for Image Generation” K. Gregor, I. Danihelka, A. Graves, D.J. Rezende, D. Wierstra. “Generating Sequences with Recurrent Neural Networks” Alex Graves. arXiv preprint, 2013. Mapping images to text and reverse “Show, Attend and Tell: Neural Image Caption Generation with Visual Attention” K. Xu, J. Ba, R. Kiros, K. Cho, A. Courville, R. Salakhutdinov, R. Zemel, Y. Bengio. “Generating Images from Captions with Attention” E. Mansimov, E. Parisotto, J. Ba, R. Salakhutdinov.

RAM Issues Memory structure and addressing Types of memory when they should be used, and how can they be learnt? How to do fast retrieval of relevant knowledge when the scale is huge? How to build hierarchical memories, e.g. multi-scale attention? How to build hierarchical reasoning, e.g. composition of functions? Knowledge representation and handling How to decide what to write and what not to write in the memory? How to represent knowledge to be stored in memories? How to incorporate forgetting/compression of information? AI evaluation How to evaluate reasoning models? Are artificial tasks a good way? Where do real tasks are needed? Can we draw inspiration from how animal or human memories work?

OK, let’s RAM on! Real-time questions & updates: http://fb.me/ram