Neural Modular Networks

Slides:



Advertisements
Similar presentations
Level 1 Recall Recall of a fact, information, or procedure. Level 2 Skill/Concept Use information or conceptual knowledge, two or more steps, etc. Level.
Advertisements

How computers answer questions An introduction to machine learning Peter Barnum August 7, 2008.
1 Agenda: 09/01/2011 Finish preparing information visualization methods. Tables, diagrams, charts, bullet points Words vs. pictures vs. numbers Visualization.
PDDL: A Language with a Purpose? Lee McCluskey Department of Computing and Mathematical Sciences, The University of Huddersfield.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
Artificial Intelligence
Copyright R. Weber INFO 629 Concepts in Artificial Intelligence Fall 2004 Professor: Dr. Rosina Weber.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
11/2008AAAI Circuit sharing and the implementation of intelligent systems Michael L. Anderson Institute for Advanced Computer Studies Program in.
11 C H A P T E R Artificial Intelligence and Expert Systems.
Markov Logic and Deep Networks Pedro Domingos Dept. of Computer Science & Eng. University of Washington.
I Robot.
Submodule construction in logics 1 Gregor v. Bochmann, University of Ottawa Using First-Order Logic to Reason about Submodule Construction Gregor v. Bochmann.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Chapter 6 – Architectural Design Lecture 1 1Chapter 6 Architectural design.
Finite State Machines (FSM) OR Finite State Automation (FSA) - are models of the behaviors of a system or a complex object, with a limited number of defined.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Learning to Answer Questions from Image Using Convolutional Neural Network Lin Ma, Zhengdong Lu, and Hang Li Huawei Noah’s Ark Lab, Hong Kong
Artificial Intelligence
Ensembling Diverse Approaches to Question Answering
Jonatas Wehrmann, Willian Becker, Henry E. L. Cagnini, and Rodrigo C
Neural Machine Translation
Support for Program Analysis as a First-Class Design Constraint in Legion Michael Bauer 02/22/17.
End-To-End Memory Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Deep Learning Amin Sobhani.
PART IV: The Potential of Algorithmic Machines.
An Artificial Intelligence Approach to Precision Oncology
Recursive Neural Networks
Functions and Top-Down Design
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Intelligent Information System Lab
Are End-to-end Systems the Ultimate Solutions for NLP?
Deep learning and applications to Natural language processing
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
Dynamic Routing Using Inter Capsule Routing Protocol Between Capsules
Image Question Answering
Efficient Image Classification on Vertically Decomposed Data
Implementing Language Extensions with Model Transformations
Recurrent Neural Networks
TA : Mubarakah Otbi, Duaa al Ofi , Huda al Hakami
The Bias Variance Tradeoff and Regularization
Lecture: Deep Convolutional Neural Networks
1.5 Linear Inequalities.
Deep Cross-media Knowledge Transfer
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Natural Language to SQL(nl2sql)
Implementation support
Learning linguistic structure with simple recurrent neural networks
Chapter 5 Architectural Design.
Logistics Project proposals are due by midnight tonight (two pages)
Implementing Language Extensions with Model Transformations
Explainable Machine Learning Joseph E. Gonzalez
Course Summary Joseph E. Gonzalez
Natural Language Processing (NLP) Systems Joseph E. Gonzalez
Attention for translation
Model Compression Joseph E. Gonzalez
Dynamic Neural Networks Joseph E. Gonzalez
Automatic Handwriting Generation
Human-object interaction
Presented by: Anurag Paul
Building Dynamic Knowledge Graphs From Text Using Machine Reading Comprehension Rajarshi Das, Tsendsuren Munkhdalai, Xingdi Yuan, Adam Trischler, Andrew.
1MIT CSAIL 2Tsinghua University 3MIT-IBM Watson AI Lab 4DeepMind
Implementation support
Visual Grounding.
ICCV 2019.
CVPR 2019 Poster.
Presentation transcript:

Neural Modular Networks Joseph E. Gonzalez Co-director of the RISE Lab jegonzal@cs.berkeley.edu

Today

What Problem is being solved? Problem Domain: Visual Question Answering “Visual Turing test”

Prior State of the Art Semantic parsing and logic: Dependent on pre-trained computer vision models to populate database Jointly Learning to Parse and Perceive: Connecting Natural Language to the Physical World

Prior State of the Art Deep Embeddings Learned end-to-end but image representation is independent of the question. Image Question Answering: A Visual Semantic Embedding Model and a new Dataset Are you Talking to a Machine? Datasets and Methods for Multilingual Image Question Answering

Proposed Solution Compose a neural network to answer question Use parser to extract logical expression from question. Reuse network components across problems

Is there a circle next to a square? Question: Is there a circle next to a square? Logical Expression: is(circle, next-to(square)) Objective: Convert question into logical expression. Conceptually  Inducing a program from a question Also probably the more brittle part of the work Addressed in follow-up paper Alternative solution: user writes logical expression  programming

Neural Modules “Learned Sub-routines/Functions” Separate weights for each argument e.g., [dog]

Composition! “What color is his tie?” Separate weights for each argument e.g., [dog] Composition! “Learned programs” “What color is his tie?”

Composition! “What color is his tie?” Separate weights for each argument e.g., [dog] Composition! “Learned programs” “What color is his tie?” “Is there a red shape above a circle?” color(tie)

Training Train multiple graphs at once with shared modules. Separate weights for each argument e.g., [dog] Training Train multiple graphs at once with shared modules. “What color is his tie?” Individual models learn through their composition. color(tie) No pre-training

Evaluation Metrics and Results Accuracy on VQA benchmarks Existing benchmarks only require limited reasoning… Introduce new Shapes Benchmark Shapes Benchmark VQA Benchmark

Qualitative Results

Impact Over 300 citations (pretty good) Follow-up work “Learning to Reason: End-to-End Module Networks for Visual Question Answering” address limitations of parsing. Uses Policy RNN to predict composition (trained using RL) CLEVR dataset

Points to a bigger opportunity… Composition of learned modules Conjecture: Increasing “non-experts” will compose existing ML models to solve new complex problems. Organizations will develop and reuse model components in multiple tasks Training will span many different neural module programs Needed? Abstractions for individual components Mechanisms for composition and joint training