Graph Neural Networks Amog Kamsetty January 30, 2019.

Slides:



Advertisements
Similar presentations
Negative Selection Algorithms at GECCO /22/2005.
Advertisements

A brief review of non-neural-network approaches to deep learning
Neural networks Introduction Fitting neural networks
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Artificial Neural Networks
Spatial Pyramid Pooling in Deep Convolutional
The Database and Info. Systems Lab. University of Illinois at Urbana-Champaign User Profiling in Ego-network: Co-profiling Attributes and Relationships.
M. Wang, T. Xiao, J. Li, J. Zhang, C. Hong, & Z. Zhang (2014)
Chapter 8: Adaptive Networks
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
Predicting the dropouts rate of online course using LSTM method
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
R-NET: Machine Reading Comprehension With Self-Matching Networks
Convolutional Sequence to Sequence Learning
SUNY Korea BioData Mining Lab - Journal Review
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 2, 2017.
Convolutional Neural Network
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
Artificial Neural Networks
Deep Learning Amin Sobhani.
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
Data Mining, Neural Network and Genetic Programming
Chilimbi, et al. (2014) Microsoft Research
Data Mining, Neural Network and Genetic Programming
Recursive Neural Networks
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Matt Gormley Lecture 16 October 24, 2016
Deep Learning: Model Summary
Spring Courses CSCI 5922 – Probabilistic Models (Mozer) CSCI Mind Reading Machines (Sidney D’Mello) CSCI 7000 – Human Centered Machine Learning.
Intelligent Information System Lab
Neural networks (3) Regularization Autoencoder
Supervised Training of Deep Networks
Neural Networks 2 CS446 Machine Learning.
Convolution Neural Networks
Convolutional Networks
Master’s Thesis defense Ming Du Advisor: Dr. Yi Shang
RNNs: Going Beyond the SRN in Language Prediction
Goodfellow: Chap 6 Deep Feedforward Networks
A First Look at Music Composition using LSTM Recurrent Neural Networks
Deep learning Introduction Classes of Deep Learning Networks
CSC 578 Neural Networks and Deep Learning
MEgo2Vec: Embedding Matched Ego Networks for User Alignment Across Social Networks Jing Zhang+, Bo Chen+, Xianming Wang+, Fengmei Jin+, Hong Chen+, Cuiping.
Neural Networks Geoff Hulten.
Other Classification Models: Recurrent Neural Network (RNN)
SVM-based Deep Stacking Networks
Lecture 16: Recurrent Neural Networks (RNNs)
Neural networks (1) Traditional multi-layer perceptrons
Attention.
实习生汇报 ——北邮 张安迪.
Event Graph Neural Network for LHC Stop Search
Neural networks (3) Regularization Autoencoder
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Attention for translation
Human-object interaction
Introduction to Neural Networks
A unified extension of lstm to deep network
Keshav Balasubramanian
Deep learning enhanced Markov State Models (MSMs)
Rgh
Neural Machine Translation using CNN
Image recognition.
Heterogeneous Graph Convolutional Network
2019/9/14 The Deep Learning Vision for Heterogeneous Network Traffic Control Proposal, Challenges, and Future Perspective Author: Nei Kato, Zubair Md.
Deep learning: Recurrent Neural Networks CV192
Peng Cui Tsinghua University
Bidirectional LSTM-CRF Models for Sequence Tagging
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Graph Neural Networks Amog Kamsetty January 30, 2019

Motivations

Traditional Deep Learning Slide from Thomas Kipf

Graph Structured Data A lot of real-world data does not live on “grids” Knowledge Graphs Social Networks Citation Networks Communication Networks Multi-Agent Systems Protein Interaction Networks Node Classification, Graph classification, Clustering, Link predictions CNNs and RNNs can’t handle graph inputs because they rely on a specific ordering Slide from Thomas Kipf

Inspiration from CNNs Advantages of CNNs Local Connections Shared Weights Use of multiple Layers But, hard to define convolutional and pooling layers for non-Euclidean data

Deep Learning on Graphs Need architecture for machine learning on graph structured data Also need to take advantage of the structural relationships in graphs, similar to CNNs Collect information via message passing and then aggregation “Graph neural networks (GNNs) are connectionist models that capture the dependence of graphs via message passing between the nodes of graphs.” Input is a graph, and output is a graph GNNs keep track of a state unlike standard neural network Image from Thomas Kipf

Types of Graph Networks

The original Graph Neural Network (GNN) Each node is defined by its own features and those of its neighbors Learn some state embedding for each node Scarselli et al., 2009

The original Graph Neural Network (GNN) Update equation if parametrized by MLP

Inductive Capability Slide from Stanford Snap Lab

Limitations of GNN Very computationally intensive to recursively compute fixed point solution Same parameters are used in every timestep, while other neural networks use different parameters in each layer Informative edge features are difficult to model Can’t alter message propagation depending on edge type These pitfalls led to many variations of GNNs, particularly in how the propagation occurs

Graph Convolutional Networks Kipf & Welling, 2017 Slide from Thomas Kipf

Graph Convolutional Networks Slide from Stanford Snap Lab

Graph Convolutional Networks Semi-supervised learning on Zachary’s Karate Club Network Only one node from each class is labeled 300 iterations Video from Thomas Kipf

Graph Convolutional Networks

Gated Graph Neural Networks Uses LSTM for aggregation Allows for deep graph networks without worrying about overfitting or vanishing/exploding gradient Li et al., 2016 Slide from Stanford Snap Lab

Many more variants...perhaps too many

Generalizations Message Passing Neural Networks (MPNN) Unifies GNN and GCN Message Passing Phase that constructs message based on local neighborhood Update function which interprets message and updates node’s hidden state Readout phase computes feature vector for the whole graph All of these functions can have different settings Non-local Neural Networks (NLNN) Unification of self-attention style methods Node update based on neighbor nodes, not edges

Graph Nets

Relational Inductive Biases Battaglia et al., 2018 Combinatorial Generalization New inferences, predictions, and behaviors from known building blocks Relies on humans’ mechanism for representing structure and reasoning about relations But, modern deep Learning relies on end-to-end design and does away with any explicit structure Instead, advocates for use end-to-end and hand-engineering jointly Combination of “nature” and “nurture” Biases/constraints on structure and relations is necessary Structured representations have entities and relations (aka graphs) at their core Thus, argues for graph networks as foundational building block for AI

Relational Inductive Biases

Relational Inductive Biases Makes use of sharing and locality Authors say there

Graph Networks While RNNs and CNNs use relational inductive biases, they do not generalize to arbitrary relations Need a framework that can learn from graphs Proposes graph networks, which generalizes MPNN, NLNN, and other variants Graph is defined as a 3-tuple Where u is global attribute V is set of vertices with attributes E is set of edges with attributes

Graph Networks Attributes are updated in a sequence of updates and aggregations, as we’ve seen before Update functions can be parametrized with neural networks Same 6 functions are used for all nodes and edges

Graph Networks

Graph Networks 3 main properties 3 design principles Nodes and edges provide a strong relational bias Entities and relations are represented as sets and are thus order invariant Per-edge and per-node functions are shared across the entire network 3 design principles Flexible representations Configurable within block structure Composable multi-block architectures

Flexible Representations

Composable Multi-block Architectures

Impact and Challenges

Impact Graph networks have applications in many areas Works well for any data that can be represented as graphs Node/Graph classification Link prediction Clustering Use case in biology, chemistry, physical systems Can be used for non-structured data but difficult to generate graphs from raw data Whether the idea of relational inductive biases will be adopted remains to be seen

Non-structural scenarios Scalability Challenges Shallow Structure Dynamic Graphs Non-structural scenarios Scalability

References “Graph Neural Networks: A Review of Methods and Applications” Zhou et al. 2019 “Gated Graph Sequence Neural Networks” Li et al. 2017 “The Graph Neural Network Model” Scarselli et al. 2009 “Relational inductive biases, deep learning ,and graph networks” Battaglia et al. 2018 The morning paper blog, Adrian Coyler Structured Deep Models: Deep Learning on Graphs and Beyond, talk by Thomas Kipf “Convolutional Networks on Graphs for Learning Molecular Fingerprints” Duvenaud et al. 2015 “Semi-Supervised Classification with Graph Convolutional Networks” Kipf & Welling 2017 “Graph Convolutional Networks”, Kipf 2016 “Representation Learning on Graphs: Methods and Applications” Hamilton et al. 2017 “Geometric Deep Learning: Going Beyond Euclidean Data” Bronstein et al. 2017 “Gated Graph Sequence Neural Networks” Li et al., 2016