CSCI 5922 Neural Networks and Deep Learning: NIPS Highlights

Slides:



Advertisements
Similar presentations
Optimizing and Learning for Super-resolution
Advertisements

Unsupervised Learning Clustering K-Means. Recall: Key Components of Intelligent Agents Representation Language: Graph, Bayes Nets, Linear functions Inference.
Partially Observable Markov Decision Process (POMDP)
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Classification Neural Networks 1
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
1 L-BFGS and Delayed Dynamical Systems Approach for Unconstrained Optimization Xiaohui XIE Supervisor: Dr. Hon Wah TAM.
Bayesian belief networks 2. PCA and ICA
Artificial Neural Networks
online convex optimization (with partial information)
Machine Learning Chapter 4. Artificial Neural Networks
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
BrainStorming 樊艳波 Outline Several papers on icml15 & cvpr15 PALM Information Theory Learning.
HAITHAM BOU AMMAR MAASTRICHT UNIVERSITY Transfer for Supervised Learning Tasks.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Sparse Coding: A Deep Learning using Unlabeled Data for High - Level Representation Dr.G.M.Nasira R. Vidya R. P. Jaia Priyankka.
Data Mining Practical Machine Learning Tools and Techniques
Big data classification using neural network
The role of optimization in machine learning
Generative Adversarial Network (GAN)
RNNs: An example applied to the prediction task
Generative Adversarial Imitation Learning
Deep Feedforward Networks
Environment Generation with GANs
CSC321 Lecture 18: Hopfield nets and simulated annealing
Principle Component Analysis (PCA) Networks (§ 5.8)
Generative Adversarial Networks
CSCI 5922 Neural Networks and Deep Learning Generative Adversarial Networks Mike Mozer Department of Computer Science and Institute of Cognitive Science.
Spring Courses CSCI 5922 – Probabilistic Models (Mozer) CSCI Mind Reading Machines (Sidney D’Mello) CSCI 7000 – Human Centered Machine Learning.
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Machine Learning for dotNET Developer Bahrudin Hrnjica, MVP
Basic machine learning background with Python scikit-learn
Machine Learning Basics
Dynamical Statistical Shape Priors for Level Set Based Tracking
CSCI 5822 Probabilistic Models of Human and Machine Learning
CSCI 5822 Probabilistic Models of Human and Machine Learning
RNNs: Going Beyond the SRN in Language Prediction
Hidden Markov Models Part 2: Algorithms
Classification Neural Networks 1
Probabilistic Models with Latent Variables
CSCI 5822 Probabilistic Models of Human and Machine Learning
Bayesian belief networks 2. PCA and ICA
CSCI 5822 Probabilistic Models of Human and Machine Learning
Overview of Machine Learning
RL for Large State Spaces: Value Function Approximation
David Healey BYU Capstone Course 15 Nov 2018
Neural Networks Geoff Hulten.
Overfitting and Underfitting
Lip movement Synthesis from Text
View Inter-Prediction GAN: Unsupervised Representation Learning for 3D Shapes by Learning Global Shape Memories to Support Local View Predictions 1,2 1.
Competitive Optimization Spectral Methods
Autoencoders hi shea autoencoders Sys-AI.
Artificial Intelligence 10. Neural Networks
Machine learning overview
Autoencoders Supervised learning uses explicit labels/correct output in order to train a network. E.g., classification of images. Unsupervised learning.
Compressive Image Recovery using Recurrent Generative Model
Batch Normalization.
Rohan Yadav and Charles Yuan (rohany) (chenhuiy)
Image recognition.
Cengizhan Can Phoebe de Nooijer
Self-Supervised Cross-View Action Synthesis
CSC 578 Neural Networks and Deep Learning
CS249: Neural Language Model
Rong Ge, Duke University
Logistic Regression Geoff Hulten.
Iterative Projection and Matching: Finding Structure-preserving Representatives and Its Application to Computer Vision.
Presentation transcript:

CSCI 5922 Neural Networks and Deep Learning: NIPS Highlights Mike Mozer Department of Computer Science and Institute of Cognitive Science University of Colorado at Boulder position audience

Y-W Teh – concrete VAE [discrete variables] Deep sets

Gradient Descent GAN Optimization is Locally Stable (Nagarajan & Kolter, 2017) Explores situation where generator and discriminator are trained simultaneously no alternation, inner/outer loops, running one to convergence, etc. This situation does not correspond to a convex-concave optimization problem (i.e., no saddle point) “Under suitable conditions on the representational powers of the discriminator and the generator, the resulting GAN dynamical system is locally exponentially stable.” gradient updates will converge to an equilibrium point at an exponential rate

Gradient Descent GAN Optimization is Locally Stable (Nagarajan & Kolter, 2017) Simple case with 𝑫 𝒙 = 𝒘 𝟐 𝒙 𝟐 𝑮 𝒛 =𝒂𝒛 Distributions 𝒙~𝐔𝐧𝐢𝐟𝐨𝐫𝐦( −𝟏,𝟏 𝟐 ) 𝒛~𝐔𝐧𝐢𝐟𝐨𝐫𝐦( −𝟏,𝟏 𝟐 ) 𝜼 is a regularizer of some sort regularizer considers the discriminator updates when updating the generator

Bayesian GAN (Saatchi & Wilson, 2017) Problem with GANs: mode collapse GAN memorizes a few examples to fool the generator GAN doesn’t reproduce full diversity of environment Traditional GAN is conditioned on a noise sample, 𝒛 instead, marginalize over 𝒛 to obtain iteratively e𝐬𝐭𝐢𝐦𝐚𝐭𝐞 𝒑 𝜽 𝒈 𝜽 𝒅 and 𝒑 𝜽 𝒅 𝜽 𝒈 with samples of 𝒛 and represent each distribution via a set of samples

Bayesian GAN (Saatchi & Wilson, 2017) PCA representation of output space (top 2 dimensions data GAN BGAN

Bayesian GAN https://www.youtube.com/watch?v=24A8tWs6aug&feature=youtu.be

Unsupervised Learning of Disentangled Representations from Video (Denton & Birodkar, 2017) their work previous work

Unsupervised Learning of Disentangled Representations from Video (Denton & Birodkar, 2017) interpolating between two views via linear interpolation in pose space

Unsupervised Learning of Disentangled Representations from Video (Denton & Birodkar, 2017) Key idea (which has been leveraged in a lot of work) in two successive frames of a video, we’re likely to see the same object(s) but in slightly different poses true whether camera is panning or objects are moving also true of an individual who is observing static scene while moving eyes

Unsupervised Learning of Disentangled Representations from Video (Denton & Birodkar, 2017) Reconstruction loss predict frame 𝒕+𝒌 from content at 𝒕 and pose at 𝒕+𝒌 Similarity loss content should not vary from 𝒕 to 𝒕+𝒌

Unsupervised Learning of Disentangled Representations from Video (Denton & Birodkar, 2017) Adversarial loss 1 good pose vectors are ones that can fool a discriminator trying to determine whether two samples are from same or different video Adversarial loss 2 good pose vectors are ones that don’t provide any information to the discriminator about whether two samples from the same video are from same or diff. video

Unsupervised Learning of Disentangled Representations from Video (Denton & Birodkar, 2017)

Gumbel distribution trick