Spring Courses CSCI 5922 – Probabilistic Models

Slides:



Advertisements
Similar presentations
Bioinspired Computing Lecture 16
Advertisements

Deep Learning: Back To The Future. Hinton NIPS 2012 Talk Slide (More Or Less) What was hot in 1987 Neural networks What happened in ML since 1987 Computers.
Deep Learning Bing-Chen Tsai 1/21.
Stochastic Neural Networks Deep Learning and Neural Nets Spring 2015.
CS590M 2008 Fall: Paper Presentation
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
Tuomas Sandholm Carnegie Mellon University Computer Science Department
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Deep Belief Networks for Spam Filtering
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Using Fast Weights to Improve Persistent Contrastive Divergence Tijmen Tieleman Geoffrey Hinton Department of Computer Science, University of Toronto ICML.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
The Boltzmann Machine Psych 419/719 March 1, 2001.
Geoffrey Hinton CSC2535: 2013 Lecture 5 Deep Boltzmann Machines.
CSC2535 Lecture 4 Boltzmann Machines, Sigmoid Belief Nets and Gibbs sampling Geoffrey Hinton.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 18 Learning Boltzmann Machines Geoffrey Hinton.
Constraint Satisfaction and Schemata Psych 205. Goodness of Network States and their Probabilities Goodness of a network state How networks maximize goodness.
CIAR Summer School Tutorial Lecture 1b Sigmoid Belief Nets Geoffrey Hinton.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
CSC321: Introduction to Neural Networks and Machine Learning Lecture 19: Learning Restricted Boltzmann Machines Geoffrey Hinton.
Boltzman Machines Stochastic Hopfield Machines Lectures 11e 1.
Introduction to Deep Learning
Chapter 6 Neural Network.
Deep learning Tsai bing-chen 10/22.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
CSC321: Computation in Neural Networks Lecture 21: Stochastic Hopfield nets and simulated annealing Geoffrey Hinton.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
CSC2535: Computation in Neural Networks Lecture 8: Hopfield nets Geoffrey Hinton.
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
1 Restricted Boltzmann Machines and Applications Pattern Recognition (IC6304) [Presentation Date: ] [ Ph.D Candidate,
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Bayesian Neural Networks
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Learning Deep Generative Models by Ruslan Salakhutdinov
CSC321 Lecture 18: Hopfield nets and simulated annealing
Energy models and Deep Belief Networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
Matt Gormley Lecture 16 October 24, 2016
Restricted Boltzmann Machines for Classification
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Spring Courses CSCI 5922 – Probabilistic Models (Mozer) CSCI Mind Reading Machines (Sidney D’Mello) CSCI 7000 – Human Centered Machine Learning.
Multimodal Learning with Deep Boltzmann Machines
Deep Learning Qing LU, Siyuan CAO.
Deep Belief Networks Psychology 209 February 22, 2013.
Structure learning with deep autoencoders
Restricted Boltzman Machines
CSSE463: Image Recognition Day 17
CSC 578 Neural Networks and Deep Learning
ECE 471/571 - Lecture 17 Back Propagation.
Biological and Artificial Neuron
Biological and Artificial Neuron
Artificial Intelligence Methods
Deep Architectures for Artificial Intelligence
ECE 599/692 – Deep Learning Lecture 9 – Autoencoder (AE)
Biological and Artificial Neuron
CSSE463: Image Recognition Day 17
Neural Networks Geoff Hulten.
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 13
Boltzmann Machine (BM) (§6.4)
CSC321 Winter 2007 Lecture 21: Some Demonstrations of Restricted Boltzmann Machines Geoffrey Hinton.
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Toward a Great Class Project: Discussion of Stoianov & Zorzi’s Numerosity Model Psych 209 – 2019 Feb 14, 2019.
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Spring Courses CSCI 5922 – Probabilistic Models CSCI 7000 -- Mind Reading Machines Sidney D’Mello W 2:00-4:30 MUEN D418  Can we teach computers how to read peoples’ thoughts and feelings to improve engagement, efficiency, and effectiveness? For example, can computers tell when their users are confused or frustrated, mentally zoned out, cognitively overloaded, or in an optimal mental state? This course will teach you how to make computers smarter and more human-like by reading their users’ minds (much like we do). In this interdisciplinary research-focused course, you will read, present, and discuss key papers in the fields of affective computing, attentional computing, augmented cognition, and multimodal interaction. You will also apply what you learned by developing your own research project. By the end of the course, you will be knowledgeable about the relevant theories and technologies, have conducted hands-on research in the area, and will have sharpened your critical thinking and scientific discourse skills.

CSCI 5922 Neural Networks and Deep Learning: Probabilistic Nets Mike Mozer Department of Computer Science and Institute of Cognitive Science University of Colorado at Boulder

Neural Net T-Shirts

Neural Net T-Shirts

Neural Net T-Shirts

Neural Net T-Shirts

Neural Net T-Shirts

A Brief History Of Deterministic And Stochastic Networks Boltzmann Machines / Harmony Nets 1985 1982 Hopfield Nets 1986 Back Propagation 1992 Sigmoid Belief Networks 2005 Restricted Boltzmann Machines and Deep Belief Nets 2009 Deep Learning With Back Propagation

Necker Cube

Selecting One Interpretation of Necker Cube See NeckerApp.jnlp

Hopfield Networks Binary-threshold units Asynchronous update Symmetric weights Solves an optimization problem minimize energy (or cost or potential) maximize harmony (or goodness-of-fit) search for parameters (activities) that produce the best solution

y h

Hopfield Net As Content Addressible Memory Won’t discuss training procedure because it’s dorky Hebbian learning Training on set of patterns causes them to become attractors Degraded input is mapped to nearest attractor

How a Boltzmann machine models data

Three Ways To Specify Inputs Use input to set initial activations bad idea: initial activations irrelevant once equilibrium is reached Use input to clamp or freeze unit activations clamped neurons effectively vanish from network and serve as bias on hidden neurons Use input to impose strong bias set bi such that unit i will (almost) always be off or on

Back To Thermal Equilibrium

statistical physics analogy

no need for back propagation Positive and negative phases positive phase clamp visible units set hidden randomly run to equilibrium for given T compute expectations <oioj>+ negative phase set visible and hidden randomly run to equilibrium for T=1 compute expectations <oioj>- didn’t scale up, computers weren’t fast enough -> people thought boltzmann machines would never be practical no need for back propagation

Why Boltzmann Machine Failed Too slow loop over training epochs loop over training examples loop over 2 phases (+ and -) loop over annealing schedule for T loop until thermal equilibrium reached loop to sample <oioj> Sensitivity to annealing schedule Difficulty determining when equilibrium is reached As learning progresses, weights get larger, energy barriers get hard to break -> becomes even slower Back prop was invented shortly after The need to perform pattern completion wasn’t necessary for most problems (feedforward nets sufficed)

Comments On Boltzmann Machine Learning Algorithm No need for back propagation reaching thermal equilibrium involves propagating information through network Positive and negative phase positive phase clamp visible units set hidden randomly run to equilibrium for T=1 compute expectations <oioj>+ negative phase set visible and hidden randomly run to equilibrium for T=1 compute expectations <oioj>- Why Boltzmann machine failed (circa 1985)

Restricted Boltzmann Machine (also known as Harmony Network) Architecture Why positive phase is trivial Contrastive divergence algorithm Example of RBM learning gibbs sampling / what if inputs are continuous: then make input units linear with gaussian noise

RBM Generative Model As A Product Of Experts

Deep RBM Autoencoder You’ve seen this before Guarantee that stacking improves Hinton & Salakhutdinov (2006)

Reminder (From Unsupervised Learning Lecture)

USED IN DEEP NETS LECTURE

Deep Belief Nets (DBNs): Using Stacked RBMs As A Generative Model H1 H2 H3 H3 H2 H1 V Generative model is not a Boltzmann machine Why do we need symmetric connections between H2 and H3?

Using A DBN For Supervised Learning 1. Train RBMs in unsupervised fashion 2. In final RBM, include additional units representing class labels 3a. Recognition model Use feedforward weights and fine tune with back prop 3b. Generative model Alternating Gibbs sampling between H3 and H4, and feedback weights elsewhere H1 H2 H3 V H1 H2 H4 L H3 H2 H3 V H1 L H4 H2 H3 V H1 L H4

Performance on MNIST (Hinton, Osindero, & Teh, 2006) generative model recognition model

Suskever, Martens, Hinton (2011) Generating Text From A Deep Belief Net Wikipedia The meaning of life is the tradition of the ancient human reproduction: it is less favorable to the good boy for when to remove her bigger. In the show’s agreement unanimously resurfaced. The wild pasteured with consistent street forests were incorporated by the 15th century BE. In 1996 the primary rapford undergoes an effort that the reserve conditioning, written into Jewish cities, sleepers to incorporate the .St Eurasia that activates the population. Mar??a Nationale, Kelli, Zedlat-Dukastoe, Florendon, Ptu’s thought is. To adapt in most parts of North America, the dynamic fairy Dan please believes, the free speech are much related to the NYT while he was giving attention to the second advantage of school building a 2-for-2 stool killed by the Cultures saddled with a half- suit defending the Bharatiya Fernall ’s office . Ms . Claire Parters will also have a history temple for him to raise jobs until naked Prodiena to paint baseball partners , provided people to ride both of Manhattan in 1978 , but what was largely directed to China in 1946 , focusing on the trademark period is the sailboat yesterday and comments on whom they obtain overheard within the 120th anniversary , where many civil rights defined , officials said early that forms , ” said Bernard J. Marco Jr. of Pennsylvania , was monitoring New York