Comp 5013 Deep Learning Architectures Daniel L. Silver March, 2014 1.

Slides:



Advertisements
Similar presentations
Scalable Learning in Computer Vision
Advertisements

Deep Learning Bing-Chen Tsai 1/21.
CS590M 2008 Fall: Paper Presentation
Advanced topics.
Rajat Raina Honglak Lee, Roger Grosse Alexis Battle, Chaitanya Ekanadham, Helen Kwong, Benjamin Packer, Narut Sereewattanawoot Andrew Y. Ng Stanford University.
Nathan Wiebe, Ashish Kapoor and Krysta Svore Microsoft Research ASCR Workshop Washington DC Quantum Deep Learning.
Stacking RBMs and Auto-encoders for Deep Architectures References:[Bengio, 2009], [Vincent et al., 2008] 2011/03/03 강병곤.
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
Comp 3503 / 5013 Dynamic Neural Networks Daniel L. Silver March,
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
Recent Developments in Deep Learning Quoc V. Le Stanford University and Google.
Deep Learning.
How to do backpropagation in a brain
Deep Belief Networks for Spam Filtering
AN ANALYSIS OF SINGLE- LAYER NETWORKS IN UNSUPERVISED FEATURE LEARNING [1] Yani Chen 10/14/
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Submitted by:Supervised by: Ankit Bhutani Prof. Amitabha Mukerjee (Y )Prof. K S Venkatesh.
MACHINE LEARNING AND ARTIFICIAL NEURAL NETWORKS FOR FACE VERIFICATION
Lifelong Machine Learning and Reasoning
Nantes Machine Learning Meet-up 2 February 2015 Stefan Knerr CogniTalk
How to do backpropagation in a brain
Deep Learning for Speech and Language Yoshua Bengio, U. Montreal NIPS’2009 Workshop on Deep Learning for Speech Recognition and Related Applications December.
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
Intelligent Information Technology Research Lab, Acadia University, Canada 1 Lifelong Machine Learning and Reasoning Daniel L. Silver Acadia University,
A shallow introduction to Deep Learning
Large-scale Deep Unsupervised Learning using Graphics Processors
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Zhenbao Liu 1, Shaoguang Cheng 1, Shuhui Bu 1, Ke Li 2 1 Northwest Polytechnical University, Xi’an, China. 2 Information Engineering University, Zhengzhou,
Constructing Knowledge Graph from Unstructured Text Image Source: Kundan Kumar Siddhant Manocha.
LeCun, Bengio, And Hinton doi: /nature14539
M. Wang, T. Xiao, J. Li, J. Zhang, C. Hong, & Z. Zhang (2014)
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 6: Applying backpropagation to shape recognition Geoffrey Hinton.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Deep learning Tsai bing-chen 10/22.
Neural Networks William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]
CSC321 Lecture 24 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
CSC321 Lecture 27 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
語音訊號處理之初步實驗 NTU Speech Lab 指導教授: 李琳山 助教: 熊信寬
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Efficient Estimation of Word Representations in Vector Space By Tomas Mikolov, Kai Chen, Greg Corrado, Jeffrey Dean. Google Inc., Mountain View, CA. Published.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Chapter 13 Artificial Intelligence. Artificial Intelligence – Figure 13.1 The Turing Test.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Deep Learning and Its Application to Signal and Image Processing and Analysis Class III - Fall 2016 Tammy Riklin Raviv, Electrical and Computer Engineering.
Vision-inspired classification
Big data classification using neural network
The Relationship between Deep Learning and Brain Function
Supplementary readings:
Deep Learning Amin Sobhani.
ECE 5424: Introduction to Machine Learning
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
Goodfellow: Chap 1 Introduction
Deep Learning Insights and Open-ended Questions
Article Review Todd Hricik.
Restricted Boltzmann Machines for Classification
Deep Learning Fundamentals online Training at GoLogica
Deep learning and applications to Natural language processing
Unsupervised Learning and Autoencoders
Deep Learning Workshop
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Restricted Boltzman Machines
Inductive Transfer, Machine Lifelong Learning, and AGI
Goodfellow: Chap 1 Introduction
State-of-the-art face recognition systems
Basic Intro Tutorial on Machine Learning and Data Mining
Limitations of Traditional Deep Network Architectures
network of simple neuron-like computing elements
Deep Learning Some slides are from Prof. Andrew Ng of Stanford.
Automatic Handwriting Generation
Do Better ImageNet Models Transfer Better?
Presentation transcript:

Comp 5013 Deep Learning Architectures Daniel L. Silver March,

Y. Bengio - McGill 2009 Deep Learning Tutorial 2013 Deep Learning towards AI Deep Learning of Representations (Y. Bengio) –

Deep Belief RBM Networks with Geoff Hinton Learning layers of features by stacking RBMs – M M Discriminative fine-tuning in DBN – What happens during fine-tuning? – 3

Deep Belief RBM Networks with Geoff Hinton Learning handwritten digits – Modeling real-value data (G.Hinton) – 4

Deep Learning Architectures Consider the problem of trying to classify these hand-written digits.

Deep Learning Architectures 2000 top-level artificial neurons neurons (higher level features) 500 neurons (higher level features) 500 neurons (low level features) 500 neurons (low level features) Images of digits 0-9 (28 x 28 pixels) Images of digits 0-9 (28 x 28 pixels) Neural Network: - Trained on 40,000 examples - Learns: * labels / recognize images * generate images from labels - Probabilistic in nature - DemoDemo 2 3 1

Deep Convolution Networks Intro - tml#lenet tml#lenet

ML and Computing Power Andrew Ng’s work on Deep Learning Networks (ICML-2012) Problem: Learn to recognize human faces, cats, etc from unlabeled data Dataset of 10 million images; each image has 200x200 pixels 9-layered locally connected neural network (1B connections) Parallel algorithm; 1,000 machines (16,000 cores) for three days 8 Building High-level Features Using Large Scale Unsupervised Learning Quoc V. Le, Marc’Aurelio Ranzato, Rajat Monga, Matthieu Devin, Kai Chen, Greg S. Corrado, Jeffrey Dean, and Andrew Y. Ng ICML 2012: 29th International Conference on Machine Learning, Edinburgh, Scotland, June, 2012.

ML and Computing Power Results: A face detector that is 81.7% accurate Robust to translation, scaling, and rotation Further results: 15.8% accuracy in recognizing 20,000 object categories from ImageNet 70% relative improvement over the previous state-of-the-art. 9

Deep Belief Convolution Networks Deep Belief Convolution Network (Javascript) – Runs well under Google Chrome –

Google and DLA FI FI 026/is-google-cornering-the-market-on-deep- learning/ 026/is-google-cornering-the-market-on-deep- learning/

Cloud-Based ML - Google 12

Additional References Coursera course – Neural Networks fro Machine Learning: – 001/lecture 001/lecture ML: Hottest Tech Trend in next 3-5 Years – Geoff Hinton’s homepage –

Open Questions in ML

Challenges & Open Questions Stability-Plasticity problem - How do we integrate new knowledge in with old? No loss of new knowledge No loss or prior knowledge Efficient methods of storage and recall ML methods that can retain learned knowledge will be approaches to “common knowledge” representation – a “Big AI” problem 15

Challenges & Open Questions Practice makes perfect ! –An LML system must be capable of learning from examples of tasks over a lifetime –Practice should increase model accuracy and overall domain knowledge –How can this be done? –Research important to AI, Psych, and Education 16

Challenges & Open Questions Scalability –Often a difficult but important challenge –Must scale with increasing: Number of inputs and outputs Number of training examples Number of tasks Complexity of tasks, size of hypothesis representation –Preferably, linear growth 17

Never-Ending Language Learner Carlson et al (2010) Each day: Extracts information from the web to populate a growing knowledge base of language semantics Learns to perform this task better than on previous day –Uses a MTL approach in which a large number of different semantic functions are trained together 18