Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.

Slides:



Advertisements
Similar presentations
NEURAL NETWORKS Biological analogy
Advertisements

A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
Deep Learning Bing-Chen Tsai 1/21.
Artificial Neural Networks (1)
Advanced topics.
Lecture 5: CNN: Regularization
ImageNet Classification with Deep Convolutional Neural Networks
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
Lecture 14 – Neural Networks
An introduction to: Deep Learning aka or related to Deep Neural Networks Deep Structural Learning Deep Belief Networks etc,
Artificial Neural Networks ECE 398BD Instructor: Shobha Vasudevan.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Overview of Back Propagation Algorithm
ICS 273A UC Irvine Instructor: Max Welling Neural Networks.
Deep Learning and its applications to Speech EE 225D - Audio Signal Processing in Humans and Machines Oriol Vinyals UC Berkeley.
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Computer Science and Engineering
Cascade Correlation Architecture and Learning Algorithm for Neural Networks.
A shallow introduction to Deep Learning
NEURAL NETWORKS FOR DATA MINING
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Classification / Regression Neural Networks 2
Deep Convolutional Nets
Neural networks in modern image processing Petra Budíková DISA seminar,
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
Introduction to Deep Learning
Neural Network and Deep Learning 王强昌 MLA lab.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
ImageNet Classification with Deep Convolutional Neural Networks Presenter: Weicong Chen.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Convolutional Neural Network
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Big data classification using neural network
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Convolutional Neural Network
The Relationship between Deep Learning and Brain Function
Deep Learning Amin Sobhani.
an introduction to: Deep Learning
Data Mining, Neural Network and Genetic Programming
Data Mining, Neural Network and Genetic Programming
Computer Science and Engineering, Seoul National University
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
Deep Learning Insights and Open-ended Questions
Matt Gormley Lecture 16 October 24, 2016
Intelligent Information System Lab
Neural networks (3) Regularization Autoencoder
ECE 6504 Deep Learning for Perception
Supervised Training of Deep Networks
Introduction to Deep Learning for neuronal data analyses
Introduction to Neural Networks
network of simple neuron-like computing elements
Object Classes Most recent work is at the object level We perceive the world in terms of objects, belonging to different classes. What are the differences.
Emre O. Neftci  iScience  Volume 5, Pages (July 2018) DOI: /j.isci
Neural Networks Geoff Hulten.
Lecture: Deep Convolutional Neural Networks
Deep Learning Some slides are from Prof. Andrew Ng of Stanford.
Neural networks (3) Regularization Autoencoder
Course Recap and What’s Next?
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
An introduction to: Deep Learning aka or related to Deep Neural Networks Deep Structural Learning Deep Belief Networks etc,
CS295: Modern Systems: Application Case Study Neural Network Accelerator Sang-Woo Jun Spring 2019 Many slides adapted from Hyoukjun Kwon‘s Gatech “Designing.
Principles of Back-Propagation
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Object Recognizing

Deep Learning Success in 2012 DeepNet and speech processing

David Corne, and Nick Taylor, Heriot-Watt University - These slides and related resources: ne/Teaching/dmml.html

ImageNet

DL is providing breakthrough results in speech recognition and image classification … From this Hinton et al 2012 paper: go here: From here:

Continuous improvement Micrososft Dec layers Error rate of 3.5% and a localization error of 9%.

10

What are Deep Nets

Neural networks in the brain Repeating layers Linear, non-linear, pooling Learning by modifying synapses

Biology: Linear and non-linear operations

Biology: Feed-forward, recurrent, feed-back DNN adopts the feed-forward path

DNN Architecture

General structure local connections, convolution, reduced sampling

Multiple filters

Repeating operations: Linear, Non-linear, Pooling

Depth – multiple filters

Repeating 3-layer arrangement

History Deep Learning

LeNet 1998 Essentially the same as the current generation

MNIST data set

Hinton Trends in Cognitive Science 2007 The goal: unsupervised Restricted Boltzmann Machines Combining generative model and inference CNN are feed-forward and massively supervised

Back-propagation 1986

The entire network is a large parametric function The parameters are the network weights (60M in AlexNet) The parameters are learned from example The learning algorithm: back-propagation Gradient descent in the space of parameters

Back Propagation

input hidden output

input hidden output N5N5 N6N6 w 13 w 35 L = linear signal L k = Σ w ik N i N = nonlinear output N = σ(L) σ: y = 1 / (1 + e -αx ) dy/dx = y(1-y) dN/dL = N(1-N) dL/dW = N N L

L = linear signal L k = Σ w ik N i N = nonlinear output N = σ(L) σ: y = 1 / (1 + e -αx ) dy/dx = y(1-y)

N5N5 N6N6 w 13 w 35 Error E = 1/2 [(T 5 – N 5 ) 2 + (T 6 – N 6 ) 2 ] dE/d w 35 = (chain rule along the path) dE/d N 5 * dN 5 /d L 5 * dL 5 /d w 35 (T 5 – N 5 )N 5 (1-N 5 ) N3N3 * * Call dE/d L k = δ k back-propagating error δw ik = δ k N i Adjust the weight: δ5δ5 L5L5

N5N5 N6N6 w 13 w 35 General rule: dE/d L k = δ k back-propagating error Adjusting weights: δw ik = δ k N i δ5δ5 N3N3

N5N5 N6N6 w 13 w 35 General rule: dE/d L k = δ k back-propagating error Adjusting weights: δw ik = δ k N i δ3δ3 N1N1 True for w 13 Compute δ 3

input hidden output N5N5 N6N6 N6N6 w 13 w 35

Compute δ 3 dE/dw 13 = dE/d L 3 dL 3 /d w 13 = δ 3 N 1 dE/dL 3 = dE 1 /dL 3 + dE 2 /dL 3 = δ 31 + δ 32 δ 31 = dE 1 /dL 3 = dE 1 /dN 3 * dN 3 /dL 3 dE 1 /dN 3 = dE 1 /dL 5 * dL 5 / dN 3 = δ 5 * w 35 δ 31 = δ 5 w 35 N 3 (1-N 3 ) δ 32 = δ 6 w 36 N 3 (1-N 3 ) Adjusting δw 13 δw 13 = (δ 5 w 35 + δ 6 w 36 ) N 3 (1-N 3 ) N 1 w 13 L3L3 N3N3 L5L5

δ 5 * w 35 Adjusting δw 13 δw 13 = (δ 5 w 35 + δ 6 w 36 ) N 3 (1-N 3 ) N 1 w 13 δ 6 * w 36 N 3 (1-N 3 ) δ3δ3 N1N1 Propagate δ 5 and δ 6 Multiply by N 3 (1-N 3 ) Get δ 3 Adjust w 13 by δ 3 N 1 Iterated for all weights over many examples Supervision is required

Dropout

Dropout: An efficient way to average many large neural nets ( Consider a neural net with one hidden layer. Each time we present a training example, we randomly omit each hidden unit with probability 0.5. So we are randomly sampling from 2^H different architectures. – All architectures share weights.

Dropout – Multi Layer For each example, set units at all levels to 0 with some probability, usually p = 0.5 Each example has a different ‘mask’ During feed-forward flow, these units are multiplied by 0, the do not participate in the computation. Similarly for the BP The intuition is to avoid over-fitting At test time all the units are used Most implementations no longer use dropout. The issue of overfitting is actively studied. For some reasons adding weights does not cause drop in test performance.

Visualizing the features at different layers Bob Fergus NIPS 2013 Best 9 patches: showing at each layer responses of 48 units. Each unit is in fact a layer of units – copies of the same unit it different locations, covering the image (a ‘convolution’ filter) They identify by a ‘deconvolution’ algorithm the patches that caused the largest activation of the unit, in a large set of test images. Showing in a 3*3 small array the 9 top-patches for each unit.

First layer in AlexNet

Layer 3 top-9 patches for each unit

Different visual tasks

Segmentation

Edge Detection

Captioning

Getting annotations mlhttp:// ml April 2015

A woman with brown hair is sitting on her head.

a man is standing in front woman in white shirt.

Some Future Recognition Challenges

Full object interpretation Headlight Window Door knob Back wheel Mirror Front wheel Headlight Window Bumper

Actions: What are these people doing?

Agents interactions Disagree: Hug: