Classification of Hand-Written Digits Using Scattering Convolutional Network Dongmian Zou Advisor: Professor Radu Balan.

Slides:



Advertisements
Similar presentations
OpenCV Introduction Hang Xiao Oct 26, History  1999 Jan : lanched by Intel, real time machine vision library for UI, optimized code for intel 
Advertisements

Neural networks Introduction Fitting neural networks
Advanced topics.
Limin Wang, Yu Qiao, and Xiaoou Tang
ImageNet Classification with Deep Convolutional Neural Networks
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
CS771 Machine Learning : Tools, Techniques & Application Gaurav Krishna Y Harshit Maheshwari Pulkit Jain Sayantan Marik
Neural Networks I CMPUT 466/551 Nilanjan Ray. Outline Projection Pursuit Regression Neural Network –Background –Vanilla Neural Networks –Back-propagation.
Adaptive Offset Subspace Self- Organizing Map with an Application to Handwritten Digit Recognition Huicheng Zheng, Pádraig Cunningham and Alexey Tsymbal.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
What is the Best Multi-Stage Architecture for Object Recognition Kevin Jarrett, Koray Kavukcuoglu, Marc’ Aurelio Ranzato and Yann LeCun Presented by Lingbo.
Presented by: Kamakhaya Argulewar Guided by: Prof. Shweta V. Jain
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
An Example of Course Project Face Identification.
Yang, Luyu.  Postal service for sorting mails by the postal code written on the envelop  Bank system for processing checks by reading the amount of.
Handwritten Hindi Numerals Recognition Kritika Singh Akarshan Sarkar Mentor- Prof. Amitabha Mukerjee.
Object detection, deep learning, and R-CNNs
Deep Convolutional Nets
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 6: Applying backpropagation to shape recognition Geoffrey Hinton.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Yann LeCun Other Methods and Applications of Deep Learning Yann Le Cun The Courant Institute of Mathematical Sciences New York University
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Deep Learning and Its Application to Signal and Image Processing and Analysis Class III - Fall 2016 Tammy Riklin Raviv, Electrical and Computer Engineering.
Neural networks and support vector machines
Big data classification using neural network
Recognition of biological cells – development
Deep Learning Amin Sobhani.
Data Mining, Neural Network and Genetic Programming
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
Efficient Image Classification on Vertically Decomposed Data
Article Review Todd Hricik.
Applications of Deep Learning and how to get started with implementation of deep learning Presentation By : Manaswi Advisor : Dr.Chinmay.
Matt Gormley Lecture 16 October 24, 2016
Restricted Boltzmann Machines for Classification
Presenter: Chu-Song Chen
Intelligent Information System Lab
Neural networks (3) Regularization Autoencoder
ECE 6504 Deep Learning for Perception
Deep learning and applications to Natural language processing
Training Techniques for Deep Neural Networks
Object Recognition in the Dynamic Link Architecture
Understanding the Difficulty of Training Deep Feedforward Neural Networks Qiyue Wang Oct 27, 2017.
A brief introduction to neural network
Efficient Image Classification on Vertically Decomposed Data
Introduction to Neural Networks
Goodfellow: Chap 6 Deep Feedforward Networks
Image Classification.
Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no.
network of simple neuron-like computing elements
Convolutional neural networks Abin - Roozgard.
[Figure taken from googleblog
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Image to Image Translation using GANs
SVM-based Deep Stacking Networks
Representation Learning with Deep Auto-Encoder
Machine Learning – Neural Networks David Fenyő
Artificial Intelligence 10. Neural Networks
实习生汇报 ——北邮 张安迪.
Neural networks (3) Regularization Autoencoder
CSE 802. Prepared by Martin Law
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Automatic Handwriting Generation
Introduction to Neural Networks
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
Learning and Memorization
Deep learning: Recurrent Neural Networks CV192
Principles of Back-Propagation
An introduction to neural network and machine learning
Presentation transcript:

Classification of Hand-Written Digits Using Scattering Convolutional Network Dongmian Zou Advisor: Professor Radu Balan

Background

Overview Image classification Feature extractor Hand-written digits Feature extractor Convolutional neural network Machine learning techniques Part of MNIST training data retrieved from http://www.cs.nyu.edu/~roweis/data.html

Convolutional Network (ConvNet) Multi-layer structure Different layers extract different features (edges, contours, parts, etc.) Input layer Hidden layer Hidden layer Hidden layer Output layer

Convolutional Network (ConvNet) Each convolution layer contains three stages Convolution stage Detector stage Pooling stage convolution with filters nonlinearity applied local max or averaging

Scattering Convolutional Network Convolution stage Detector stage Pooling stage dilation of wavelets low-pass filter

Scattering Convolutional Network Scattering propagator Scattering transform Input Output

Scattering Convolutional Network S. Mallat proved that the scattering network is a good representation for image classification invariant to translation (approximately) stable to small-scale deformation Mallat’s theory requires a strict condition on the wavelet T. Wiatowski and H. Bölcskei give a relaxed condition See below for reference S. Mallat, Group Invariant Scattering. T. Wiatowski and H. Bölcskei, Deep Convolutional Neural Networks Based on Semi-Discrete Frames.

Machine Learning Techniques Gradient descent find the minimum of a loss function iterative method

Machine Learning Techniques Error back-propagation based on the chain rule of taking derivatives

Machine Learning Techniques Support vector machine (SVM) two-class classification problem {+1, -1} determined by the sign of

Approach

Network Structure Scattering network SVM One input layer Two convolution layer One output layer SVM

images of 28-by-28 pixels SVM classification result determined by the sign of

The Optimization Problem Our optimization problem for training the model is where hinge-loss and is the grouping of the following where fixed wavelet

images of 28-by-28 pixels Input: pairs of labels (which class x belongs to) Unknown parameters: SVM L is down-sampling factor g is the low-pass filter use cross validation for both

Two step optimization The above problem is non-convex difficult to solve with respect to {λ, w, b} We can iterate between two problems First, fix w and b, train the filters in the scattering network Second, fix the filters in the scattering network, train the SVM We use libSVM library to train the SVM Publicly available at https://www.csie.ntu.edu.tw/~cjlin/libsvm/ easier convex

Convolution layer Filters to train: Gradient descent Each filter contains two parameters Gradient descent starting value: randomly generated; distant parameters in the same layer learning rate: may change with step back propagation

Project Details

Implementation Personal Laptop MATLAB R2015b CPU: 2GHz Intel Core i7 Memory: 8 GB 1600 MHz DDR3 OS X El Capitan Version 10.11 MATLAB R2015b

Database MNIST database Publicly available at http://yann.lecun.com/exdb/mnist/ Image: 28×28 pixels (each pixels value between 0~255) 60,000 training examples 10,000 testing examples

Validation and Testing Cross-validation MatConvNet toolbox publicly available at http://www.vlfeat.org/matconvnet/ Testing Testing examples in MNIST database Measure: percentage of errors Compare with libSVM results

Timeline Sept - Oct: define the project, design the algorithm Nov: write codes for network training Dec: write codes for 2-class classification Jan - Feb: complete validation Mar: complete testing Apr: wrap up the project 2015 Multi-class if time permits 2016

Deliverables Datasets Toolboxes MATLAB codes Trained network Results Proposal, Reports, Presentation slides, etc.

Bibliography [1] Y. Bengio, I. J. Goodfellow and A. Courville, Deep Learning, Book in preparation for MIT press, 2015. [2] C. M. Bishop, Pattern Recognition and Machine Learning, New York: Springer, 2006. [3] J. Bruna and S. Mallat, Invariant Scattering Convolution Networks, Pattern Analysis and Machine Intelligence, IEEE Transactions 35(8)(2013), 1872-1886. [4] S. Mallat, Group Invariant Scattering, Comm. Pure Appl. Math., 65 (2012): 1331-1398. [5] A. Krizhevsky, I. Sutskever, G. E. Hinton, ImageNet Classification with Deep Convolutional Neural Networks, Advances in Neural Information Processing Systems 2 (2012): 1097-1105. [6] C. Szegedy et al, Going Deeper with Convolutions, Open Access Version available at http://www.cv- foundation.org/openaccess/content_cvpr_2015/papers/Szegedy_Going_Deeper_With_2015_CVPR_paper.pdf [7] A. Vedaldi and K. Lenc, MatConvNet - Convolutional Neural Networks for MATLAB, Proc. of the ACM Int. Conf. on Multimedia, 2015. [8] T. Wiatowski and H. Bölcskei, Deep Convolutional Neural Networks Based on Semi-Discrete Frames, Proc. of IEEE International Symposium on Information Theory (ISIT), Hong Kong, China, pp. 1212-1216, June 2015.

THANK YOU