Compression of CNNs Mooyeol Baek Xiangyu Zhang, Jianhua Zou, Xiang Ming, Kaiming He, Jian Sun: Efficient and Accurate Approximations of Nonlinear Convolutional.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Chapter 28 – Part II Matrix Operations. Gaussian elimination Gaussian elimination LU factorization LU factorization Gaussian elimination with partial.
Advanced data assimilation methods with evolving forecast error covariance Four-dimensional variational analysis (4D-Var) Shu-Chih Yang (with EK)
Spatial Pyramid Pooling in Deep Convolutional
Mining Discriminative Components With Low-Rank and Sparsity Constraints for Face Recognition Qiang Zhang, Baoxin Li Computer Science and Engineering Arizona.
Introduction to tensor, tensor factorization and its applications
1 Sparsity Control for Robust Principal Component Analysis Gonzalo Mateos and Georgios B. Giannakis ECE Department, University of Minnesota Acknowledgments:
SAND C 1/17 Coupled Matrix Factorizations using Optimization Daniel M. Dunlavy, Tamara G. Kolda, Evrim Acar Sandia National Laboratories SIAM Conference.
Instance Construction via Likelihood- Based Data Squashing Madigan D., Madigan D., et. al. (Ch 12, Instance selection and Construction for Data Mining.
Orthogonalization via Deflation By Achiya Dax Hydrological Service Jerusalem, Israel
Efficient computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm Anders Eriksson and Anton van den Hengel.
Research Vignette: The TransCom3 Time-Dependent Global CO 2 Flux Inversion … and More David F. Baker NCAR 12 July 2007 David F. Baker NCAR 12 July 2007.
Mingyang Zhu, Huaijiang Sun, Zhigang Deng Quaternion Space Sparse Decomposition for Motion Compression and Retrieval SCA 2012.
A Note on Rectangular Quotients By Achiya Dax Hydrological Service Jerusalem, Israel
Principal Manifolds and Probabilistic Subspaces for Visual Recognition Baback Moghaddam TPAMI, June John Galeotti Advanced Perception February 12,
The Viola/Jones Face Detector A “paradigmatic” method for real-time object detection Training is slow, but detection is very fast Key ideas Integral images.
CS434/534: Mobile Computing and Wireless Networks Y. Richard Yang 08/30/2012.
Hybrid Load Forecasting Method With Analysis of Temperature Sensitivities Authors: Kyung-Bin Song, Seong-Kwan Ha, Jung-Wook Park, Dong-Jin Kweon, Kyu-Ho.
Convolutional Restricted Boltzmann Machines for Feature Learning Mohammad Norouzi Advisor: Dr. Greg Mori Simon Fraser University 27 Nov
Facets: Fast Comprehensive Mining of Coevolving High-order Time Series Hanghang TongPing JiYongjie CaiWei FanQing He Joint Work by Presenter:Wei Fan.
1 LES of Turbulent Flows: Lecture 12 (ME EN ) Prof. Rob Stoll Department of Mechanical Engineering University of Utah Fall 2014.
Cascade Region Regression for Robust Object Detection
Philipp Gysel ECE Department University of California, Davis
Deep Residual Learning for Image Recognition
Understanding and Predicting Image Memorability at a Large Scale A. Khosla, A. S. Raju, A. Torralba and A. Oliva International Conference on Computer Vision.
Face Recognition based on 2D-PCA and CNN
Convolutional Neural Networks at Constrained Time Cost (CVPR 2015) Authors : Kaiming He, Jian Sun (MSR) Presenter : Hyunjun Ju 1.
Deeply-Recursive Convolutional Network for Image Super-Resolution
Pining for Data II: The Empirical Results Strike Back
Dimensionality Reduction and Principle Components Analysis
Deep Learning for Dual-Energy X-Ray
Demo.
Zhu Han University of Houston Thanks for Dr. Hung Nguyen’s Slides
Data Mining, Neural Network and Genetic Programming
LES of Turbulent Flows: Lecture 16 (ME EN )
Lecture 8:Eigenfaces and Shared Features
Unrolling: A principled method to develop deep neural networks
Overcoming Resource Underutilization in Spatial CNN Accelerators
Deep Learning with TensorFlow online Training at GoLogica Technologies
Lecture 5 Smaller Network: CNN
Training Techniques for Deep Neural Networks
Development of mean value engine model using ANN
Deep Residual Learning for Image Recognition
CNNs and compressive sensing Theoretical analysis
Bilinear Classifiers for Visual Recognition
SBNet: Sparse Blocks Network for Fast Inference
Towards Understanding the Invertibility of Convolutional Neural Networks Anna C. Gilbert1, Yi Zhang1, Kibok Lee1, Yuting Zhang1, Honglak Lee1,2 1University.
Filtering and State Estimation: Basic Concepts
Principal Component Analysis (PCA)
8-3 RRAM Based Convolutional Neural Networks for High Accuracy Pattern Recognition and Online Learning Tasks Z. Dong, Z. Zhou, Z.F. Li, C. Liu, Y.N. Jiang,
Bucket Renormalization for Approximate Inference
Connecting Data with Domain Knowledge in Neural Networks -- Use Deep learning in Conventional problems Lizhong Zheng.
Outline Background Motivation Proposed Model Experimental Results
Example of a simple deep network architecture.
Asymmetric Transitivity Preserving Graph Embedding
Lecture 13: Singular Value Decomposition (SVD)
Going Deeper with Convolutions
The 9 Deep Learning Papers You Need To Know About (Understanding CNNs Part 3) Fait un tour historique du domaine: quels articles/travaux ont été marquants.
EE 193: Parallel Computing
Model Compression Joseph E. Gonzalez
Natalie Lang Tomer Malach
Deep learning enhanced Markov State Models (MSMs)
Learning and Memorization
Example of a simple deep network architecture.
Mohammad Samragh Mojan Javaheripi Farinaz Koushanfar
Single image super-resolution with limited number of filters
Learned Convolutional Sparse Coding
Recent Developments on Super-Resolution
SDSEN: Self-Refining Deep Symmetry Enhanced Network
Rong Ge, Duke University
Presentation transcript:

Compression of CNNs Mooyeol Baek Xiangyu Zhang, Jianhua Zou, Xiang Ming, Kaiming He, Jian Sun: Efficient and Accurate Approximations of Nonlinear Convolutional Networks. Yong-Deok Kim, Eunhyeok Park, Sungjoo Yoo, Taelim Choi, Lu Yang, Dongjun Shin: Compression of Deep Convolutional Neural Networks for Fast and Low Power Mobile Applications.

Motivation It’s practically important to accelerate the test-time computation of CNNs. CNN filters can be approximately decomposed into a series of smaller filters by row-rank approximation.

Approaches Zhang et al. Kim et al. m m c n d k k c m m c 1 1 c m m c’ n n d’ 1 1 n d n n 1 1 k k c’ n n

Efficient and Accurate Approximations of Nonlinear Convolutional Networks. Xiangyu Zhang, Jianhua Zou, Xiang Ming, Kaiming He, Jian Sun

Contribution Low-rank approximation minimizing the reconstruction error of nonlinear responses. Asymmetric reconstruction to reduce the accumulated error of multiple approximated layers. Empirical observation of PCA energy to select proper rank.

Low-rank Approximation m m c n d k k c n n d’ 1 1 n

Low-rank Approximation Relaxation

Asymmetric Reconstruction Uses non-approximate responses to reduce the accumulated error of multiple approximated layers. OriginalApproximated

Rank Selection

Experiments [1] Linear vs. Nonlinear

Experiments [2] Symmetric vs. Asymmetric

Experiments [3] Rank selection

Compression of Deep Convolutional Neural Networks for Fast and Low Power Mobile Applications. Yong-Deok Kim, Eunhyeok Park, Sungjoo Yoo, Taelim Choi, Lu Yang, Dongjun Shin

Contribution One-shot whole network compression scheme which consists of simple three steps: 1.Rank selection (Variational Bayesian matrix factorization) 2.Low-rank tensor decomposition (Tucker decomposition) 3.Fine-tuning.

Tensor Decomposition Tucker decomposition

Tensor Decomposition Zhang et al. Kim et al. m m c n d k k c m m c 1 1 c m m c’ n n d’ 1 1 n d n n 1 1 k k c’ n n

Fine-tuning

Experiments [1]

Experiments [2]