低比特卷积神经网络的量化研究介绍 主讲人:朱锋.

Slides:



Advertisements
Similar presentations
A brief review of non-neural-network approaches to deep learning
Advertisements

Neural networks Introduction Fitting neural networks
ImageNet Classification with Deep Convolutional Neural Networks
Spatial Pyramid Pooling in Deep Convolutional
Deep Convolutional Nets
ImageNet Classification with Deep Convolutional Neural Networks Presenter: Weicong Chen.
Convolutional Neural Network
Deep Residual Learning for Image Recognition
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Convolutional Neural Networks at Constrained Time Cost (CVPR 2015) Authors : Kaiming He, Jian Sun (MSR) Presenter : Hyunjun Ju 1.
Convolutional Neural Networks
Network Compression and Speedup
When deep learning meets object detection: Introduction to two technologies: SSD and YOLO Wenchi Ma.
Wenchi MA CV Group EECS,KU 03/20/2017
Deep Residual Learning for Image Recognition
Deep Learning for Dual-Energy X-Ray
Deep Residual Networks
Object Detection based on Segment Masks
Compact Bilinear Pooling
Data Mining, Neural Network and Genetic Programming
The Problem: Classification
Generative Adversarial Networks
Regularizing Face Verification Nets To Discrete-Valued Pain Regression
Inception and Residual Architecture in Deep Convolutional Networks
Neural Networks 2 CS446 Machine Learning.
Training Techniques for Deep Neural Networks
Efficient Deep Model for Monocular Road Segmentation
Deep Residual Learning for Image Recognition
ECE 599/692 – Deep Learning Lecture 6 – CNN: The Variants
Fully Convolutional Networks for Semantic Segmentation
Layer-wise Performance Bottleneck Analysis of Deep Neural Networks
Bird-species Recognition Using Convolutional Neural Network
Computer Vision James Hays
Introduction to Neural Networks
Goodfellow: Chap 6 Deep Feedforward Networks
Image Classification.
NormFace:
Toward improved document classification and retrieval
A Comparative Study of Convolutional Neural Network Models with Rosenblatt’s Brain Model Abu Kamruzzaman, Atik Khatri , Milind Ikke, Damiano Mastrandrea,
Deep Learning Tutorial
Jia-Bin Huang Virginia Tech ECE 6554 Advanced Computer Vision
Two-Stream Convolutional Networks for Action Recognition in Videos
Neural Network Compression
Object Detection + Deep Learning
ECE 599/692 – Deep Learning Lecture 5 – CNN: The Representative Power
Smart Robots, Drones, IoT
8-3 RRAM Based Convolutional Neural Networks for High Accuracy Pattern Recognition and Online Learning Tasks Z. Dong, Z. Zhou, Z.F. Li, C. Liu, Y.N. Jiang,
Single Image Rolling Shutter Distortion Correction
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Lecture: Deep Convolutional Neural Networks
Outline Background Motivation Proposed Model Experimental Results
Neural networks (3) Regularization Autoencoder
ImageNet Classification with Deep Convolutional Neural Networks
Heterogeneous convolutional neural networks for visual recognition
Course Recap and What’s Next?
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Neural Architecture Search: Basic Approach, Acceleration and Tricks
Model Compression Joseph E. Gonzalez
Dynamic Neural Networks Joseph E. Gonzalez
Authors: Chaim Baskin, Natan Liss, Evgenii Zheltonozhskii, Alex M
Deep Object Co-Segmentation
Natalie Lang Tomer Malach
CS295: Modern Systems: Application Case Study Neural Network Accelerator Sang-Woo Jun Spring 2019 Many slides adapted from Hyoukjun Kwon‘s Gatech “Designing.
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
Learning and Memorization
End-to-End Facial Alignment and Recognition
CRCV REU 2019 Kara Schatz.
Principles of Back-Propagation
Mohsen Imani, Saransh Gupta, Yeseong Kim, Tajana Rosing
ICLR, 2019 Jiahe Li
Presentation transcript:

低比特卷积神经网络的量化研究介绍 主讲人:朱锋

目 录CONTENTS 1 研究背景 目 录 CONTENTS 2 现有量化研究 3 量化训练加速

研究背景

• Language understanding Video • Video understanding … Background Image • Classification • Localization • Segmentation Audio • Speech recognition • Language understanding Video • Video understanding …

How to get better performance ? Background How to get better performance ? Complicated Models

Challenges of Deploying: • Limited computing resources Background Challenges of Deploying: • Limited computing resources • Short response time • Millions of parameters • Complicated model architecture Model Architecture Parameters Top-1 ERR Top-5 ERR AlexNet 8 Layers (5conv + 3fc) ~ 60 million 40.7% 15.3% VGG 19 Layers (16conv + 3fc) ~ 144 million 24.4% 7.1% GoogLeNet 22 Layers ~ 6.8 million - 7.9% ResNet 52 Layers (50conv + 2fc) ~ 200 million 21.29% 5.71%

现有量化研究

Quantization Quantization is the process of constraining an input from a continuous set to a discrete set. Quantization of Neural Network For model param: 32-bit float lower bit int

non-uniform quantization HWGQ(CVPR2017) LQ-Net(ECCV2018) uniform quantization Quantization Dorefa-Net(CVPR2016) PACT(arxiv 2018) binarization BNN(NIPS2016) XNOR-Net(ECCV2016) Bi-Real(ECCV2018)

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) Zhaowei Cai (UCSD) Xiaodong He(MRR) Jian Sun(Megvii Inc) Nuno Vasconcelos(UCSD) Half-Wave Gaussian Quantization(HWGQ-net)--> Quantization of Activation layer and Bacth Normalization layer. Quantization of activation is more difficult than that of weights, because of activation function.

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) Binary Networks Binary Activation Quantization Derivative of C with respect to w A problem: derivative almost zero.

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) Backward use hard tanh Still two problems: 1、gradient varnish 2、gradient mismatch

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) ReLU——half-wave rectifier Forward approximation

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) Half-Wave Gaussian Quantization(HWGQ-net) Backward approximation Vanila ReLU Clipped ReLU

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) Log-tailed ReLU

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) Experiments Dataset: ImageNet FW: full precision weight, BW: binary weight

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) Experiments

Deep Learning with Low Precision by Half-wave Gaussian Quantization (CVPR 2017) Experiments

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) Jungwook Choi, Pierce I-Jen Chuang, Zhuo Wang, Swagath Venkataramani, Vijayalakshmi Srinivasan, Kailash Gopalakrishnan IBM Research AI Challenge in Activation Quantization

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) Jungwook Choi, Pierce I-Jen Chuang, Zhuo Wang, Swagath Venkataramani, Vijayalakshmi Srinivasan, Kailash Gopalakrishnan IBM Research AI Challenge in Activation Quantization

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) Parameterized Clipping Activation to replace ReLU Quantization STE to get derivative

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) Parameterized Clipping Activation to replace ReLU

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) Balancing Clipping and Quantization Error

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) Statistics-Aware Weight Binning

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) Experiment

Binarized Neural Networks: Training Neural Networks with Weights and Activations Constrained to +1 or -1(NIPS 2016) Matthieu Courbariaux, Yoshua Bengio(Universite de Montreal) Deterministic Stochastic Binarization

Binarized Neural Networks: Training Neural Networks with Weights and Activations Constrained to +1 or -1(NIPS 2016) Matthieu Courbariaux, Yoshua Bengio(Universite de Montreal) Deterministic STE- hard tanh XNOR replace mul

XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks(ECCV 2016) Mohammad Rastegariy, Vicente Ordonezy, Joseph Redmon, Ali Farhadiy (University of Washington) Binary-Weight-Networks Estimating Binary Weights

XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks(ECCV 2016) XNOR-Networks Binary Dot Product Optimal Solution

XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks(ECCV 2016) Experiment

XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks(ECCV 2016) Experiment AlexNet Dataset: Cifar10 ResNet-18 & GooLenet DataSet: ImageNet

Dorefa-Net: Training Low Bitwidth Convolutional Neural Networks With Low Bitwidth Gradients Shuchang Zhou, Yuxin Wu, Zekun Ni, Xinyu Zhou, He Wen, Yuheng Zou (Megvii Inc) activation weight

Dorefa-Net: Training Low Bitwidth Convolutional Neural Networks With Low Bitwidth Gradients FIRST AND THE LAST LAYER DO NOT QUANTIZE Experiment

Bi-Real Net: Enhancing the Performance of 1-bit CNNs With Improved Representational Capability and Advanced Training Algorithm(ECCV 2018) Zechun Liu(HKUST), Baoyuan Wu(Tencent AI), Wenhan Luo(Tencen t AI), Xin Yang(HUST), Wei Liu(Tencent AI), and Kwang-Ting Cheng(HKUST)

Bi-Real Net: Enhancing the Performance of 1-bit CNNs With Improved Representational Capability and Advanced Training Algorithm(ECCV 2018) Zechun Liu(HKUST), Baoyuan Wu(Tencent AI), Wenhan Luo(Tencen t AI), Xin Yang(HUST), Wei Liu(Tencent AI), and Kwang-Ting Cheng(HKUST) 1、representation capability

Bi-Real Net: Enhancing the Performance of 1-bit CNNs With Improved Representational Capability and Advanced Training Algorithm(ECCV 2018)

Bi-Real Net: Enhancing the Performance of 1-bit CNNs With Improved Representational Capability and Advanced Training Algorithm(ECCV 2018) 2、gradient mismatch Weight initialization use Clip

Bi-Real Net: Enhancing the Performance of 1-bit CNNs With Improved Representational Capability and Advanced Training Algorithm(ECCV 2018) Experiment

Bi-Real Net(ECCV 2018) Experiment

LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks (ECCV 2018) Dongqing Zhang, Jiaolong Yang, Dongqiangzi Ye, and Gang Hua (Microsoft Research)

LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks (ECCV 2018) Learnable Quantization Function

LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks (ECCV 2018) Training Process Two Params to optimize, two step: 1.Fix v learn B 2.Fix B and update v

LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks (ECCV 2018) Experiment

LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks (ECCV 2018) Experiment

THANKS 谢谢大家聆听