Approximate Fully Connected Neural Network Generation

Slides:



Advertisements
Similar presentations
Marković Miljan 3139/2011
Advertisements

ELE 523E COMPUTATIONAL NANOELECTRONICS W3: Quantum Computing, 22/9/2014 FALL 2014 Mustafa Altun Electronics & Communication Engineering Istanbul Technical.
An Overview of the Application of Neural Networks to the Monitoring of Civil Engineering Structures By Brian Walsh & Arturo González With thanks thanks.
Tuomas Sandholm Carnegie Mellon University Computer Science Department
Characterization Presentation Neural Network Implementation On FPGA Supervisor: Chen Koren Maria Nemets Maxim Zavodchik
Handwritten Character Recognition Using Artificial Neural Networks Shimie Atkins & Daniel Marco Supervisor: Johanan Erez Technion - Israel Institute of.
Moving NN Triggers to Level-1 at LHC Rates Triggering Problem in HEP Adopted neural solutions Specifications for Level 1 Triggering Hardware Implementation.
Presenting: Itai Avron Supervisor: Chen Koren Final Presentation Spring 2005 Implementation of Artificial Intelligence System on FPGA.
Embedded Systems Laboratory Informatics Institute Federal University of Rio Grande do Sul Porto Alegre – RS – Brazil SRC TechCon 2005 Portland, Oregon,
Presenting: Itai Avron Supervisor: Chen Koren Characterization Presentation Spring 2005 Implementation of Artificial Intelligence System on FPGA.
Hopefully a clearer version of Neural Network. With Actual Weights.
Presenting: Itai Avron Supervisor: Chen Koren Mid Semester Presentation Spring 2005 Implementation of Artificial Intelligence System on FPGA.
Information Fusion Yu Cai. Research Article “Comparative Analysis of Some Neural Network Architectures for Data Fusion”, Authors: Juan Cires, PA Romo,
GPGPU platforms GP - General Purpose computation using GPU
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
A Genetic Algorithms Approach to Feature Subset Selection Problem by Hasan Doğu TAŞKIRAN CS 550 – Machine Learning Workshop Department of Computer Engineering.
03/12/20101 Analysis of FPGA based Kalman Filter Architectures Arvind Sudarsanam Dissertation Defense 12 March 2010.
Accuracy-Configurable Adder for Approximate Arithmetic Designs
1 st Neural Network: AND function Threshold(Y) = 2 X1 Y X Y.
Evolving a Sigma-Pi Network as a Network Simulator by Justin Basilico.
Floating Point vs. Fixed Point for FPGA 1. Applications Digital Signal Processing -Encoders/Decoders -Compression -Encryption Control -Automotive/Aerospace.
A comprehensive method for the evaluation of the sensitivity to SEUs of FPGA-based applications A comprehensive method for the evaluation of the sensitivity.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Network Theory and Application Ashish Venugopal Sriram Gollapalli Ulas Bardak.
Hybrid AI & Machine Learning Systems Using Ne ural Network and Subsumption Architecture Libraries By Logan Kearsley.
Backpropagation An efficient way to compute the gradient Hung-yi Lee.
1 Rapid Estimation of Power Consumption for Hybrid FPGAs Chun Hok Ho 1, Philip Leong 2, Wayne Luk 1, Steve Wilton 3 1 Department of Computing, Imperial.
Back-Propagation MLP Neural Network Optimizer ECE 539 Andrew Beckwith.
Implementation of Finite Field Inversion
Design of an 8-bit Carry-Skip Adder Using Reversible Gates Vinothini Velusamy, Advisor: Prof. Xingguo Xiong Department of Electrical Engineering, University.
ITERATIVE CHANNEL ESTIMATION AND DECODING OF TURBO/CONVOLUTIONALLY CODED STBC-OFDM SYSTEMS Hakan Doğan 1, Hakan Ali Çırpan 1, Erdal Panayırcı 2 1 Istanbul.
Dr. Sudharman K. Jayaweera and Amila Kariyapperuma ECE Department University of New Mexico Ankur Sharma Department of ECE Indian Institute of Technology,
PARALLELIZATION OF ARTIFICIAL NEURAL NETWORKS Joe Bradish CS5802 Fall 2015.
Copyright © 2010 Houman Homayoun Houman Homayoun National Science Foundation Computing Innovation Fellow Department of Computer Science University of California.
Implementing Local Relative Sensitivity Pruning Paul Victorey.
Neural Networks Vladimir Pleskonjić 3188/ /20 Vladimir Pleskonjić General Feedforward neural networks Inputs are numeric features Outputs are in.
Perceptrons Michael J. Watts
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
Ghent University Backpropagation for Population-Temporal Coded Spiking Neural Networks July WCCI/IJCNN 2006 Benjamin Schrauwen and Jan Van Campenhout.
A Synthesis Method for MVL Reversible Logic by 1 Department of Computer Science, University of Victoria, Canada M. Miller 1, G. Dueck 2, and D. Maslov.
Multiple-Layer Networks and Backpropagation Algorithms
Neural Networks.
OPTIMIZATION OF MODELS: LOOKING FOR THE BEST STRATEGY
Chilimbi, et al. (2014) Microsoft Research
CSE 473 Introduction to Artificial Intelligence Neural Networks
MLP Based Feedback System for Gas Valve Control in a Madison Symmetric Torus Andrew Seltzman Dec 14, 2010.
Neural Networks: Improving Performance in X-ray Lithography Applications ECE 539 Ryan T. Hogg May 10, 2000.
Lecture 11. MLP (III): Back-Propagation
Training a Neural Network
Deep Learning Hierarchical Representations for Image Steganalysis
Artificial Neural Network & Backpropagation Algorithm
BACKPROPAGATION Multlayer Network.
Camera Calibration Using Neural Network for Image-Based Soil Deformation Measurement Systems Zhao, Honghua Ge, Louis Civil, Architectural, and Environmental.
Face Recognition with Neural Networks
network of simple neuron-like computing elements
Emre O. Neftci  iScience  Volume 5, Pages (July 2018) DOI: /j.isci
Optimization for Fully Connected Neural Network for FPGA application
Final Project presentation
Capabilities of Threshold Neurons
Department of Electrical Engineering
ECE 434 Advanced Digital System L11
ELE 523E COMPUTATIONAL NANOELECTRONICS
Tuba Ayhan, Fırat Kula, Mustafa Altun
August 8, 2006 Danny Budik, Itamar Elhanany Machine Intelligence Lab
Introduction to Neural Network
Department of Computer Science Ben-Gurion University of the Negev
Surrogate Modeling for Predicting FPGA Place and Route
Learning and Memorization
Circuit-Based Intrinsic Methods to Detect Overfitting
Implementation of CMOS Logic Circuits with
Presentation transcript:

Approximate Fully Connected Neural Network Generation Tuba Ayhan, Mustafa Altun tuba.ayhan@itu.edu.tr Istanbul Technical University Electronics and Communication Engineering Department Emerging Circuits and Computation (ECC) Group – ITU

Approximate computing Exact implementation High performance High area and power consumption Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Approximate computing Exact implementation High performance High area and power consumption Approximate implementation Lower performance Low area and power consumption Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Approximate FCNN Generation tuba.ayhan@itu.edu.tr Outline Area reduction Network generation Approximation model Network implementation Multiplier library Architecture template Layer information Error tolerance w, b Performance loss estimation Area cost estimation Network architecture Layer architecture Constant multiplier block Area reduction Error propagation Area reduction algorithm Test results Conclusion Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Approximate FCNN Generation tuba.ayhan@itu.edu.tr Outline Area reduction Network generation Approximation model Network implementation Multiplier library Architecture template Layer information Error tolerance w, b Performance loss estimation Area cost estimation Network architecture Layer architecture Constant multiplier block Area reduction Error propagation Area reduction algorithm Test results Conclusion Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Approximate FCNN Generation tuba.ayhan@itu.edu.tr Outline Area reduction Network generation Approximation model Network implementation Multiplier library Architecture template Layer information Error tolerance w, b Performance loss estimation Area cost estimation Network architecture Layer architecture Constant multiplier block Area reduction Error propagation Area reduction algorithm Test results Conclusion Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Approximate FCNN Generation tuba.ayhan@itu.edu.tr Outline Area reduction Network generation Approximation model Network implementation Multiplier library Architecture template Layer information Error tolerance w, b Performance loss estimation Area cost estimation Network architecture Layer architecture Constant multiplier block Area reduction Error propagation Area reduction algorithm Test results Conclusion Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Network architecture – Layer architecture 1/2 X0 X1 1 LAYER 1 w00 + X w01 w02 w04 b0 w11 w12 w14 w10 b1 b2 b3 b4 w03 w13 X0 X1 O0 O1 y 𝑛 =𝑓 𝑤 0𝑛 . 𝑥 0 + 𝑤 1𝑛 . 𝑥 1 + 𝑏 𝑛 .1 O 0 = 𝑤 00 . 𝑥 0 + 𝑤 01 . 𝑥 1 + 𝑏 0 .1 O 1 = 𝑤 01 . 𝑥 0 + 𝑤 11 . 𝑥 1 + 𝑏 1 .1 .... .... Layer 1 architecture O 4 = 𝑤 04 . 𝑥 0 + 𝑤 14 . 𝑥 1 + 𝑏 4 .1 Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Network architecture – Layer architecture 2/2 X X0 X1 1 LAYER 1 LAYER 2 LAYER 3 w00 + w01 w02 w03 w04 b0 w11 w12 w13 w14 w10 b1 b2 b3 b4 X0 X1 O0 O1 Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Network architecture – Constant multiplier block 1/2 <<0 + <<1 <<2 <<3 <<4 <<5 <<6 <<7 CMB-L1 Xi O0 O1 O2 O3 O4 𝐎= 𝒏=𝟏 𝑲 𝒂 𝒏 𝒙 Constant multiplier block for layer 1: 5 outputs  5 adders 8 bit input  8 shifted versions and a 0. 𝒂 𝒏 = 𝟐 𝒊 ,𝟎 Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Network architecture – Constant multiplier block 2/2 8-bit signed multiplication Error: Mean = 0 std = depends on # addends Implemented on Spartan6 FPGA Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Area reduction – Error propagation Take binary classification problem X0 X1 O0 O1 O0-O1 Class0 Class1 Exact: A3 Error H3 A2 H2 Approx.: Class0 Class1 A1 H1 Approximation level of layer i: Ai Noise added at layer i: Hi # neurons in layer i: ni O0-O1 Error tolerance, eT 𝑯 𝒊 = 𝑨 𝒊 + 𝑯 𝒊−𝟏 𝟐 𝒏 𝒊−𝟏 𝑯 𝟑 ≤ 𝐞 𝐓 Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Area reduction – Area reduction algorithm 𝐦𝐢𝐧 𝑵 𝑻 .𝑪 𝒙 s.t. 𝑯 𝑳 ≤ 𝒆 𝑻 𝒍𝒃≤𝒙≤𝒖𝒃 Curve fitting applied: C(A)=𝟓𝟓 𝒆 −𝟐.𝟓𝑨 +𝟔𝟑 𝒆 −𝟎.𝟏𝑨 Example 𝑳=𝟑 X0 X1 O0 O1 𝒙= 𝑨 𝟏 𝑨 𝟐 𝑨 𝟑 𝑵= 𝟏𝟎 𝟏𝟓 𝟔 𝑯 𝟑 = 𝑨 𝟑 + 𝑨 𝟐 + 𝑨 𝟏 𝟐 .𝟓 𝟏 𝟐 .𝟑 ≤ 𝒆 𝑻 Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Approximate FCNN Generation tuba.ayhan@itu.edu.tr Simulation results Simulation Setup #neurons Exact eT = 0.5% eT = 5% eT = 10% 120 1826 1521 513 380 96 1463 1431 529 328 72 1026 927 497 286 60 745 617 314 216 36 614 292 165 163 12 252 106 104 Define architecture Implement Exact Approx.with eT=0.5 Approx.with eT=5 Approx.with eT=10 Include weights Problem 1 Problem 10 Problem 2 . 12 Neurons X0 X1 O0 O1 120 Neurons Approximation Xilinx Spartan6 FPGA with ISE 14.7 design tools Networks have 12 ~120 neurons. 1 exact and 3 approximate implementations are available. Each network is trained for 10 different classification problems. Results are given as average over 10 problems. Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Approximate FCNN Generation tuba.ayhan@itu.edu.tr Conclusion Approximate FCNN generation framework is proposed with: An FCN structure and An area reduction algorithm utilizing the proposed structure The approximation error of the constant multiplier blocks can be altered. The area consumption of the network is reduced, yet the network can provide targeted classification performance. The area optimization can be completely detached from FCN training. Without prior information on network constants, the framework can still achieve up to 73% area saving with less than 5% loss in classification accuracy, in our examples. This project is funded by TUBITAK (The Scientific and Technological Research Council of Turkey) 1001 Project # 117E078. Approximate FCNN Generation tuba.ayhan@itu.edu.tr

Approximate FCNN Generation tuba.ayhan@itu.edu.tr Thank you. tuba.ayhan@itu.edu.tr Istanbul Technical University Emerging Circuits and Computation (ECC) Group Approximate FCNN Generation tuba.ayhan@itu.edu.tr