Analysis of Trained CNN (Receptive Field & Weights of Network)

Slides:



Advertisements
Similar presentations
Rich feature Hierarchies for Accurate object detection and semantic segmentation Ross Girshick, Jeff Donahue, Trevor Darrell, Jitandra Malik (UC Berkeley)
Advertisements

Thesis title: “Studies in Pattern Classification – Biological Modeling, Uncertainty Reasoning, and Statistical Learning” 3 parts: (1)Handwritten Digit.
Aula 5 Alguns Exemplos PMR5406 Redes Neurais e Lógica Fuzzy.
Example: ZIP Code Recognition Classification of handwritten numerals.
Radial-Basis Function Networks
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Avoiding Segmentation in Multi-digit Numeral String Recognition by Combining Single and Two-digit Classifiers Trained without Negative Examples Dan Ciresan.
Feedforward semantic segmentation with zoom-out features
Rich feature hierarchies for accurate object detection and semantic segmentation 2014 IEEE Conference on Computer Vision and Pattern Recognition Ross Girshick,
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Hybrid Deep Learning for Reflectance Confocal Microscopy Skin Images
Handwritten Digit Recognition Using Stacked Autoencoders
Demo.
The Relationship between Deep Learning and Brain Function
CS 6501: 3D Reconstruction and Understanding Convolutional Neural Networks Connelly Barnes.
Summary of “Efficient Deep Learning for Stereo Matching”
Object Detection based on Segment Masks
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
Gradient-based Learning Applied to Document Recognition
Data Mining, Neural Network and Genetic Programming
Data Mining, Neural Network and Genetic Programming
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
Article Review Todd Hricik.
Applications of Deep Learning and how to get started with implementation of deep learning Presentation By : Manaswi Advisor : Dr.Chinmay.
Combining CNN with RNN for scene labeling (segmentation)
Rotational Rectification Network for Robust Pedestrian Detection
Lecture 5 Smaller Network: CNN
Neural Networks 2 CS446 Machine Learning.
How it Works: Convolutional Neural Networks
Training Techniques for Deep Neural Networks
Convolutional Networks
CS 698 | Current Topics in Data Science
CS6890 Deep Learning Weizhen Cai
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
Non-linear classifiers Neural networks
Introduction to Convolutional Neural Network (CNN/ConvNET)-insights from amateur George (Tian Zhou)
Fully Convolutional Networks for Semantic Segmentation
A Convolutional Neural Network Cascade For Face Detection
Computer Vision James Hays
Introduction to Neural Networks
Image Classification.
Grid Long Short-Term Memory
Counting in Dense Crowds using Deep Learning
Convolutional Neural Networks
Deep Learning Hierarchical Representations for Image Steganalysis
Object Classification through Deconvolutional Neural Networks
Smart Robots, Drones, IoT
network of simple neuron-like computing elements
Convolutional neural networks Abin - Roozgard.
CSC 578 Neural Networks and Deep Learning
Object Detection Creation from Scratch Samsung R&D Institute Ukraine
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Faster R-CNN By Anthony Martinez.
LECTURE 35: Introduction to EEG Processing
On Convolutional Neural Network
Use 3D Convolutional Neural Network to Inspect Solder Ball Defects
LECTURE 33: Alternative OPTIMIZERS
RCNN, Fast-RCNN, Faster-RCNN
Deep Learning Some slides are from Prof. Andrew Ng of Stanford.
Mihir Patel and Nikhil Sardana
Convolutional Neural Network
CSC 578 Neural Networks and Deep Learning
CIS 519 Recitation 11/15/18.
Department of Computer Science Ben-Gurion University of the Negev
Deep Object Co-Segmentation
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
Learning and Memorization
Image recognition.
CSC 578 Neural Networks and Deep Learning
Shengcong Chen, Changxing Ding, Minfeng Liu 2018
Presentation transcript:

Analysis of Trained CNN (Receptive Field & Weights of Network) Bukweon Kim

Basic Example:Mnist Data Mnist data is data of 28 by 28 sized handwritten digit images, all labeled which is which. Let us observe the structural characteristics with this Mnist data using simple CNN.

Example CNN structure Input Image 24 X 24 28 X 28 5 X 5 X 4 5 X 5 X 4 First Convolution Weights 24 X 24 12 X 12 Convolution Pooling X 8 5 X 5 X 4 Second Convolution Weights 8 X 8 4 X 4 Fully Connected Classify 4 X 4 X 8 X 16 Fully Connected Layer Weights Classify Dictionary 16 X 1 16 X 10 ReLU Softmax ReLU ReLU

Receptive Field Receptive field is the area that this value is determined from. The green box is determined by red red boxes and green box affects cyan box or values. Any value that is not inside red box does not affect the value of the green box.

Classification Similar (high inner product value) 0 1 2 3 4 5 6 7 8 9 1 2 3 4 5 6 7 8 9 Inner product and softmax 0 0 0 0 0 0 1 0 0 0 Result achieved by applying softmax on inner-product result of signal and library This feature strongly suggest input is either 0, 4, 5, 6, or 8 Signals of 16 features extracted from previous steps of CNN Library of 16 features for each digit

Outputs of several 6s 6 We will focus on this 13th signal and analysis what it means. Library of 16 features for 6

6 6 like image examples for explanation We will focus on this 13th signal and analysis what it means. Library of 16 features for 6 14th Mnist Data Confuses between 6 and 0 Classified as 6 Classified as 6 Classified as 1

What is the meaning of this 13th features? 0.2 + 0.1 0.4 0.7 Inner Product ReLU 1.7 For the convenience of understanding I will focus on the strongest signal Output of previous pooling layer Weights for 13th signal of next layer

Weight and signal analysis : Weight value Receptive field of each values. positive negative Main signal given from previous layer 0.22 0.18 0.04 0.05 2.22 2.97 0.07 0.66 0.83 0.61 -1.04 0.15 4.27 -1.18 -1.19 -0.21 -1.22 -1.33 -0.55 -0.67 0.07 0.05 0.12 0.22 0.18 0.04 Value translated as image moved 8 pixels below. Inner Product Inner Product 0.7 -0.7

Weight and signal analysis : Pooling 0.22 0.18 0.04 0.05 The max pooling made signal a bit local translation invariant. Even though we moved the image 2 pixels, the signal of selected pixel did not change 0.14 0.22 0.05 2 pixels 0.22 0.18 0.04 2 pixels 2×2 max pooling with stride 2

Weight and signal analysis : Weight value deeper understanding 1 Increased value of input corresponding to positive weight enhances the signal. 0.42099 0.27828 Increased value of input corresponding to negative weight suppress the signal. 0.03 -0.01 -0.26 -0.09 -0.12 -0.03 -0.85 -0.04 -0.02 0.15 -0.43 0.11 0.14 0.06 0.04 -0.21 0.02 -0.05 0.05 Value change of input corresponding to weight near 0 does not effect the signal much. 0.26084

Weight and signal analysis : Pooling deeper understanding Strongest signal output from one of this inner product or

+ Weight and signal analysis : Weight value deeper understanding 2 Map of the pattern determined from 2 previous filters Convolution & pooling Convolution Map of the signal of combination of any of 4 patterns we looked for ReLU + Weights of first convolution layer Outputs of first pooling layer 8th Weight in second convolutional layer

Weight and signal analysis : Weight value deeper understanding 3 ReLU + Each maps are looking for the patterns somewhat similar to these (these are not exact because it is not linear) the final output may be considered as the value derived from taking many combination of patterns in account. These accounted patterns may not only enhance the value, but also suppress the value. ReLU( + + + )= suppress enhance ReLU( + + + )= Input for fully connected layer!! suppress enhance

Why was CNN fooled/not fooled for examples? The strong 13th signal usually tells if input is 6 or not because of what it looks for. The strong 13th signal usually tells if input is 6 or not because of what it looks for. ignored Library of 16 features for 6 Existence of / pattern on middle fooled them to think it is 1. Confuses between 6 and 0 14th Mnist Data Classified as 6 Classified as 6 Classified as 1

The CNN with ReLU looks for combination of patterns as it gets deeper. Conclusion The CNN with ReLU looks for combination of patterns as it gets deeper. The pooling layer tells CNN that we are looking for local translation invariant features. Deeper layers of CNN allow the network to look for more complex combination of patterns. Also, it allow wider invariance for local patterns. With knowing what exactly CNN looks for, we can tell have deeper understanding of how the CNN works, and what can or can’t it do.

Semantic Segmentation Using Image Classification Pixelwise Classification Amniotic Fluid Umbilical Vein Stomach Bubble Shadowing Artifact Bone Other white region Classification CNN Stomach Bubble Repeat for every pixel Extract Patch centered at pixel Classify pixel

Comparison for Segmentation Result with and without spine position With some change of CNN structure, we could give the spine position information into the CNN structure where we wanted them to be applied