Neural Network Pipeline CONTACT & ACKNOWLEDGEMENTS

Slides:



Advertisements
Similar presentations
Université du Québec École de technologie supérieure Face Recognition in Video Using What- and-Where Fusion Neural Network Mamoudou Barry and Eric Granger.
Advertisements

Spatial Pyramid Pooling in Deep Convolutional
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
From R-CNN to Fast R-CNN
Electrical and Computer Systems Engineering Postgraduate Student Research Forum 2001 WAVELET ANALYSIS FOR CONDITION MONITORING OF CIRCUIT BREAKERS Author:
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
BING: Binarized Normed Gradients for Objectness Estimation at 300fps
Fully Convolutional Networks for Semantic Segmentation
Feedforward semantic segmentation with zoom-out features
Cascade Region Regression for Robust Object Detection
Rich feature hierarchies for accurate object detection and semantic segmentation 2014 IEEE Conference on Computer Vision and Pattern Recognition Ross Girshick,
Philipp Gysel ECE Department University of California, Davis
Spatial Localization and Detection
Deep Residual Learning for Image Recognition
Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition arXiv: v4 [cs.CV(CVPR)] 23 Apr 2015 Kaiming He, Xiangyu Zhang, Shaoqing.
Facial Smile Detection Based on Deep Learning Features Authors: Kaihao Zhang, Yongzhen Huang, Hong Wu and Liang Wang Center for Research on Intelligent.
When deep learning meets object detection: Introduction to two technologies: SSD and YOLO Wenchi Ma.
Recent developments in object detection
Deep Learning for Dual-Energy X-Ray
Faster R-CNN – Concepts
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
The Problem: Classification
Week III: Deep Tracking
Gender Classification Using Scaled Conjugate Gradient Back Propagation
Article Review Todd Hricik.
Mentor: Afshin Dehghan
Robust Lung Nodule Classification using 2
Regularizing Face Verification Nets To Discrete-Valued Pain Regression
Combining CNN with RNN for scene labeling (segmentation)
YOLO9000:Better, Faster, Stronger
Compositional Human Pose Regression
Unrolling: A principled method to develop deep neural networks
Part-Based Room Categorization for Household Service Robots
Ajita Rattani and Reza Derakhshani,
Hierarchical Deep Convolutional Neural Network
Synthesis of X-ray Projections via Deep Learning
Deep Residual Learning for Image Recognition
Project Implementation for ITCS4122
Low Dose CT Image Denoising Using WGAN and Perceptual Loss
Using Tensorflow to Detect Objects in an Image
Layer-wise Performance Bottleneck Analysis of Deep Neural Networks
Bird-species Recognition Using Convolutional Neural Network
Normalized Cut Loss for Weakly-supervised CNN Segmentation
Neural network systems
Toward improved document classification and retrieval
Figure 4. Testing minimal configurations with existing models for spatiotemporal recognition. (A-B) A binary classifier is trained to separate a positive.
Jia-Bin Huang Virginia Tech ECE 6554 Advanced Computer Vision
Object Detection + Deep Learning
Hairong Qi, Gonzalez Family Professor
Pose Estimation for non-cooperative Spacecraft Rendevous using CNN
8-3 RRAM Based Convolutional Neural Networks for High Accuracy Pattern Recognition and Online Learning Tasks Z. Dong, Z. Zhou, Z.F. Li, C. Liu, Y.N. Jiang,
YOLO-LITE: A Real-Time Object Detection Web Implementation
Outline Background Motivation Proposed Model Experimental Results
Object Tracking: Comparison of
RCNN, Fast-RCNN, Faster-RCNN
Heterogeneous convolutional neural networks for visual recognition
Course Recap and What’s Next?
Department of Computer Science Ben-Gurion University of the Negev
Human-object interaction
Natalie Lang Tomer Malach
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
Object Detection Implementations
Unrolling the shutter: CNN to correct motion distortions
End-to-End Facial Alignment and Recognition
CRCV REU 2019 Kara Schatz.
Week 3 Volodymyr Bobyr.
Jiahe Li
Deep Video Quality Assessor: From Spatio-temporal Visual Sensitivity to A convolutional Neural Aggregation Network Woojae Kim1, Jongyoo Kim2, Sewoong Ahn1,Jinwoo.
Adrian E. Gonzalez , David Parra Department of Computer Science
ICLR, 2019 Jiahe Li
Presentation transcript:

Neural Network Pipeline CONTACT & ACKNOWLEDGEMENTS Reconstruction-free deep convolutional neural networks for partially observed images Arun Asokan Nair*, Luoluo Liu*, Akshay Rangamani*, Peter Chin*^, Muyinatu A. Lediju Bell *+#, Trac D. Tran* *Department of Electrical and Computer Engineering, +Department of Computer Science, #Department of Biomedical Engineering - Johns Hopkins University ^Department of Computer Science - Boston University Introduction Neural Network Pipeline We propose a framework to extract information from visual data with an unknown fraction of pixels missing using CNNs, without performing reconstruction or re-training the neural network fon every possible partial observation ratio. We prove that by training a neural network on a few observation ratios, it generalizes to unseen observation ratios for both classification and object detection tasks. Experiments Image Classification on CIFAR-10 using VGG-16 Object Detection on Pascal VOC 2007 using Faster R-CNN [3] Fig. Image classification on partially observed images. During training, we input fully sampled images and images with missing pixel ratios of 0.5, 0.25, 0.125 to a VGG-16[2] network. The test data to the network has observation ratios randomly generated between 0 and 1. Networks were trained using SGD with momentum=0.9, learning rate=0.1, learning rate decay=10-6 and batch size=128 for 250 epochs Conclusions Benefits of our approach: Faster than reconstruct-then-classify Higher accuracies for challenging partial observation ratios (0.2 and 0.1). Generalizes to arbitrary unseen observation ratios. Motivation Pixel-wise Coded Exposure (PCE) Imaging [1] is a technique used to sense and record video sequences under power and sensor size constraints. The PCE measurements are close to partially observed frames, which we study in this work through the classification and object detection tasks. References Results [1] Zhang, J., Xiong, T., Tran, T., Chin, S., & Etienne-Cummings, R. (2016). Compact all-CMOS spatiotemporal compressive sensing video camera with pixel-wise coded exposure. Optics express, 24(8), 9013-9024. [2] Karen Simonyan and Andrew Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv:1409.1556, 2014. [3] Shaoqing Ren et al., “Faster R-CNN: Towards real-time object detection with region proposal networks,” in Advances in Neural Information Processing Systems, 2015 Exemplars Object Detection Results from a Faster-RCNN [3] network trained on different partial observation ratios Observation ratio: 50% Observation ratio: 25% Fig. Testing times Fig. Averaged classification accuracies Fig. Mean Average Precision(mAP) CONTACT & ACKNOWLEDGEMENTS Contact author – Arun Asokan Nair & Akshay Rangamani E-mail – anair8@jhu.edu, rangamani.akshay@jhu.edu This work is partially supported by the National Science Foundation (NSF) under Grant CCF-1422995. Table: Averaged classification accuracies with various partial observation ratios -> DICE score!! ------------------------ Text and figures could be more integrated.