Vessel Extraction in X-Ray Angiograms Using Deep Learning

Slides:



Advertisements
Similar presentations
Deep Learning and Neural Nets Spring 2015
Advertisements

Unsupervised Learning With Neural Nets Deep Learning and Neural Nets Spring 2015.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Spatial Pyramid Pooling in Deep Convolutional
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Object detection, deep learning, and R-CNNs
Feedforward semantic segmentation with zoom-out features
Image Captioning Approaches
Rich feature hierarchies for accurate object detection and semantic segmentation 2014 IEEE Conference on Computer Vision and Pattern Recognition Ross Girshick,
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
S.Bengio, O.Vinyals, N.Jaitly, N.Shazeer
When deep learning meets object detection: Introduction to two technologies: SSD and YOLO Wenchi Ma.
Recent developments in object detection
Hybrid Deep Learning for Reflectance Confocal Microscopy Skin Images
Convolutional Sequence to Sequence Learning
CS 4501: Introduction to Computer Vision Object Localization, Detection, Semantic Segmentation Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy.
Convolutional Neural Network
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
The Relationship between Deep Learning and Brain Function
Object Detection based on Segment Masks
Object detection with deformable part-based models
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
From Vision to Grasping: Adapting Visual Networks
A practical guide to learning Autoencoders
Krishna Kumar Singh, Yong Jae Lee University of California, Davis
Article Review Todd Hricik.
Neural Machine Translation by Jointly Learning to Align and Translate
Intelligent Information System Lab
Natural Language Processing of Knee MRI Reports
Huazhong University of Science and Technology
Lecture 5 Smaller Network: CNN
Convolutional Networks
CS6890 Deep Learning Weizhen Cai
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
Dynamic Routing Using Inter Capsule Routing Protocol Between Capsules
Image Question Answering
A Convolutional Neural Network Cascade For Face Detection
By: Kevin Yu Ph.D. in Computer Engineering
Attention-based Caption Description Mun Jonghwan.
Introduction to Neural Networks
Image Classification.
Convolutional Neural Networks for Visual Tracking
Creating Data Representations
The Big Health Data–Intelligent Machine Paradox
Papers 15/08.
Outline Background Motivation Proposed Model Experimental Results
Object Tracking: Comparison of
RCNN, Fast-RCNN, Faster-RCNN
Introduction to Artificial Intelligence Lecture 24: Computer Vision IV
Convolutional Neural Networks
Heterogeneous convolutional neural networks for visual recognition
Learn to Comment Mentor: Mahdi M. Kalayeh
by Khaled Nasr, Pooja Viswanathan, and Andreas Nieder
Department of Computer Science Ben-Gurion University of the Negev
Jointly Generating Captions to Aid Visual Question Answering
Sequence to Sequence Video to Text
Presented by: Anurag Paul
Deep Object Co-Segmentation
Image Processing and Multi-domain Translation
Feature Selective Anchor-Free Module for Single-Shot Object Detection
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Object Detection Implementations
Presented By: Harshul Gupta
Multi-UAV to UAV Tracking
End-to-End Facial Alignment and Recognition
Example of training and deployment of deep convolutional neural networks. Example of training and deployment of deep convolutional neural networks. During.
Week 7 Presentation Ngoc Ta Aidean Sharghi
Neural Machine Translation by Jointly Learning to Align and Translate
Presented By: Firas Gerges (fg92)
Presentation transcript:

Vessel Extraction in X-Ray Angiograms Using Deep Learning Ziyan Li Nasr-Esfahani E, Samavi S, Karimi N, et al. Vessel extraction in X-ray angiograms using deep learning[C]//Engineering in Medicine and Biology Society (EMBC), 2016 IEEE 38th Annual International Conference of the. IEEE, 2016: 643-646.

Overview Vessel extraction:deep CNN Preprocess:top-hat transform WTH and BTH Extract small elements

CNN Architecture input image:patch architecture: output: (33*33, 1040000 patches) architecture: two convolve layers(depth 50) two max pooling layers two fully connected layers output: two probabilities for the center pixel of the patch (vessel region and background)

Results Get binary mask:threshold T Dataset:44 X-ray images (resolution of 512*512) Train:26 Test:18 The ratio of foreground and background: 1:1 Metrics: accuracy,sensitivity,specificity,PPV(阳性预测率),NPV

Results

From Captions to Visual Concepts and Back Fang H, Gupta S, Iandola F, et al. From captions to visual concepts and back[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015: 1473-1482.

Overview Main approaches:retrieval of existing human-written captions, and generation of novel captions Caption Generation Pipeline Detect words Generate sentences Re-rank sentences

Word Detection Training Word Detectors Vocabulary :1000 most common words in the training captions Train detectors: Not standard supervised learning:no words bounding boxes Weakly-supervised approach of Multiple Instance Learning (多示例 学习) Compute 𝑝 𝑖𝑗 𝜔 :a logistic function on top of the fc7 layer The probability of word 𝜔 in bag 𝑏 𝑖 bounding boxes may not be easily defined, such as open or beautiful 每张图像即一个bag,图像中的patch就是示例,单词w出现在图像的标注中,图像 i 就标记为正; 单词使用one-hot编码

Word Detection MIL: 𝑝 𝑖𝑗 𝜔 —> 𝑝 𝑖 𝜔 Get coarse spatial response map Each location expresses the 𝑝 𝑖𝑗 𝜔 MIL: 𝑝 𝑖𝑗 𝜔 —> 𝑝 𝑖 𝜔 Generating Word Scores for a Test Image for all words 𝜔 in the vocabulary, put the image into CNN to obtain 𝑝 𝑖 𝜔 Output all words 𝜔 if 𝑝 𝑖 𝜔 >𝑡ℎ𝑟𝑒𝑠ℎ𝑜𝑙𝑑 𝜏 bounding boxes may not be easily defined, such as open or beautiful 每张图像即一个bag,图像中的patch就是示例,单词w出现在图像的标注中,图像 i 就标记为正; 单词使用one-hot编码

Word Detection Generating Word Scores for a Test Image

Language Generation To generate candidate captions, we use a maximum entropy (ME) LM conditioned on the set of visually detected words objective function of MELM: log likelihood of the of the captions where <s> denotes the start-of-sentence token, ¯ wj 2 V [ </s>, and fk(wl, · · · , w1, ˜Vl−1) and #k respectively denote the k-th max-entropy feature and its weight

Language Generation Generation Process Choose M-best sentences Beam Search(束搜索) When </s>(the end of sentence token) is generated, the full path to </s> is removed from the stack and set aside as a completed sentence. The process continues until a maximum sentence length L is reached Choose M-best sentences Given a target number of T image attributes to be mentioned, the sequences in C covering at least T objects are added to the M-best list, sorted in descending order by the log- likelihood. This maintains a stack of length l

Sentence Re-Ranking Use MERT(最小错误率训练法) to re-rank the M sentences uses a linear combination of features a new feature: DMSM(多模态相似模型) Deep Multimodal Similarity Model image model text model learn two neural networks, map image and text to vectors measure cosine similarity 1. The log-likelihood of the sequence. 2. The length of the sequence. 3. The log-probability per word of the sequence. 4. The logarithm of the sequence’s rank in the log-likelihood. 5. 11 binary features indicating whether the number of mentioned objects is x (x = 0,...,10). 6. The DMSM score between the sequence and the image.

Results Datasets 1:Microsoft COCO dataset PPLX、BLEU、METEOR 82783 training images 40504 validation images PPLX、BLEU、METEOR PPLX (perplexity) measures the uncertainty of the language model METEOR measures unigram precision and re- call Uncon- ditioned, which generates sentences by sampling an N- gram LM without knowledge of the visual word detectors; and Shuffled Human, which randomly picks another human generated caption from another image. Both the BLEU and METEOR scores are very low for these ap- proaches, demonstrating the variation and complexity of the Microsoft COCO dataset.