Low Dose CT Image Denoising Using WGAN and Perceptual Loss

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Duke University COPYRIGHT © DUKE UNIVERSITY 2012 Sparsity Based Denoising of Spectral Domain Optical Coherence Tomography Images Leyuan Fang, Shutao Li,
Unsupervised Learning With Neural Nets Deep Learning and Neural Nets Spring 2015.
1 Blind Image Quality Assessment Based on Machine Learning 陈 欣
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
Learning to perceive how hand-written digits were drawn Geoffrey Hinton Canadian Institute for Advanced Research and University of Toronto.
ImageNet Classification with Deep Convolutional Neural Networks Presenter: Weicong Chen.
SUPER RESOLUTION USING NEURAL NETS Hila Levi & Eran Amar Weizmann Ins
Conditional Generative Adversarial Networks
Generative Adversarial Nets
Deep Learning for Dual-Energy X-Ray
Generative Adversarial Network (GAN)
Generative Adversarial Imitation Learning
Environment Generation with GANs
Deep Neural Net Scenery Generation
Automatic Lung Cancer Diagnosis from CT Scans (Week 2)
Deep Reinforcement Learning
Cascade of multi-scale convolutional neural networks for bone suppression of chest radiographs in gradient domain  Yang Wei , Chen Yingyin , Liu Yunbi.
Yuanke Zhang1,2, Hongbing Lu1, Junyan Rong1, Yuxiang Xing3, Jing Meng2
Article Review Todd Hricik.
Perceptual Loss Deep Feature Interpolation for Image Content Changes
Generative Adversarial Networks
Regularizing Face Verification Nets To Discrete-Valued Pain Regression
Image Quality Assessment on CT Reconstruction
Combining CNN with RNN for scene labeling (segmentation)
Chaoyun Zhang, Xi Ouyang, and Paul Patras
Synthesis of X-ray Projections via Deep Learning
Super-resolution Image Reconstruction
Single Image Super-Resolution
Generative adversarial networks (GANs) for edge detection
Dipartimento di Ingegneria «Enzo Ferrari»
For Monochromatic Imaging
Authors: Jun-Yan Zhu*, Taesun Park*, Phillip Isola, Alexei A. Efros
Presenter: Hajar Emami
Adversarially Tuned Scene Generation
State-of-the-art face recognition systems
Normalized Cut Loss for Weakly-supervised CNN Segmentation
CNNs and compressive sensing Theoretical analysis
Deep CNN of JPEG 2000 電信所R 林俊廷.
Towards Understanding the Invertibility of Convolutional Neural Networks Anna C. Gilbert1, Yi Zhang1, Kibok Lee1, Yuting Zhang1, Honglak Lee1,2 1University.
Goodfellow: Chapter 14 Autoencoders
Image recognition: Defense adversarial attacks
David Healey BYU Capstone Course 15 Nov 2018
Image to Image Translation using GANs
GAN Applications.
Outline Background Motivation Proposed Model Experimental Results
Lip movement Synthesis from Text
Machine Learning / Deep Learning
Machine Learning based Data Analysis
Machine learning overview
Textual Video Prediction
Neural Network Pipeline CONTACT & ACKNOWLEDGEMENTS
Compressive Image Recovery using Recurrent Generative Model
Course Recap and What’s Next?
Abnormally Detection
Department of Computer Science Ben-Gurion University of the Negev
Ch 14. Generative adversarial networks (GANs) for edge detection
Chuan Wang1, Haibin Huang1, Xiaoguang Han2, Jue Wang1
Image Processing and Multi-domain Translation
Angel A. Cantu, Nami Akazawa Department of Computer Science
Learning Deconvolution Network for Semantic Segmentation
Deep screen image crop and enhance
End-to-End Facial Alignment and Recognition
Cengizhan Can Phoebe de Nooijer
Text-to-speech (TTS) Traditional approaches (before 2016) Neural TTS
End-to-End Speech-Driven Facial Animation with Temporal GANs
SDSEN: Self-Refining Deep Symmetry Enhanced Network
Deep screen image crop and enhance
Directional Occlusion with Neural Network
Shengcong Chen, Changxing Ding, Minfeng Liu 2018
Presentation transcript:

Low Dose CT Image Denoising Using WGAN and Perceptual Loss Qingsong Yang, Pingkun Yan, Ge Wang Biomedical Imaging Center, CBIS/BME, RPI yangq4@rpi.edu Nov 19, 2017

Statistical Reconstruction Low Dose CT Low Dose FBP Recon Full Dose Statistical Recon Reduced X-ray Exposure Increased Noise and Artifacts Noise Suppressing Methods: Sinogram Reconstructed Image FBP Statistical Reconstruction

General flowchart of denoising networking training process Deep Neural Network G Low dose CT image Difference/Errors Full dose/Noiseless CT image General flowchart of denoising networking training process

Normal Dose Low Dose ASD-POCS KSVD BM3D CNN 1 red arrow indicates a small structural detail, maybe a lesion, maybe a calcification. Only the result from CNN, we can see it. All the other methods smooth it. You can enlarge more. 2 blue arrow indicates a region between two material with high attenuation coefficient. ASD-POCS has blocky effect. None of the methods except CNN can effectively eliminate the streak-like artifacts. 3 green arrow also indicates a region that there is no artifacts in CNN’s result but all other methods still have obvious artifacts. Chen, Hu, et al. "Low-dose CT via convolutional neural network." Biomedical optics express 8.2 (2017): 679-694.

A RED-CNN network using paired convolutional and de-convolutional layers for low dose CT denoising Chen, Hu, et al. "Low-Dose CT with a Residual Encoder-Decoder Convolutional Neural Network (RED-CNN)." arXiv preprint arXiv:1702.00288 (2017).

A wavelet domain deep convolutional neural network architecture for low-dose CT denoising Kang, Eunhee, Junhong Min, and Jong Chul Ye. "A deep convolutional neural network using directional wavelets for low‐dose X‐ray CT reconstruction." Medical Physics 44.10 (2017).

Normal-dose TV-POCS K-SVD BM3D WaveNet RED-CNN Chen, Hu, et al. "Low-Dose CT with a Residual Encoder-Decoder Convolutional Neural Network (RED-CNN)." arXiv preprint arXiv:1702.00288 (2017).

MSE Loss: pixel-wise errors Perceptual Loss: errors in a defined feature space Deep Neural Network G Low dose CT image Difference/Errors Full dose/Noiseless CT image

Proposed by Visual Geometry Group, University of Oxford Very deep neural network Trained on natural images for image classification VGG-19 network Network structure with perceptual loss.

Network Training MGH dataset Training dataset VGG extractor GE Discovery CT750HD Over 40 Cadavers’ body CT volumes Four noise levels: 10NI, 20NI, 30NI, and 40NI Three reconstruction algorithms: FBP, ASIR and VEO Training dataset Inputs: FBP30NI Labels: VEO10NI Over 10,000 Image patches 80x80 VGG extractor VGG_11, VGG_31, VGG_34 FBP30NI VEO10NI

Comparison of zoomed ROI FBP30NI VEO30NI CNN-MSE Comparison of zoomed ROI CNN-VGG11 CNN-VGG31 CNN-VGG34 Using perceptual can avoid oversmoothing Deep VGG layer capture more details An example of denoising results using MSE loss and different layers of VGG network as feature extractors

Generative Adversarial Network - GAN A game between two players: Discriminator D Generator G D tries to discriminate between: A sample from the real data A sample from the generated data G tries to “trick” D by generating samples that are hard for D to distinguish from real data Goodfellow, Ian, et al. "Generative adversarial nets." Advances in neural information processing systems. 2014.

Wasserstein GAN - WGAN Pitfall of GAN WGAN No guarantee to equilibrium The discriminator only gives 0 or 1 but cannot describe how good or bad the image is https://github.com/soumith/ganhacks WGAN Wasserstein distance between two data distributions The discriminator gives a continuous evaluation describe how good or bad the image is Arjovsky, Martin, and Léon Bottou. "Towards principled methods for training generative adversarial networks." arXiv preprint arXiv:1701.04862 (2017). M. Arjovsky, S. Chintala, and L. Bottou, “Wasserstein gan,” arXiv preprint arXiv:1701.07875, 2017. I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. Courville, “Improved training of wasserstein gans,” arXiv preprint arXiv:1704.00028, 2017.

CNN-MSE CNN-VGG WGAN-MSE WGAN-VGG WGAN Overall structure of the denoising network.

Network Training MGH dataset Training dataset Networks: GE Discovery CT750HD Over 40 Cadavers’ body CT volumes Four noise levels: 10NI, 20NI, 30NI, and 40NI Three reconstruction algorithms: FBP, ASIR and VEO Training dataset Inputs: FBP30NI Labels: VEO10NI Over 10,000 Image patches 80x80 Networks: CNN-MSE / CNN-VGG WGAN-MSE / WGAN-VGG / WGAN FBP30NI VEO10NI

Comparison of zoomed in ROI FBP30NI VEO30NI CNN-MSE CNN-VGG WGAN WGAN-VGG FBP30NI VEO30NI CNN-MSE CNN-VGG WGAN WGAN-VGG Comparison of zoomed in ROI An example of denoising results using different loss functions

Network Training Mayo data Training dataset Networks: Two noise levels: full dose and simulated quarter dose FBP reconstruction Training dataset Inputs: quarter dose images Labels: full dose Over 10,000 Image patches 80x80 Networks: CNN-MSE / CNN-VGG WGAN-MSE / WGAN-VGG / WGAN Quarter dose Full dose AAPM, “Low dose ct grand challenge,” 2017. [Online]. Available: http://www.aapm.org/GrandChallenge/LowDoseCT/#

An example of denoisng result using different loss functions Full Dose Quarter Dose CNN-MSE PSNR SSIM ROI Mean (HU) ROI Variance (HU) Full Dose 9 36 Quarter Dose 19.7904 0.7496 11 74 CNN-MSE 24.4894 0.7966 12 18 CNN-VGG 23.2322 0.7926 4 30 WGAN-MSE 24.0637 0.8090 28 WGAN-VGG 23.3942 0.7923 31 WGAN 22.0168 0.7745 23 37 CNN-VGG WGAN-MSE WGAN-VGG WGAN Quantitative analysis using PSNR and SSIM and statistical properties of a small ROI An example of denoisng result using different loss functions

An example of denoisng result using different loss functions Full Dose Quarter Dose CNN-MSE PSNR SSIM ROI Mean (HU) ROI Variance (HU) Full Dose 9 36 Quarter Dose 18.4519 0.6471 118 38 CNN-MSE 23.2649 0.7022 120 15 CNN-VGG 22.0950 0.6972 104 28 WGAN-MSE 22.7255 0.7122 115 25 WGAN-VGG 22.1620 0.6759 111 29 WGAN 20.9051 135 33 CNN-VGG WGAN-MSE WGAN-VGG WGAN Quantitative analysis using PSNR and SSIM and statistical properties of a small ROI An example of denoisng result using different loss functions

Summary Simple Network Structure Perceptual Loss - Image Content WGAN Framework – Data Distribution Outlook - Evaluation