1,677 research outputs found
A Deep Learning Approach to Denoise Optical Coherence Tomography Images of the Optic Nerve Head
Purpose: To develop a deep learning approach to de-noise optical coherence
tomography (OCT) B-scans of the optic nerve head (ONH).
Methods: Volume scans consisting of 97 horizontal B-scans were acquired
through the center of the ONH using a commercial OCT device (Spectralis) for
both eyes of 20 subjects. For each eye, single-frame (without signal
averaging), and multi-frame (75x signal averaging) volume scans were obtained.
A custom deep learning network was then designed and trained with 2,328 "clean
B-scans" (multi-frame B-scans), and their corresponding "noisy B-scans" (clean
B-scans + gaussian noise) to de-noise the single-frame B-scans. The performance
of the de-noising algorithm was assessed qualitatively, and quantitatively on
1,552 B-scans using the signal to noise ratio (SNR), contrast to noise ratio
(CNR), and mean structural similarity index metrics (MSSIM).
Results: The proposed algorithm successfully denoised unseen single-frame OCT
B-scans. The denoised B-scans were qualitatively similar to their corresponding
multi-frame B-scans, with enhanced visibility of the ONH tissues. The mean SNR
increased from dB (single-frame) to dB
(denoised). For all the ONH tissues, the mean CNR increased from (single-frame) to (denoised). The MSSIM increased from
(single frame) to (denoised) when compared with
the corresponding multi-frame B-scans.
Conclusions: Our deep learning algorithm can denoise a single-frame OCT
B-scan of the ONH in under 20 ms, thus offering a framework to obtain superior
quality OCT B-scans with reduced scanning times and minimal patient discomfort
MedGAN: Medical Image Translation using GANs
Image-to-image translation is considered a new frontier in the field of
medical image analysis, with numerous potential applications. However, a large
portion of recent approaches offers individualized solutions based on
specialized task-specific architectures or require refinement through
non-end-to-end training. In this paper, we propose a new framework, named
MedGAN, for medical image-to-image translation which operates on the image
level in an end-to-end manner. MedGAN builds upon recent advances in the field
of generative adversarial networks (GANs) by merging the adversarial framework
with a new combination of non-adversarial losses. We utilize a discriminator
network as a trainable feature extractor which penalizes the discrepancy
between the translated medical images and the desired modalities. Moreover,
style-transfer losses are utilized to match the textures and fine-structures of
the desired target images to the translated images. Additionally, we present a
new generator architecture, titled CasNet, which enhances the sharpness of the
translated medical outputs through progressive refinement via encoder-decoder
pairs. Without any application-specific modifications, we apply MedGAN on three
different tasks: PET-CT translation, correction of MR motion artefacts and PET
image denoising. Perceptual analysis by radiologists and quantitative
evaluations illustrate that the MedGAN outperforms other existing translation
approaches.Comment: 16 pages, 8 figure
Combined Denoising and Suppression of Transient Artifacts in Arterial Spin Labeling MRI Using Deep Learning
Background: Arterial spin labeling (ASL) is a useful tool for measuring cerebral blood flow (CBF). However, due to the low signal-to-noise ratio (SNR) of the technique, multiple repetitions are required, which results in prolonged scan times and increased susceptibility to artifacts. Purpose: To develop a deep-learning-based algorithm for simultaneous denoising and suppression of transient artifacts in ASL images. Study Type: Retrospective. Subjects: 131 pediatric neuro-oncology patients for model training and 11 healthy adult subjects for model evaluation. Field Strength/Sequence: 3T / pseudo-continuous and pulsed ASL with 3D gradient-and-spin-echo readout. Assessment: A denoising autoencoder (DAE) model was designed with stacked encoding/decoding convolutional layers. Reference standard images were generated by averaging 10 pairwise ASL subtraction images. The model was trained to produce perfusion images of a similar quality using a single subtraction image. Performance was compared against Gaussian and non-local means (NLM) filters. Evaluation metrics included SNR, peak SNR (PSNR), and structural similarity index (SSIM) of the CBF images, compared to the reference standard. Statistical Tests: One-way analysis of variance (ANOVA) tests for group comparisons. Results: The DAE model was the only model to produce a significant increase in SNR compared to the raw images (P < 0.05), providing an average SNR gain of 62%. The DAE model was also effective at suppressing transient artifacts, and was the only model to show a significant improvement in accuracy in the generated CBF images, as assessed using PSNR values (P < 0.05). In addition, using data from multiple inflow time acquisitions, the DAE images produced the best fit to the Buxton kinetic model, offering a 75% reduction in the fitting error compared to the raw images. Data Conclusion: Deep-learning-based algorithms provide superior accuracy when denoising ASL images, due to their ability to simultaneously increase SNR and suppress artifactual signals in raw ASL images. Level of Evidence: 3. Technical Efficacy Stage: 1
Multi-modal Image Processing based on Coupled Dictionary Learning
In real-world scenarios, many data processing problems often involve
heterogeneous images associated with different imaging modalities. Since these
multimodal images originate from the same phenomenon, it is realistic to assume
that they share common attributes or characteristics. In this paper, we propose
a multi-modal image processing framework based on coupled dictionary learning
to capture similarities and disparities between different image modalities. In
particular, our framework can capture favorable structure similarities across
different image modalities such as edges, corners, and other elementary
primitives in a learned sparse transform domain, instead of the original pixel
domain, that can be used to improve a number of image processing tasks such as
denoising, inpainting, or super-resolution. Practical experiments demonstrate
that incorporating multimodal information using our framework brings notable
benefits.Comment: SPAWC 2018, 19th IEEE International Workshop On Signal Processing
Advances In Wireless Communication
- …