19 research outputs found
Generating Diffusion MRI scalar maps from T1 weighted images using generative adversarial networks
Diffusion magnetic resonance imaging (diffusion MRI) is a non-invasive
microstructure assessment technique. Scalar measures, such as FA (fractional
anisotropy) and MD (mean diffusivity), quantifying micro-structural tissue
properties can be obtained using diffusion models and data processing
pipelines. However, it is costly and time consuming to collect high quality
diffusion data. Here, we therefore demonstrate how Generative Adversarial
Networks (GANs) can be used to generate synthetic diffusion scalar measures
from structural T1-weighted images in a single optimized step. Specifically, we
train the popular CycleGAN model to learn to map a T1 image to FA or MD, and
vice versa. As an application, we show that synthetic FA images can be used as
a target for non-linear registration, to correct for geometric distortions
common in diffusion MRI
Disentangled Uncertainty and Out of Distribution Detection in Medical Generative Models
Trusting the predictions of deep learning models in safety critical settings such as the medical domain is still not a viable option. Distentangled uncertainty quantification in the field of medical imaging has received little attention. In this paper, we study disentangled uncertainties in image to image translation tasks in the medical domain. We compare multiple uncertainty quantification methods, namely Ensembles, Flipout, Dropout, and DropConnect, while using CycleGAN to convert T1-weighted brain MRI scans to T2-weighted brain MRI scans. We further evaluate uncertainty behavior in the presence of out of distribution data (Brain CT and RGB Face Images), showing that epistemic uncertainty can be used to detect out of distribution inputs, which should increase reliability of model outputs