14,324 research outputs found
Recurrent Fully Convolutional Neural Networks for Multi-slice MRI Cardiac Segmentation
In cardiac magnetic resonance imaging, fully-automatic segmentation of the
heart enables precise structural and functional measurements to be taken, e.g.
from short-axis MR images of the left-ventricle. In this work we propose a
recurrent fully-convolutional network (RFCN) that learns image representations
from the full stack of 2D slices and has the ability to leverage inter-slice
spatial dependences through internal memory units. RFCN combines anatomical
detection and segmentation into a single architecture that is trained
end-to-end thus significantly reducing computational time, simplifying the
segmentation pipeline, and potentially enabling real-time applications. We
report on an investigation of RFCN using two datasets, including the publicly
available MICCAI 2009 Challenge dataset. Comparisons have been carried out
between fully convolutional networks and deep restricted Boltzmann machines,
including a recurrent version that leverages inter-slice spatial correlation.
Our studies suggest that RFCN produces state-of-the-art results and can
substantially improve the delineation of contours near the apex of the heart.Comment: MICCAI Workshop RAMBO 201
A Deep Cascade of Convolutional Neural Networks for MR Image Reconstruction
The acquisition of Magnetic Resonance Imaging (MRI) is inherently slow.
Inspired by recent advances in deep learning, we propose a framework for
reconstructing MR images from undersampled data using a deep cascade of
convolutional neural networks to accelerate the data acquisition process. We
show that for Cartesian undersampling of 2D cardiac MR images, the proposed
method outperforms the state-of-the-art compressed sensing approaches, such as
dictionary learning-based MRI (DLMRI) reconstruction, in terms of
reconstruction error, perceptual quality and reconstruction speed for both
3-fold and 6-fold undersampling. Compared to DLMRI, the error produced by the
method proposed is approximately twice as small, allowing to preserve
anatomical structures more faithfully. Using our method, each image can be
reconstructed in 23 ms, which is fast enough to enable real-time applications
Motion processing deficits in migraine are related to contrast sensitivity
Background: There are conflicting reports concerning the ability of people with migraine to detect and discriminate visual motion. Previous studies used different displays and none adequately assessed other parameters that could affect performance, such as those that could indicate precortical dysfunction.
Methods: Motion-direction detection, discrimination and relative motion thresholds were compared from participants with and without migraine. Potentially relevant visual covariates were included (contrast sensitivity; acuity; stereopsis; visual discomfort, stress, triggers; dyslexia).
Results: For each task, migraine participants were less accurate than a control group and had impaired contrast sensitivity, greater visual discomfort, visual stress and visual triggers. Only contrast sensitivity correlated with performance on each motion task; it also mediated performance.
Conclusions: Impaired performance on certain motion tasks can be attributed to impaired contrast sensitivity early in the visual system rather than a deficit in cortical motion processing per se. There were, however, additional differences for global and relative motion thresholds embedded in noise, suggesting changes in extrastriate cortex in migraine. Tasks to study the effects of noise on performance at different levels of the visual system and across modalities are recommended. A battery of standard visual tests should be included in any future work on the visual system and migraine
Segmentation of Intracranial Arterial Calcification with Deeply Supervised Residual Dropout Networks
Intracranial carotid artery calcification (ICAC) is a major risk factor for
stroke, and might contribute to dementia and cognitive decline. Reliance on
time-consuming manual annotation of ICAC hampers much demanded further research
into the relationship between ICAC and neurological diseases. Automation of
ICAC segmentation is therefore highly desirable, but difficult due to the
proximity of the lesions to bony structures with a similar attenuation
coefficient. In this paper, we propose a method for automatic segmentation of
ICAC; the first to our knowledge. Our method is based on a 3D fully
convolutional neural network that we extend with two regularization techniques.
Firstly, we use deep supervision (hidden layers supervision) to encourage
discriminative features in the hidden layers. Secondly, we augment the network
with skip connections, as in the recently developed ResNet, and dropout layers,
inserted in a way that skip connections circumvent them. We investigate the
effect of skip connections and dropout. In addition, we propose a simple
problem-specific modification of the network objective function that restricts
the focus to the most important image regions and simplifies the optimization.
We train and validate our model using 882 CT scans and test on 1,000. Our
regularization techniques and objective improve the average Dice score by 7.1%,
yielding an average Dice of 76.2% and 97.7% correlation between predicted ICAC
volumes and manual annotations.Comment: Accepted for MICCAI 201
Sub-token ViT Embedding via Stochastic Resonance Transformers
We discover the presence of quantization artifacts in Vision Transformers
(ViTs), which arise due to the image tokenization step inherent in these
architectures. These artifacts result in coarsely quantized features, which
negatively impact performance, especially on downstream dense prediction tasks.
We present a zero-shot method to improve how pre-trained ViTs handle spatial
quantization. In particular, we propose to ensemble the features obtained from
perturbing input images via sub-token spatial translations, inspired by
Stochastic Resonance, a method traditionally applied to climate dynamics and
signal processing. We term our method ``Stochastic Resonance Transformer"
(SRT), which we show can effectively super-resolve features of pre-trained
ViTs, capturing more of the local fine-grained structures that might otherwise
be neglected as a result of tokenization. SRT can be applied at any layer, on
any task, and does not require any fine-tuning. The advantage of the former is
evident when applied to monocular depth prediction, where we show that
ensembling model outputs are detrimental while applying SRT on intermediate ViT
features outperforms the baseline models by an average of 4.7% and 14.9% on the
RMSE and RMSE-log metrics across three different architectures. When applied to
semi-supervised video object segmentation, SRT also improves over the baseline
models uniformly across all metrics, and by an average of 2.4% in F&J score. We
further show that these quantization artifacts can be attenuated to some extent
via self-distillation. On the unsupervised salient region segmentation, SRT
improves upon the base model by an average of 2.1% on the maxF metric. Finally,
despite operating purely on pixel-level features, SRT generalizes to non-dense
prediction tasks such as image retrieval and object discovery, yielding
consistent improvements of up to 2.6% and 1.0% respectively
- …