24 research outputs found
Learning to Address Intra-segment Misclassification in Retinal Imaging
Accurate multi-class segmentation is a long-standing challenge in medical imaging, especially in scenarios where classes share strong similarity. Segmenting retinal blood vessels in retinal photographs is one such scenario, in which arteries and veins need to be identified and differentiated from each other and from the background. Intra-segment misclassification, i.e. veins classified as arteries or vice versa, frequently occurs when arteries and veins intersect, whereas in binary retinal vessel segmentation, error rates are much lower. We thus propose a new approach that decomposes multi-class segmentation into multiple binary, followed by a binary-to-multi-class fusion network. The network merges representations of artery, vein, and multi-class feature maps, each of which are supervised by expert vessel annotation in adversarial training. A skip-connection based merging process explicitly maintains class-specific gradients to avoid gradient vanishing in deep layers, to favor the discriminative features. The results show that, our model respectively improves F1-score by 4.4%, 5.1%, and 4.2% compared with three state-of-the-art deep learning based methods on DRIVE-AV, LES-AV, and HRF-AV data sets. Code: https://github.com/rmaphoh/Learning-AVSegmentatio
Learning to Address Intra-segment Misclassification in Retinal Imaging
Accurate multi-class segmentation is a long-standing challenge in medical imaging, especially in scenarios where classes share strong similarity. Segmenting retinal blood vessels in retinal photographs is one such scenario, in which arteries and veins need to be identified and differentiated from each other and from the background. Intra-segment misclassification, i.e. veins classified as arteries or vice versa, frequently occurs when arteries and veins intersect, whereas in binary retinal vessel segmentation, error rates are much lower. We thus propose a new approach that decomposes multi-class segmentation into multiple binary, followed by a binary-to-multi-class fusion network. The network merges representations of artery, vein, and multi-class feature maps, each of which are supervised by expert vessel annotation in adversarial training. A skip-connection based merging process explicitly maintains class-specific gradients to avoid gradient vanishing in deep layers, to favor the discriminative features. The results show that, our model respectively improves F1-score by 4.4%, 5.1%, and 4.2% compared with three state-of-the-art deep learning based methods on DRIVE-AV, LES-AV, and HRF-AV data sets. Code: https://github.com/rmaphoh/Learning-AVSegmentatio
EDDense-Net: Fully Dense Encoder Decoder Network for Joint Segmentation of Optic Cup and Disc
Glaucoma is an eye disease that causes damage to the optic nerve, which can
lead to visual loss and permanent blindness. Early glaucoma detection is
therefore critical in order to avoid permanent blindness. The estimation of the
cup-to-disc ratio (CDR) during an examination of the optical disc (OD) is used
for the diagnosis of glaucoma. In this paper, we present the EDDense-Net
segmentation network for the joint segmentation of OC and OD. The encoder and
decoder in this network are made up of dense blocks with a grouped
convolutional layer in each block, allowing the network to acquire and convey
spatial information from the image while simultaneously reducing the network's
complexity. To reduce spatial information loss, the optimal number of filters
in all convolution layers were utilised. In semantic segmentation, dice pixel
classification is employed in the decoder to alleviate the problem of class
imbalance. The proposed network was evaluated on two publicly available
datasets where it outperformed existing state-of-the-art methods in terms of
accuracy and efficiency. For the diagnosis and analysis of glaucoma, this
method can be used as a second opinion system to assist medical
ophthalmologists
Recommended from our members
From Fully-Supervised, Single-Task to Scarcely-Supervised, Multi-Task Deep Learning for Medical Image Analysis
Image analysis based on machine learning has gained prominence with the advent of deep learning, particularly in medical imaging. To be effective in addressing challenging image analysis tasks, however, conventional deep neural networks require large corpora of annotated training data, which are unfortunately scarce in the medical domain, thus often rendering fully-supervised learning strategies ineffective.This thesis devises for use in a variety of medical image analysis applications a series of novel deep learning methods, ranging from fully-supervised, single-task learning to scarcely-supervised, multi-task learning that makes efficient use of annotated training data. Specifically, its main contributions include (1) fully-supervised, single-task learning for the segmentation of pulmonary lobes from chest CT scans and the analysis of scoliosis from spine X-ray images; (2) supervised, single-task, domain-generalized pulmonary segmentation in chest X-ray images and retinal vasculature segmentation in fundoscopic images; (3) largely-unsupervised, multiple-task learning via deep generative modeling for the joint synthesis and classification of medical image data; and (4) partly-supervised, multiple-task learning for the combined segmentation and classification of chest and spine X-ray images
Improving foveal avascular zone segmentation in fluorescein angiograms by leveraging manual vessel labels from public color fundus pictures
In clinical routine, ophthalmologists frequently analyze the shape and size of the foveal avascular zone (FAZ) to detect and monitor retinal diseases. In order to extract those parameters, the contours of the FAZ need to be segmented, which is normally achieved by analyzing the retinal vasculature (RV) around the macula in fluorescein angiograms (FA). Computer-aided segmentation methods based on deep learning (DL) can automate this task. However, current approaches for segmenting the FAZ are often tailored to a specific dataset or require manual initialization. Furthermore, they do not take the variability and challenges of clinical FA into account, which are often of low quality and difficult to analyze. In this paper we propose a DL-based framework to automatically segment the FAZ in challenging FA scans from clinical routine. Our approach mimics the workflow of retinal experts by using additional RV labels as a guidance during training. Hence, our model is able to produce RV segmentations simultaneously. We minimize the annotation work by using a multi-modal approach that leverages already available public datasets of color fundus pictures (CFPs) and their respective manual RV labels. Our experimental evaluation on two datasets with FA from 1) clinical routine and 2) large multicenter clinical trials shows that the addition of weak RV labels as a guidance during training improves the FAZ segmentation significantly with respect to using only manual FAZ annotations.Fil: Hofer, Dominik. Medizinische Universität Wien; AustriaFil: Schmidt Erfurth, Ursula. Medizinische Universität Wien; AustriaFil: Orlando, JosĂ© Ignacio. Universidad Nacional del Centro de la Provincia de Buenos Aires. Facultad de Ciencias Exactas. Grupo de Plasmas Densos Magnetizados. Provincia de Buenos Aires. GobernaciĂłn. Comision de Investigaciones CientĂficas. Grupo de Plasmas Densos Magnetizados; Argentina. Medizinische Universität Wien; AustriaFil: Goldbach, Felix. Medizinische Universität Wien; AustriaFil: Gerendas, Bianca S.. Medizinische Universität Wien; AustriaFil: Seeböck, Philipp. Medizinische Universität Wien; Austri
The Little W-Net That Could: State-of-the-Art Retinal Vessel Segmentation with Minimalistic Models
The segmentation of the retinal vasculature from eye fundus images represents
one of the most fundamental tasks in retinal image analysis. Over recent years,
increasingly complex approaches based on sophisticated Convolutional Neural
Network architectures have been slowly pushing performance on well-established
benchmark datasets. In this paper, we take a step back and analyze the real
need of such complexity. Specifically, we demonstrate that a minimalistic
version of a standard U-Net with several orders of magnitude less parameters,
carefully trained and rigorously evaluated, closely approximates the
performance of current best techniques. In addition, we propose a simple
extension, dubbed W-Net, which reaches outstanding performance on several
popular datasets, still using orders of magnitude less learnable weights than
any previously published approach. Furthermore, we provide the most
comprehensive cross-dataset performance analysis to date, involving up to 10
different databases. Our analysis demonstrates that the retinal vessel
segmentation problem is far from solved when considering test images that
differ substantially from the training data, and that this task represents an
ideal scenario for the exploration of domain adaptation techniques. In this
context, we experiment with a simple self-labeling strategy that allows us to
moderately enhance cross-dataset performance, indicating that there is still
much room for improvement in this area. Finally, we also test our approach on
the Artery/Vein segmentation problem, where we again achieve results
well-aligned with the state-of-the-art, at a fraction of the model complexity
in recent literature. All the code to reproduce the results in this paper is
released