4 research outputs found
PMP-Swin: Multi-Scale Patch Message Passing Swin Transformer for Retinal Disease Classification
Retinal disease is one of the primary causes of visual impairment, and early
diagnosis is essential for preventing further deterioration. Nowadays, many
works have explored Transformers for diagnosing diseases due to their strong
visual representation capabilities. However, retinal diseases exhibit milder
forms and often present with overlapping signs, which pose great difficulties
for accurate multi-class classification. Therefore, we propose a new framework
named Multi-Scale Patch Message Passing Swin Transformer for multi-class
retinal disease classification. Specifically, we design a Patch Message Passing
(PMP) module based on the Message Passing mechanism to establish global
interaction for pathological semantic features and to exploit the subtle
differences further between different diseases. Moreover, considering the
various scale of pathological features we integrate multiple PMP modules for
different patch sizes. For evaluation, we have constructed a new dataset, named
OPTOS dataset, consisting of 1,033 high-resolution fundus images photographed
by Optos camera and conducted comprehensive experiments to validate the
efficacy of our proposed method. And the results on both the public dataset and
our dataset demonstrate that our method achieves remarkable performance
compared to state-of-the-art methods.Comment: 9 pages, 7 figure
Enhancing Point Annotations with Superpixel and Confidence Learning Guided for Improving Semi-Supervised OCT Fluid Segmentation
Automatic segmentation of fluid in Optical Coherence Tomography (OCT) images
is beneficial for ophthalmologists to make an accurate diagnosis. Although
semi-supervised OCT fluid segmentation networks enhance their performance by
introducing additional unlabeled data, the performance enhancement is limited.
To address this, we propose Superpixel and Confident Learning Guide Point
Annotations Network (SCLGPA-Net) based on the teacher-student architecture,
which can learn OCT fluid segmentation from limited fully-annotated data and
abundant point-annotated data. Specifically, we use points to annotate fluid
regions in unlabeled OCT images and the Superpixel-Guided Pseudo-Label
Generation (SGPLG) module generates pseudo-labels and pixel-level label trust
maps from the point annotations. The label trust maps provide an indication of
the reliability of the pseudo-labels. Furthermore, we propose the Confident
Learning Guided Label Refinement (CLGLR) module identifies error information in
the pseudo-labels and leads to further refinement. Experiments on the RETOUCH
dataset show that we are able to reduce the need for fully-annotated data by
94.22\%, closing the gap with the best fully supervised baselines to a mean IoU
of only 2\%. Furthermore, We constructed a private 2D OCT fluid segmentation
dataset for evaluation. Compared with other methods, comprehensive experimental
results demonstrate that the proposed method can achieve excellent performance
in OCT fluid segmentation.Comment: Submission to BSP
MSE-Nets: Multi-annotated Semi-supervised Ensemble Networks for Improving Segmentation of Medical Image with Ambiguous Boundaries
Medical image segmentation annotations exhibit variations among experts due
to the ambiguous boundaries of segmented objects and backgrounds in medical
images. Although using multiple annotations for each image in the
fully-supervised has been extensively studied for training deep models,
obtaining a large amount of multi-annotated data is challenging due to the
substantial time and manpower costs required for segmentation annotations,
resulting in most images lacking any annotations. To address this, we propose
Multi-annotated Semi-supervised Ensemble Networks (MSE-Nets) for learning
segmentation from limited multi-annotated and abundant unannotated data.
Specifically, we introduce the Network Pairwise Consistency Enhancement (NPCE)
module and Multi-Network Pseudo Supervised (MNPS) module to enhance MSE-Nets
for the segmentation task by considering two major factors: (1) to optimize the
utilization of all accessible multi-annotated data, the NPCE separates
(dis)agreement annotations of multi-annotated data at the pixel level and
handles agreement and disagreement annotations in different ways, (2) to
mitigate the introduction of imprecise pseudo-labels, the MNPS extends the
training data by leveraging consistent pseudo-labels from unannotated data.
Finally, we improve confidence calibration by averaging the predictions of base
networks. Experiments on the ISIC dataset show that we reduced the demand for
multi-annotated data by 97.75\% and narrowed the gap with the best
fully-supervised baseline to just a Jaccard index of 4\%. Furthermore, compared
to other semi-supervised methods that rely only on a single annotation or a
combined fusion approach, the comprehensive experimental results on ISIC and
RIGA datasets demonstrate the superior performance of our proposed method in
medical image segmentation with ambiguous boundaries
Accurate Segmentation of Optic Disc And Cup from Multiple Pseudo-labels by Noise-aware Learning
Optic disc and cup segmentation plays a crucial role in automating the
screening and diagnosis of optic glaucoma. While data-driven convolutional
neural networks (CNNs) show promise in this area, the inherent ambiguity of
segmenting objects and background boundaries in the task of optic disc and cup
segmentation leads to noisy annotations that impact model performance. To
address this, we propose an innovative label-denoising method of Multiple
Pseudo-labels Noise-aware Network (MPNN) for accurate optic disc and cup
segmentation. Specifically, the Multiple Pseudo-labels Generation and Guided
Denoising (MPGGD) module generates pseudo-labels by multiple different
initialization networks trained on true labels, and the pixel-level consensus
information extracted from these pseudo-labels guides to differentiate clean
pixels from noisy pixels. The training framework of the MPNN is constructed by
a teacher-student architecture to learn segmentation from clean pixels and
noisy pixels. Particularly, such a framework adeptly leverages (i) reliable and
fundamental insight from clean pixels and (ii) the supplementary knowledge
within noisy pixels via multiple perturbation-based unsupervised consistency.
Compared to other label-denoising methods, comprehensive experimental results
on the RIGA dataset demonstrate our method's excellent performance. The code is
available at https://github.com/wwwtttjjj/MPNNComment: CSCWD 202