2,550 research outputs found
Brain tumor segmentation with missing modalities via latent multi-source correlation representation
Multimodal MR images can provide complementary information for accurate brain
tumor segmentation. However, it's common to have missing imaging modalities in
clinical practice. Since there exists a strong correlation between multi
modalities, a novel correlation representation block is proposed to specially
discover the latent multi-source correlation. Thanks to the obtained
correlation representation, the segmentation becomes more robust in the case of
missing modalities. The model parameter estimation module first maps the
individual representation produced by each encoder to obtain independent
parameters, then, under these parameters, the correlation expression module
transforms all the individual representations to form a latent multi-source
correlation representation. Finally, the correlation representations across
modalities are fused via the attention mechanism into a shared representation
to emphasize the most important features for segmentation. We evaluate our
model on BraTS 2018 datasets, it outperforms the current state-of-the-art
method and produces robust results when one or more modalities are missing.Comment: 9 pages, 6 figures, accepted by MICCAI 202
CKD-TransBTS: Clinical Knowledge-Driven Hybrid Transformer with Modality-Correlated Cross-Attention for Brain Tumor Segmentation
Brain tumor segmentation (BTS) in magnetic resonance image (MRI) is crucial
for brain tumor diagnosis, cancer management and research purposes. With the
great success of the ten-year BraTS challenges as well as the advances of CNN
and Transformer algorithms, a lot of outstanding BTS models have been proposed
to tackle the difficulties of BTS in different technical aspects. However,
existing studies hardly consider how to fuse the multi-modality images in a
reasonable manner. In this paper, we leverage the clinical knowledge of how
radiologists diagnose brain tumors from multiple MRI modalities and propose a
clinical knowledge-driven brain tumor segmentation model, called CKD-TransBTS.
Instead of directly concatenating all the modalities, we re-organize the input
modalities by separating them into two groups according to the imaging
principle of MRI. A dual-branch hybrid encoder with the proposed
modality-correlated cross-attention block (MCCA) is designed to extract the
multi-modality image features. The proposed model inherits the strengths from
both Transformer and CNN with the local feature representation ability for
precise lesion boundaries and long-range feature extraction for 3D volumetric
images. To bridge the gap between Transformer and CNN features, we propose a
Trans&CNN Feature Calibration block (TCFC) in the decoder. We compare the
proposed model with five CNN-based models and six transformer-based models on
the BraTS 2021 challenge dataset. Extensive experiments demonstrate that the
proposed model achieves state-of-the-art brain tumor segmentation performance
compared with all the competitors
ISA-Net: Improved spatial attention network for PET-CT tumor segmentation
Achieving accurate and automated tumor segmentation plays an important role
in both clinical practice and radiomics research. Segmentation in medicine is
now often performed manually by experts, which is a laborious, expensive and
error-prone task. Manual annotation relies heavily on the experience and
knowledge of these experts. In addition, there is much intra- and interobserver
variation. Therefore, it is of great significance to develop a method that can
automatically segment tumor target regions. In this paper, we propose a deep
learning segmentation method based on multimodal positron emission
tomography-computed tomography (PET-CT), which combines the high sensitivity of
PET and the precise anatomical information of CT. We design an improved spatial
attention network(ISA-Net) to increase the accuracy of PET or CT in detecting
tumors, which uses multi-scale convolution operation to extract feature
information and can highlight the tumor region location information and
suppress the non-tumor region location information. In addition, our network
uses dual-channel inputs in the coding stage and fuses them in the decoding
stage, which can take advantage of the differences and complementarities
between PET and CT. We validated the proposed ISA-Net method on two clinical
datasets, a soft tissue sarcoma(STS) and a head and neck tumor(HECKTOR)
dataset, and compared with other attention methods for tumor segmentation. The
DSC score of 0.8378 on STS dataset and 0.8076 on HECKTOR dataset show that
ISA-Net method achieves better segmentation performance and has better
generalization. Conclusions: The method proposed in this paper is based on
multi-modal medical image tumor segmentation, which can effectively utilize the
difference and complementarity of different modes. The method can also be
applied to other multi-modal data or single-modal data by proper adjustment
- …