488 research outputs found
Multimodal Spatial Attention Module for Targeting Multimodal PET-CT Lung Tumor Segmentation
Multimodal positron emission tomography-computed tomography (PET-CT) is used
routinely in the assessment of cancer. PET-CT combines the high sensitivity for
tumor detection with PET and anatomical information from CT. Tumor segmentation
is a critical element of PET-CT but at present, there is not an accurate
automated segmentation method. Segmentation tends to be done manually by
different imaging experts and it is labor-intensive and prone to errors and
inconsistency. Previous automated segmentation methods largely focused on
fusing information that is extracted separately from the PET and CT modalities,
with the underlying assumption that each modality contains complementary
information. However, these methods do not fully exploit the high PET tumor
sensitivity that can guide the segmentation. We introduce a multimodal spatial
attention module (MSAM) that automatically learns to emphasize regions (spatial
areas) related to tumors and suppress normal regions with physiologic
high-uptake. The resulting spatial attention maps are subsequently employed to
target a convolutional neural network (CNN) for segmentation of areas with
higher tumor likelihood. Our MSAM can be applied to common backbone
architectures and trained end-to-end. Our experimental results on two clinical
PET-CT datasets of non-small cell lung cancer (NSCLC) and soft tissue sarcoma
(STS) validate the effectiveness of the MSAM in these different cancer types.
We show that our MSAM, with a conventional U-Net backbone, surpasses the
state-of-the-art lung tumor segmentation approach by a margin of 7.6% in Dice
similarity coefficient (DSC)
ISA-Net: Improved spatial attention network for PET-CT tumor segmentation
Achieving accurate and automated tumor segmentation plays an important role
in both clinical practice and radiomics research. Segmentation in medicine is
now often performed manually by experts, which is a laborious, expensive and
error-prone task. Manual annotation relies heavily on the experience and
knowledge of these experts. In addition, there is much intra- and interobserver
variation. Therefore, it is of great significance to develop a method that can
automatically segment tumor target regions. In this paper, we propose a deep
learning segmentation method based on multimodal positron emission
tomography-computed tomography (PET-CT), which combines the high sensitivity of
PET and the precise anatomical information of CT. We design an improved spatial
attention network(ISA-Net) to increase the accuracy of PET or CT in detecting
tumors, which uses multi-scale convolution operation to extract feature
information and can highlight the tumor region location information and
suppress the non-tumor region location information. In addition, our network
uses dual-channel inputs in the coding stage and fuses them in the decoding
stage, which can take advantage of the differences and complementarities
between PET and CT. We validated the proposed ISA-Net method on two clinical
datasets, a soft tissue sarcoma(STS) and a head and neck tumor(HECKTOR)
dataset, and compared with other attention methods for tumor segmentation. The
DSC score of 0.8378 on STS dataset and 0.8076 on HECKTOR dataset show that
ISA-Net method achieves better segmentation performance and has better
generalization. Conclusions: The method proposed in this paper is based on
multi-modal medical image tumor segmentation, which can effectively utilize the
difference and complementarity of different modes. The method can also be
applied to other multi-modal data or single-modal data by proper adjustment
Deep Learning-based Radiomics Framework for Multi-Modality PET-CT Images
Multimodal positron emission tomography - computed tomography (PET-CT) imaging is widely regarded as the imaging modality of choice for cancer management. This is because PET-CT combines the high sensitivity of PET in detecting regions of abnormal functions and the specificity of CT in depicting the underlying anatomy of where the abnormal functions are occurring.
Radiomics is an emerging research field that enables the extraction and analysis of quantitative features from medical images, providing valuable insights into the underlying pathophysiology that cannot be discerned by the naked eyes. This information is capable of assisting decision-making in clinical practice, leading to better personalised treatment planning, patient outcome prediction, and therapy response assessment.
The aim of this thesis is to propose a new deep learning-based radiomics framework for multimodal PET-CT images. The proposed framework comprises of three methods: 1) a tumour segmentation method via a self-supervision enabled false positive and false negative reduction network; 2) a constrained hierarchical multi-modality feature learning is constructed to predict the patient outcome with multimodal PET-CT images; 3) an automatic neural architecture search method to automatically find the optimal network architecture for both patient outcome prediction and tumour segmentation.
Extensive experiments have been conducted on three datasets, including one public soft-tissue sarcomas dataset, one public challenge dataset, and one in-house lung cancer data. The results demonstrated that the proposed methods obtained better performance in all tasks when compared to the state-of-the-art methods
State-of-the-Art and Development Trend of Interventional Ultrasound in China
Interventional ultrasound (IUS) is an important branch of modern minimally invasive medicine that has been widely applied in clinical practice due to its unique techniques and advantages. As a relatively emerging field, IUS has progressed towards standardization, precision, intelligence, and cutting-edge directions alone with more than 40 years of development, which is becoming increasingly important techniques in clinical medicine. This article will briefly review the development and advancement of IUS for diagnosis and treatment in China in the era of precision medicine from the aspects of artificial intelligence, virtual navigation, molecular imaging, and nanotechnology
Teacher-student approach for lung tumor segmentation from mixed-supervised datasets
Purpose: Cancer is among the leading causes of death in the developed world, and lung cancer is the most lethal type. Early detection is crucial for better prognosis, but can be resource intensive to achieve. Automating tasks such as lung tumor localization and segmentation in radiological images can free valuable time for radiologists and other clinical personnel. Convolutional neural networks may be suited for such tasks, but require substantial amounts of labeled data to train. Obtaining labeled data is a challenge, especially in the medical domain.
Methods: This paper investigates the use of a teacher-student design to utilize datasets with different types of supervision to train an automatic model performing pulmonary tumor segmentation on computed tomography images. The framework consists of two models: the student that performs end-to-end automatic tumor segmentation and the teacher that supplies the student additional pseudo-annotated data during training.
Results: Using only a small proportion of semantically labeled data and a large number of bounding box annotated data, we achieved competitive performance using a teacher-student design. Models trained on larger amounts of semantic annotations did not perform better than those trained on teacher-annotated data. Our model trained on a small number of semantically labeled data achieved a mean dice similarity coefficient of 71.0 on the MSD Lung dataset.
Conclusions: Our results demonstrate the potential of utilizing teacher-student designs to reduce the annotation load, as less supervised annotation schemes may be performed, without any real degradation in segmentation accuracy.publishedVersio
Molecular Imaging
The present book gives an exceptional overview of molecular imaging. Practical approach represents the red thread through the whole book, covering at the same time detailed background information that goes very deep into molecular as well as cellular level. Ideas how molecular imaging will develop in the near future present a special delicacy. This should be of special interest as the contributors are members of leading research groups from all over the world
Deep Learning in Breast Cancer Imaging: A Decade of Progress and Future Directions
Breast cancer has reached the highest incidence rate worldwide among all
malignancies since 2020. Breast imaging plays a significant role in early
diagnosis and intervention to improve the outcome of breast cancer patients. In
the past decade, deep learning has shown remarkable progress in breast cancer
imaging analysis, holding great promise in interpreting the rich information
and complex context of breast imaging modalities. Considering the rapid
improvement in the deep learning technology and the increasing severity of
breast cancer, it is critical to summarize past progress and identify future
challenges to be addressed. In this paper, we provide an extensive survey of
deep learning-based breast cancer imaging research, covering studies on
mammogram, ultrasound, magnetic resonance imaging, and digital pathology images
over the past decade. The major deep learning methods, publicly available
datasets, and applications on imaging-based screening, diagnosis, treatment
response prediction, and prognosis are described in detail. Drawn from the
findings of this survey, we present a comprehensive discussion of the
challenges and potential avenues for future research in deep learning-based
breast cancer imaging.Comment: Survey, 41 page
A Survey on Deep Learning in Medical Image Registration: New Technologies, Uncertainty, Evaluation Metrics, and Beyond
Over the past decade, deep learning technologies have greatly advanced the
field of medical image registration. The initial developments, such as
ResNet-based and U-Net-based networks, laid the groundwork for deep
learning-driven image registration. Subsequent progress has been made in
various aspects of deep learning-based registration, including similarity
measures, deformation regularizations, and uncertainty estimation. These
advancements have not only enriched the field of deformable image registration
but have also facilitated its application in a wide range of tasks, including
atlas construction, multi-atlas segmentation, motion estimation, and 2D-3D
registration. In this paper, we present a comprehensive overview of the most
recent advancements in deep learning-based image registration. We begin with a
concise introduction to the core concepts of deep learning-based image
registration. Then, we delve into innovative network architectures, loss
functions specific to registration, and methods for estimating registration
uncertainty. Additionally, this paper explores appropriate evaluation metrics
for assessing the performance of deep learning models in registration tasks.
Finally, we highlight the practical applications of these novel techniques in
medical imaging and discuss the future prospects of deep learning-based image
registration
- …