15 research outputs found
LSSANet: A Long Short Slice-Aware Network for Pulmonary Nodule Detection
Convolutional neural networks (CNNs) have been demonstrated to be highly
effective in the field of pulmonary nodule detection. However, existing CNN
based pulmonary nodule detection methods lack the ability to capture long-range
dependencies, which is vital for global information extraction. In computer
vision tasks, non-local operations have been widely utilized, but the
computational cost could be very high for 3D computed tomography (CT) images.
To address this issue, we propose a long short slice-aware network (LSSANet)
for the detection of pulmonary nodules. In particular, we develop a new
non-local mechanism termed long short slice grouping (LSSG), which splits the
compact non-local embeddings into a short-distance slice grouped one and a
long-distance slice grouped counterpart. This not only reduces the
computational burden, but also keeps long-range dependencies among any elements
across slices and in the whole feature map. The proposed LSSG is easy-to-use
and can be plugged into many pulmonary nodule detection networks. To verify the
performance of LSSANet, we compare with several recently proposed and
competitive detection approaches based on 2D/3D CNN. Promising evaluation
results on the large-scale PN9 dataset demonstrate the effectiveness of our
method. Code is at https://github.com/Ruixxxx/LSSANet.Comment: MICCAI 202
OpenShape: Scaling Up 3D Shape Representation Towards Open-World Understanding
We introduce OpenShape, a method for learning multi-modal joint
representations of text, image, and point clouds. We adopt the commonly used
multi-modal contrastive learning framework for representation alignment, but
with a specific focus on scaling up 3D representations to enable open-world 3D
shape understanding. To achieve this, we scale up training data by ensembling
multiple 3D datasets and propose several strategies to automatically filter and
enrich noisy text descriptions. We also explore and compare strategies for
scaling 3D backbone networks and introduce a novel hard negative mining module
for more efficient training. We evaluate OpenShape on zero-shot 3D
classification benchmarks and demonstrate its superior capabilities for
open-world recognition. Specifically, OpenShape achieves a zero-shot accuracy
of 46.8% on the 1,156-category Objaverse-LVIS benchmark, compared to less than
10% for existing methods. OpenShape also achieves an accuracy of 85.3% on
ModelNet40, outperforming previous zero-shot baseline methods by 20% and
performing on par with some fully-supervised methods. Furthermore, we show that
our learned embeddings encode a wide range of visual and semantic concepts
(e.g., subcategories, color, shape, style) and facilitate fine-grained text-3D
and image-3D interactions. Due to their alignment with CLIP embeddings, our
learned shape representations can also be integrated with off-the-shelf
CLIP-based models for various applications, such as point cloud captioning and
point cloud-conditioned image generation.Comment: Project Website: https://colin97.github.io/OpenShape
RibSeg v2: A Large-scale Benchmark for Rib Labeling and Anatomical Centerline Extraction
Automatic rib labeling and anatomical centerline extraction are common
prerequisites for various clinical applications. Prior studies either use
in-house datasets that are inaccessible to communities, or focus on rib
segmentation that neglects the clinical significance of rib labeling. To
address these issues, we extend our prior dataset (RibSeg) on the binary rib
segmentation task to a comprehensive benchmark, named RibSeg v2, with 660 CT
scans (15,466 individual ribs in total) and annotations manually inspected by
experts for rib labeling and anatomical centerline extraction. Based on the
RibSeg v2, we develop a pipeline including deep learning-based methods for rib
labeling, and a skeletonization-based method for centerline extraction. To
improve computational efficiency, we propose a sparse point cloud
representation of CT scans and compare it with standard dense voxel grids.
Moreover, we design and analyze evaluation metrics to address the key
challenges of each task. Our dataset, code, and model are available online to
facilitate open research at https://github.com/M3DV/RibSegComment: 10 pages, 6 figures, journa
LSSANet: A Long Short Slice-Aware Network for Pulmonary Nodule Detection
Convolutional neural networks (CNNs) have been demonstrated to be highly effective in the field of pulmonary nodule detection. However, existing CNN based pulmonary nodule detection methods lack the ability to capture long-range dependencies, which is vital for global information extraction. In computer vision tasks, non-local operations have been widely utilized, but the computational cost could be very high for 3D computed tomography (CT) images. To address this issue, we propose a long short slice-aware network (LSSANet) for the detection of pulmonary nodules. In particular, we develop a new non-local mechanism termed long short slice grouping (LSSG), which splits the compact non-local embeddings into a short-distance slice grouped one and a long-distance slice grouped counterpart. This not only reduces the computational burden, but also keeps long-range dependencies among any elements across slices and in the whole feature map. The proposed LSSG is easy-to-use and can be plugged into many pulmonary nodule detection networks. To verify the performance of LSSANet, we compare with several recently proposed and competitive detection approaches based on 2D/3D CNN. Promising evaluation results on the large-scale PN9 dataset demonstrate the effectiveness of our method. Code is at https:// github.com/Ruixxxx/LSSANet.CVLA
Current Visceral Leishmaniasis Research: A Research Review to Inspire Future Study
Visceral leishmaniasis (VL), one of the deadliest parasitic diseases in the world, causes more than 50,000 human deaths each year and afflicts millions of people throughout South America, East Africa, South Asia, and Mediterranean Region. In 2015 the World Health Organization classified VL as a neglected tropical disease (NTD), prompting concentrated study of the VL epidemic using mathematical and simulation models. This paper reviews literature related to prevalence and prevention control strategies. More than thirty current research works were reviewed and classified based on VL epidemic study methods, including modeling approaches, control strategies, and simulation techniques since 2013. A summarization of these technical methods, major findings, and contributions from existing works revealed that VL epidemic research efforts must improve in the areas of validating and verifying VL mathematical models with real-world epidemic data. In addition, more dynamic disease control strategies must be explored and advanced simulation techniques must be used to predict VL pandemics
Deep learning predicts malignancy and metastasis of solid pulmonary nodules from CT scans
In the clinic, it is difficult to distinguish the malignancy and aggressiveness of solid pulmonary nodules (PNs). Incorrect assessments may lead to delayed diagnosis and an increased risk of complications. We developed and validated a deep learning-based model for the prediction of malignancy as well as local or distant metastasis in solid PNs based on CT images of primary lesions during initial diagnosis. In this study, we reviewed the data from multiple patients with solid PNs at our institution from 1 January 2019 to 30 April 2022. The patients were divided into three groups: benign, Ia-stage lung cancer, and T1-stage lung cancer with metastasis. Each cohort was further split into training and testing groups. The deep learning system predicted the malignancy and metastasis status of solid PNs based on CT images, and then we compared the malignancy prediction results among four different levels of clinicians. Experiments confirmed that human–computer collaboration can further enhance diagnostic accuracy. We made a held-out testing set of 134 cases, with 689 cases in total. Our convolutional neural network model reached an area under the ROC (AUC) of 80.37% for malignancy prediction and an AUC of 86.44% for metastasis prediction. In observer studies involving four clinicians, the proposed deep learning method outperformed a junior respiratory clinician and a 5-year respiratory clinician by considerable margins; it was on par with a senior respiratory clinician and was only slightly inferior to a senior radiologist. Our human–computer collaboration experiment showed that by simply adding binary human diagnosis into model prediction probabilities, model AUC scores improved to 81.80–88.70% when combined with three out of four clinicians. In summary, the deep learning method can accurately diagnose the malignancy of solid PNs, improve its performance when collaborating with human experts, predict local or distant metastasis in patients with T1-stage lung cancer, and facilitate the application of precision medicine
What Makes for Automatic Reconstruction of Pulmonary Segments
3D reconstruction of pulmonary segments plays an important role in surgical treatment planning of lung cancer, which facilitates preservation of pulmonary function and helps ensure low recurrence rates. However, automatic reconstruction of pulmonary segments remains unexplored in the era of deep learning. In this paper, we investigate what makes for automatic reconstruction of pulmonary segments. First and foremost, we formulate, clinically and geometrically, the anatomical definitions of pulmonary segments, and propose evaluation metrics adhering to these definitions. Second, we propose ImPulSe (Implicit Pulmonary Segment), a deep implicit surface model designed for pulmonary segment reconstruction. The automatic reconstruction of pulmonary segments by ImPulSe is accurate in metrics and visually appealing. Compared with canonical segmentation methods, ImPulSe outputs continuous predictions of arbitrary resolutions with higher training efficiency and fewer parameters. Lastly, we experiment with different network inputs to analyze what matters in the task of pulmonary segment reconstruction. Our code is available at https://github.com/M3DV/ImPulSe.CVLA
Improvement of Photocatalytic Performance for the g-C3N4/MoS2 Composite Used for Hypophosphite Oxidation
The synthesized g-C3N4/MoS2 composite was a high-efficiency photocatalytic for hypophosphite oxidation. In this work, a stable and cheap g-C3N4 worked as the chelating agent and combined with the MoS2 materials. The structures of the fabricated g-C3N4/MoS2 photocatalyst were characterized by some methods including X-ray diffraction (XRD), scanning electron microscopy (SEM), and X-ray photoelectron spectra (XPS). Moreover, the photocatalytic performances of various photocatalysts were measured by analyzing the oxidation efficiency of hypophosphite under visible light irradiation and the oxidation efficiency of hypophosphite using the g-C3N4/MoS2 photocatalyst which was 93.45%. According to the results, the g-C3N4/MoS2 composite showed a promising photocatalytic performance for hypophosphite oxidation. The improved photocatalytic performance for hypophosphite oxidation was due to the effective charge separation analyzed by the photoluminescence (PL) emission spectra. The transient photocurrent response measurement indicated that the g-C3N4/MoS2 composites (2.5 μA cm–2) were 10 times improved photocurrent intensity and 2 times improved photocurrent intensity comparing with the pure g-C3N4 (0.25 μA cm–2) and MoS2 (1.25 μA cm–2), respectively. The photocatalytic mechanism of hypophosphite oxidation was analyzed by adding some scavengers, and the recycle experiments indicated that the g-C3N4/MoS2 composite had a good stability
ViSTA: A Novel Network Improving Lung Adenocarcinoma Invasiveness Prediction from Follow-Up CT Series
To investigate the value of the deep learning method in predicting the invasiveness of early lung adenocarcinoma based on irregularly sampled follow-up computed tomography (CT) scans. In total, 351 nodules were enrolled in the study. A new deep learning network based on temporal attention, named Visual Simple Temporal Attention (ViSTA), was proposed to process irregularly sampled follow-up CT scans. We conducted substantial experiments to investigate the supplemental value in predicting the invasiveness using serial CTs. A test set composed of 69 lung nodules was reviewed by three radiologists. The performance of the model and radiologists were compared and analyzed. We also performed a visual investigation to explore the inherent growth pattern of the early adenocarcinomas. Among counterpart models, ViSTA showed the best performance (AUC: 86.4% vs. 60.6%, 75.9%, 66.9%, 73.9%, 76.5%, 78.3%). ViSTA also outperformed the model based on Volume Doubling Time (AUC: 60.6%). ViSTA scored higher than two junior radiologists (accuracy of 81.2% vs. 75.4% and 71.0%) and came close to the senior radiologist (85.5%). Our proposed model using irregularly sampled follow-up CT scans achieved promising accuracy in evaluating the invasiveness of the early stage lung adenocarcinoma. Its performance is comparable with senior experts and better than junior experts and traditional deep learning models. With further validation, it can potentially be applied in clinical practice