2,953 research outputs found
Automatic Tooth Segmentation from 3D Dental Model using Deep Learning: A Quantitative Analysis of what can be learnt from a Single 3D Dental Model
3D tooth segmentation is an important task for digital orthodontics. Several
Deep Learning methods have been proposed for automatic tooth segmentation from
3D dental models or intraoral scans. These methods require annotated 3D
intraoral scans. Manually annotating 3D intraoral scans is a laborious task.
One approach is to devise self-supervision methods to reduce the manual
labeling effort. Compared to other types of point cloud data like scene point
cloud or shape point cloud data, 3D tooth point cloud data has a very regular
structure and a strong shape prior. We look at how much representative
information can be learnt from a single 3D intraoral scan. We evaluate this
quantitatively with the help of ten different methods of which six are generic
point cloud segmentation methods whereas the other four are tooth segmentation
specific methods. Surprisingly, we find that with a single 3D intraoral scan
training, the Dice score can be as high as 0.86 whereas the full training set
gives Dice score of 0.94. We conclude that the segmentation methods can learn a
great deal of information from a single 3D tooth point cloud scan under
suitable conditions e.g. data augmentation. We are the first to quantitatively
evaluate and demonstrate the representation learning capability of Deep
Learning methods from a single 3D intraoral scan. This can enable building
self-supervision methods for tooth segmentation under extreme data limitation
scenario by leveraging the available data to the fullest possible extent.Comment: accepted to SIPAIM 202
Recommended from our members
Deep learning for cardiac image segmentation: A review
Deep learning has become the most widely used approach for cardiac image segmentation in recent years. In this paper, we provide a review of over 100 cardiac image segmentation papers using deep learning, which covers common imaging modalities including magnetic resonance imaging (MRI), computed tomography (CT), and ultrasound (US) and major anatomical structures of interest (ventricles, atria and vessels). In addition, a summary of publicly available cardiac image datasets and code repositories are included to provide a base for encouraging reproducible research. Finally, we discuss the challenges and limitations with current deep learning-based approaches (scarcity of labels, model generalizability across different domains, interpretability) and suggest potential directions for future research
A Generative Shape Compositional Framework to Synthesise Populations of Virtual Chimaeras
Generating virtual populations of anatomy that capture sufficient variability while remaining plausible is essential for conducting in-silico trials of medical devices. However, not all anatomical shapes of interest are always available for each individual in a population. Hence, missing/partially-overlapping anatomical information is often available across individuals in a population. We introduce a generative shape model for complex anatomical structures, learnable from datasets of unpaired datasets. The proposed generative model can synthesise complete whole complex shape assemblies coined virtual chimaeras, as opposed to natural human chimaeras. We applied this framework to build virtual chimaeras from databases of whole-heart shape assemblies that each contribute samples for heart substructures. Specifically, we propose a generative shape compositional framework which comprises two components - a part-aware generative shape model which captures the variability in shape observed for each structure of interest in the training population; and a spatial composition network which assembles/composes the structures synthesised by the former into multi-part shape assemblies (viz. virtual chimaeras). We also propose a novel self supervised learning scheme that enables the spatial composition network to be trained with partially overlapping data and weak labels. We trained and validated our approach using shapes of cardiac structures derived from cardiac magnetic resonance images available in the UK Biobank. Our approach significantly outperforms a PCA-based shape model (trained with complete data) in terms of generalisability and specificity. This demonstrates the superiority of the proposed approach as the synthesised cardiac virtual populations are more plausible and capture a greater degree of variability in shape than those generated by the PCA-based shape model
Graph Laplacian for Image Anomaly Detection
Reed-Xiaoli detector (RXD) is recognized as the benchmark algorithm for image
anomaly detection; however, it presents known limitations, namely the
dependence over the image following a multivariate Gaussian model, the
estimation and inversion of a high-dimensional covariance matrix, and the
inability to effectively include spatial awareness in its evaluation. In this
work, a novel graph-based solution to the image anomaly detection problem is
proposed; leveraging the graph Fourier transform, we are able to overcome some
of RXD's limitations while reducing computational cost at the same time. Tests
over both hyperspectral and medical images, using both synthetic and real
anomalies, prove the proposed technique is able to obtain significant gains
over performance by other algorithms in the state of the art.Comment: Published in Machine Vision and Applications (Springer
DeepCut: Object Segmentation from Bounding Box Annotations using Convolutional Neural Networks
In this paper, we propose DeepCut, a method to obtain pixelwise object
segmentations given an image dataset labelled with bounding box annotations. It
extends the approach of the well-known GrabCut method to include machine
learning by training a neural network classifier from bounding box annotations.
We formulate the problem as an energy minimisation problem over a
densely-connected conditional random field and iteratively update the training
targets to obtain pixelwise object segmentations. Additionally, we propose
variants of the DeepCut method and compare those to a naive approach to CNN
training under weak supervision. We test its applicability to solve brain and
lung segmentation problems on a challenging fetal magnetic resonance dataset
and obtain encouraging results in terms of accuracy
Multi-site, Multi-domain Airway Tree Modeling (ATM'22): A Public Benchmark for Pulmonary Airway Segmentation
Open international challenges are becoming the de facto standard for
assessing computer vision and image analysis algorithms. In recent years, new
methods have extended the reach of pulmonary airway segmentation that is closer
to the limit of image resolution. Since EXACT'09 pulmonary airway segmentation,
limited effort has been directed to quantitative comparison of newly emerged
algorithms driven by the maturity of deep learning based approaches and
clinical drive for resolving finer details of distal airways for early
intervention of pulmonary diseases. Thus far, public annotated datasets are
extremely limited, hindering the development of data-driven methods and
detailed performance evaluation of new algorithms. To provide a benchmark for
the medical imaging community, we organized the Multi-site, Multi-domain Airway
Tree Modeling (ATM'22), which was held as an official challenge event during
the MICCAI 2022 conference. ATM'22 provides large-scale CT scans with detailed
pulmonary airway annotation, including 500 CT scans (300 for training, 50 for
validation, and 150 for testing). The dataset was collected from different
sites and it further included a portion of noisy COVID-19 CTs with ground-glass
opacity and consolidation. Twenty-three teams participated in the entire phase
of the challenge and the algorithms for the top ten teams are reviewed in this
paper. Quantitative and qualitative results revealed that deep learning models
embedded with the topological continuity enhancement achieved superior
performance in general. ATM'22 challenge holds as an open-call design, the
training data and the gold standard evaluation are available upon successful
registration via its homepage.Comment: 32 pages, 16 figures. Homepage: https://atm22.grand-challenge.org/.
Submitte
Tubular Structure Segmentation Using Spatial Fully Connected Network with Radial Distance Loss for 3D Medical Images
This paper presents a new spatial fully connected tubular network for 3D tubular-structure segmentation. Automatic and complete segmentation of intricate tubular structures remains an unsolved challenge in the medical image analysis. Airways and vasculature pose high demands on medical image analysis as they are elongated fine structures with calibers ranging from several tens of voxels to voxel-level resolution, branching in deeply multi-scale fashion, and with complex topological and spatial relationships. Most machine/deep learning approaches are based on intensity features and ignore spatial consistency across the network that are otherwise distinct in tubular structures. In this work, we introduce 3D slice-by-slice convolutional layers in a U-Net architecture to capture the spatial information of elongated structures. Furthermore, we present a novel loss function, coined radial distance loss, specifically designed for tubular structures. The commonly used methods of cross-entropy loss and generalized Dice loss are sensitive to volumetric variation. However, in tiny tubular structure segmentation, topological errors are as important as volumetric errors. The proposed radial distance loss places higher weight to the centerline, and this weight decreases along the radial direction. Radial distance loss can help networks focus more attention on tiny structures than on thicker tubular structures. We perform experiments on bronchus segmentation on 3D CT images. The experimental results show that compared to the baseline U-Net, our proposed network achieved improvement about 24% and 30% in Dice index and centerline over ratio
- …