67,385 research outputs found
Large-Scale Automatic Reconstruction of Neuronal Processes from Electron Microscopy Images
Automated sample preparation and electron microscopy enables acquisition of
very large image data sets. These technical advances are of special importance
to the field of neuroanatomy, as 3D reconstructions of neuronal processes at
the nm scale can provide new insight into the fine grained structure of the
brain. Segmentation of large-scale electron microscopy data is the main
bottleneck in the analysis of these data sets. In this paper we present a
pipeline that provides state-of-the art reconstruction performance while
scaling to data sets in the GB-TB range. First, we train a random forest
classifier on interactive sparse user annotations. The classifier output is
combined with an anisotropic smoothing prior in a Conditional Random Field
framework to generate multiple segmentation hypotheses per image. These
segmentations are then combined into geometrically consistent 3D objects by
segmentation fusion. We provide qualitative and quantitative evaluation of the
automatic segmentation and demonstrate large-scale 3D reconstructions of
neuronal processes from a volume of brain
tissue over a cube of in each dimension corresponding to
1000 consecutive image sections. We also introduce Mojo, a proofreading tool
including semi-automated correction of merge errors based on sparse user
scribbles
COVLIAS 1.0: Lung segmentation in COVID-19 computed tomography scans using hybrid deep learning artificial intelligence models
Background: COVID-19 lung segmentation using Computed Tomography (CT) scans is important for the diagnosis of lung severity. The process of automated lung segmentation is challenging due to (a) CT radiation dosage and (b) ground-glass opacities caused by COVID-19. The lung segmentation methodologies proposed in 2020 were semi-or automated but not reliable, accurate, and user-friendly. The proposed study presents a COVID Lung Image Analysis System (COVLIAS 1.0, AtheroPointâą, Roseville, CA, USA) consisting of hybrid deep learning (HDL) models for lung segmentation. Methodology: The COVLIAS 1.0 consists of three methods based on solo deep learning (SDL) or hybrid deep learning (HDL). SegNet is proposed in the SDL category while VGG-SegNet and ResNet-SegNet are designed under the HDL paradigm. The three proposed AI approaches were benchmarked against the National Institute of Health (NIH)-based conventional segmentation model using fuzzy-connectedness. A cross-validation protocol with a 40:60 ratio between training and testing was designed, with 10% validation data. The ground truth (GT) was manually traced by a radiologist trained personnel. For performance evaluation, nine different criteria were selected to perform the evaluation of SDL or HDL lung segmentation regions and lungs long axis against GT. Results: Using the database of 5000 chest CT images (from 72 patients), COVLIAS 1.0 yielded AUC of ~0.96, ~0.97, ~0.98, and ~0.96 (p-value < 0.001), respectively within 5% range of GT area, for SegNet, VGG-SegNet, ResNet-SegNet, and NIH. The mean Figure of Merit using four models (left and right lung) was above 94%. On benchmarking against the National Institute of Health (NIH) segmentation method, the proposed model demonstrated a 58% and 44% improvement in ResNet-SegNet, 52% and 36% improvement in VGG-SegNet for lung area, and lung long axis, respectively. The PE statistics performance was in the following order: ResNet-SegNet > VGG-SegNet > NIH > SegNet. The HDL runs in <1 s on test data per image. Conclusions: The COVLIAS 1.0 system can be applied in real-time for radiology-based clinical settings
Interactive medical image segmentation - towards integrating human guidance and deep learning
Medical image segmentation is an essential step in many clinical workflows involving diagnostics and patient treatment planning. Deep learning has advanced the field of medical image segmentation, particularly with respect to automating contouring. However, some anatomical structures, such as tumours, are challenging for fully automated methods. When automatic methods fail, manual contouring is required. In such cases, semi-automatic tools can support clinicians in contouring tasks. The objective of this thesis was to leverage cliniciansâ expert knowledge when performing segmentation tasks, allowing for interactions along the segmentation workflow and improving deep learning predictions.
In this thesis, a deep learning approach is proposed that produces a 3D segmentation of a structure of interest based on a user-provided input. If trained on a diverse set of structures, state-of-the-art performance was achieved for structures included in the training set. More importantly, the model was also able to generalize and make predictions for unseen structures that were not represented in the training set. Various avenues to guide user interaction and leverage multiple user inputs more effectively were also investigated. These further improved the segmentation performance and demonstrated the ability to accurately segment a broad range of anatomical structures.
An evaluation by clinicians demonstrated that time spent contouring was reduced when using the contextual deep learning tool as compared to conventional contouring tools. This evaluation also revealed that the majority of contouring time is observation time, which is only indirectly affected by the segmentation approach. This suggests, that user interface design and guiding the userâs attention to critical areas can have a large impact on time taken on the contouring task.
Overall, this thesis proposes an interactive deep learning segmentation method, demonstrates its clinical impact, and highlights the potential synergies between clinicians and artificial intelligence
Automated quantification and evaluation of motion artifact on coronary CT angiography images
Abstract Purpose
This study developed and validated a Motion Artifact Quantification algorithm to automatically quantify the severity of motion artifacts on coronary computed tomography angiography (CCTA) images. The algorithm was then used to develop a Motion IQ Decision method to automatically identify whether a CCTA dataset is of sufficient diagnostic image quality or requires further correction. Method
The developed Motion Artifact Quantification algorithm includes steps to identify the right coronary artery (RCA) regions of interest (ROIs), segment vessel and shading artifacts, and to calculate the motion artifact score (MAS) metric. The segmentation algorithms were verified against groundâtruth manual segmentations. The segmentation algorithms were also verified by comparing and analyzing the MAS calculated from groundâtruth segmentations and the algorithmâgenerated segmentations. The Motion IQ Decision algorithm first identifies slices with unsatisfactory image quality using a MAS threshold. The algorithm then uses an artifactâlength threshold to determine whether the degraded vessel segment is large enough to cause the dataset to be nondiagnostic. An observer study on 30 clinical CCTA datasets was performed to obtain the groundâtruth decisions of whether the datasets were of sufficient image quality. A fiveâfold crossâvalidation was used to identify the thresholds and to evaluate the Motion IQ Decision algorithm. Results
The automated segmentation algorithms in the Motion Artifact Quantification algorithm resulted in Dice coefficients of 0.84 for the segmented vessel regions and 0.75 for the segmented shading artifact regions. The MAS calculated using the automated algorithm was within 10% of the values obtained using groundâtruth segmentations. The MAS threshold and artifactâlength thresholds were determined by the ROC analysis to be 0.6 and 6.25 mm by all folds. The Motion IQ Decision algorithm demonstrated 100% sensitivity, 66.7% ± 27.9% specificity, and a total accuracy of 86.7% ± 12.5% for identifying datasets in which the RCA required correction. The Motion IQ Decision algorithm demonstrated 91.3% sensitivity, 71.4% specificity, and a total accuracy of 86.7% for identifying CCTA datasets that need correction for any of the three main vessels. Conclusion
The Motion Artifact Quantification algorithm calculated accurate
Shallow vs deep learning architectures for white matter lesion segmentation in the early stages of multiple sclerosis
In this work, we present a comparison of a shallow and a deep learning
architecture for the automated segmentation of white matter lesions in MR
images of multiple sclerosis patients. In particular, we train and test both
methods on early stage disease patients, to verify their performance in
challenging conditions, more similar to a clinical setting than what is
typically provided in multiple sclerosis segmentation challenges. Furthermore,
we evaluate a prototype naive combination of the two methods, which refines the
final segmentation. All methods were trained on 32 patients, and the evaluation
was performed on a pure test set of 73 cases. Results show low lesion-wise
false positives (30%) for the deep learning architecture, whereas the shallow
architecture yields the best Dice coefficient (63%) and volume difference
(19%). Combining both shallow and deep architectures further improves the
lesion-wise metrics (69% and 26% lesion-wise true and false positive rate,
respectively).Comment: Accepted to the MICCAI 2018 Brain Lesion (BrainLes) worksho
- âŠ