3,270 research outputs found
Computational Anatomy for Multi-Organ Analysis in Medical Imaging: A Review
The medical image analysis field has traditionally been focused on the
development of organ-, and disease-specific methods. Recently, the interest in
the development of more 20 comprehensive computational anatomical models has
grown, leading to the creation of multi-organ models. Multi-organ approaches,
unlike traditional organ-specific strategies, incorporate inter-organ relations
into the model, thus leading to a more accurate representation of the complex
human anatomy. Inter-organ relations are not only spatial, but also functional
and physiological. Over the years, the strategies 25 proposed to efficiently
model multi-organ structures have evolved from the simple global modeling, to
more sophisticated approaches such as sequential, hierarchical, or machine
learning-based models. In this paper, we present a review of the state of the
art on multi-organ analysis and associated computation anatomy methodology. The
manuscript follows a methodology-based classification of the different
techniques 30 available for the analysis of multi-organs and multi-anatomical
structures, from techniques using point distribution models to the most recent
deep learning-based approaches. With more than 300 papers included in this
review, we reflect on the trends and challenges of the field of computational
anatomy, the particularities of each anatomical region, and the potential of
multi-organ analysis to increase the impact of 35 medical imaging applications
on the future of healthcare.Comment: Paper under revie
3D mesh processing using GAMer 2 to enable reaction-diffusion simulations in realistic cellular geometries
Recent advances in electron microscopy have enabled the imaging of single
cells in 3D at nanometer length scale resolutions. An uncharted frontier for in
silico biology is the ability to simulate cellular processes using these
observed geometries. Enabling such simulations requires watertight meshing of
electron micrograph images into 3D volume meshes, which can then form the basis
of computer simulations of such processes using numerical techniques such as
the Finite Element Method. In this paper, we describe the use of our recently
rewritten mesh processing software, GAMer 2, to bridge the gap between poorly
conditioned meshes generated from segmented micrographs and boundary marked
tetrahedral meshes which are compatible with simulation. We demonstrate the
application of a workflow using GAMer 2 to a series of electron micrographs of
neuronal dendrite morphology explored at three different length scales and show
that the resulting meshes are suitable for finite element simulations. This
work is an important step towards making physical simulations of biological
processes in realistic geometries routine. Innovations in algorithms to
reconstruct and simulate cellular length scale phenomena based on emerging
structural data will enable realistic physical models and advance discovery at
the interface of geometry and cellular processes. We posit that a new frontier
at the intersection of computational technologies and single cell biology is
now open.Comment: 39 pages, 14 figures. High resolution figures and supplemental movies
available upon reques
Recommended from our members
Registration of the endoluminal surfaces of the colon derived from prone and supine CT colonography
Purpose: Computed tomographic (CT) colonography is a relatively new technique for detecting bowel cancer or potentially precancerous polyps. CT scanning is combined with three-dimensional (3D) image reconstruction to produce a virtual endoluminal representation similar to optical colonoscopy. Because retained fluid and stool can mimic pathology, CT data are acquired with the bowel cleansed and insufflated with gas and patient in both prone and supine positions. Radiologists then match visually endoluminal locations between the two acquisitions in order to determine whether apparent pathology is real or not. This process is hindered by the fact that the colon, essentially a long tube, can undergo considerable deformation between acquisitions. The authors present a novel approach to automatically establish spatial correspondence between prone and supine endoluminal colonic surfaces after surface parameterization, even in the case of local colon collapse.Methods: The complexity of the registration task was reduced from a 3D to a 2D problem by mapping the surfaces extracted from prone and supine CT colonography onto a cylindrical parameterization. A nonrigid cylindrical registration was then performed to align the full colonic surfaces. The curvature information from the original 3D surfaces was used to determine correspondence. The method can also be applied to cases with regions of local colonic collapse by ignoring the collapsed regions during the registration.Results: Using a development set, suitable parameters were found to constrain the cylindrical registration method. Then, the same registration parameters were applied to a different set of 13 validation cases, consisting of 8 fully distended cases and 5 cases exhibiting multiple colonic collapses. All polyps present were well aligned, with a mean (+/- std. dev.) registration error of 5.7 (+/- 3.4) mm. An additional set of 1175 reference points on haustral folds spread over the full endoluminal colon surfaces resulted in an error of 7.7 (+/- 7.4) mm. Here, 82% of folds were aligned correctly after registration with a further 15% misregistered by just onefold.Conclusions: The proposed method reduces the 3D registration task to a cylindrical registration representing the endoluminal surface of the colon. Our algorithm uses surface curvature information as a similarity measure to drive registration to compensate for the large colorectal deformations that occur between prone and supine data acquisitions. The method has the potential to both enhance polyp detection and decrease the radiologist's interpretation time. (C) 2011 American Association of Physicists in Medicine. [DOI: 10.1118/1.3577603
Locally Adaptive Frames in the Roto-Translation Group and their Applications in Medical Imaging
Locally adaptive differential frames (gauge frames) are a well-known
effective tool in image analysis, used in differential invariants and
PDE-flows. However, at complex structures such as crossings or junctions, these
frames are not well-defined. Therefore, we generalize the notion of gauge
frames on images to gauge frames on data representations defined on the extended space of positions and
orientations, which we relate to data on the roto-translation group ,
. This allows to define multiple frames per position, one per
orientation. We compute these frames via exponential curve fits in the extended
data representations in . These curve fits minimize first or second
order variational problems which are solved by spectral decomposition of,
respectively, a structure tensor or Hessian of data on . We include
these gauge frames in differential invariants and crossing preserving PDE-flows
acting on extended data representation and we show their advantage compared
to the standard left-invariant frame on . Applications include
crossing-preserving filtering and improved segmentations of the vascular tree
in retinal images, and new 3D extensions of coherence-enhancing diffusion via
invertible orientation scores
Geometrically Induced Force Interaction for Three-Dimensional Deformable Models
This work introduces a novel 3D deformable model that is based on a geometrically induced external force field, which can be conveniently generalised to arbitrary dimensions. This external force field is based on hypothesised interactions between the relative geometries of the deformable model and the object boundary. The relative geometrical configurations contribute to a dynamic vector force field that changes accordingly as the deformable model evolves. In addition, we show that by enhancing the geometrical interaction field with a nonlocal edge preserving algorithm, the new model can effectively overcome image noise. We provide a comprehensive comparative study and show that the proposed method achieves significant improvements against existing techniques
The whole mesh deformation model: A fast image segmentation method suitable for effective parallelization
In this article, we propose a novel image segmentation method called the whole mesh deformation (WMD) model, which aims at addressing the problems of modern medical imaging. Such problems have raised from the combination of several factors: (1) significant growth of medical image volumes sizes due to increasing capabilities of medical acquisition devices; (2) the will to increase the complexity of image processing algorithms in order to explore new functionality; (3) change in processor development and turn towards multi processing units instead of growing bus speeds and the number of operations per second of a single processing unit. Our solution is based on the concept of deformable models and is characterized by a very effective and precise segmentation capability. The proposed WMD model uses a volumetric mesh instead of a contour or a surface to represent the segmented shapes of interest, which allows exploiting more information in the image and obtaining results in shorter times, independently of image contents. The model also offers a good ability for topology changes and allows effective parallelization of workflow, which makes it a very good choice for large datasets. We present a precise model description, followed by experiments on artificial images and real medical data
- …