1,342 research outputs found
Discrete Imaging Models for Three-Dimensional Optoacoustic Tomography using Radially Symmetric Expansion Functions
Optoacoustic tomography (OAT), also known as photoacoustic tomography, is an
emerging computed biomedical imaging modality that exploits optical contrast
and ultrasonic detection principles. Iterative image reconstruction algorithms
that are based on discrete imaging models are actively being developed for OAT
due to their ability to improve image quality by incorporating accurate models
of the imaging physics, instrument response, and measurement noise. In this
work, we investigate the use of discrete imaging models based on Kaiser-Bessel
window functions for iterative image reconstruction in OAT. A closed-form
expression for the pressure produced by a Kaiser-Bessel function is calculated,
which facilitates accurate computation of the system matrix.
Computer-simulation and experimental studies are employed to demonstrate the
potential advantages of Kaiser-Bessel function-based iterative image
reconstruction in OAT
Tomographic image quality of rotating slat versus parallel hole-collimated SPECT
Parallel and converging hole collimators are most frequently used in nuclear medicine. Less common is the use of rotating slat collimators for single photon emission computed tomography (SPECT). The higher photon collection efficiency, inherent to the geometry of rotating slat collimators, results in much lower noise in the data. However, plane integrals contain spatial information in only one direction, whereas line integrals provide two-dimensional information. It is not a trivial question whether the initial gain in efficiency will compensate for the lower information content in the plane integrals. Therefore, a comparison of the performance of parallel hole and rotating slat collimation is needed. This study compares SPECT with rotating slat and parallel hole collimation in combination with MLEM reconstruction with accurate system modeling and correction for scatter and attenuation. A contrast-to-noise study revealed an improvement of a factor 2-3 for hot lesions and more than a factor of 4 for cold lesion. Furthermore, a clinically relevant case of heart lesion detection is simulated for rotating slat and parallel hole collimators. In this case, rotating slat collimators outperform the traditional parallel hole collimators. We conclude that rotating slat collimators are a valuable alternative for parallel hole collimators
Recommended from our members
Quantitative Statistical Methods for Image Quality Assessment
Quantitative measures of image quality and reliability are critical for both qualitative interpretation and quantitative analysis of medical images. While, in theory, it is possible to analyze reconstructed images by means of Monte Carlo simulations using a large number of noise realizations, the associated computational burden makes this approach impractical. Additionally, this approach is less meaningful in clinical scenarios, where multiple noise realizations are generally unavailable. The practical alternative is to compute closed-form analytical expressions for image quality measures. The objective of this paper is to review statistical analysis techniques that enable us to compute two key metrics: resolution (determined from the local impulse response) and covariance. The underlying methods include fixed-point approaches, which compute these metrics at a fixed point (the unique and stable solution) independent of the iterative algorithm employed, and iteration-based approaches, which yield results that are dependent on the algorithm, initialization, and number of iterations. We also explore extensions of some of these methods to a range of special contexts, including dynamic and motion-compensated image reconstruction. While most of the discussed techniques were developed for emission tomography, the general methods are extensible to other imaging modalities as well. In addition to enabling image characterization, these analysis techniques allow us to control and enhance imaging system performance. We review practical applications where performance improvement is achieved by applying these ideas to the contexts of both hardware (optimizing scanner design) and image reconstruction (designing regularization functions that produce uniform resolution or maximize task-specific figures of merit)
Inversion using a new low-dimensional representation of complex binary geological media based on a deep neural network
Efficient and high-fidelity prior sampling and inversion for complex
geological media is still a largely unsolved challenge. Here, we use a deep
neural network of the variational autoencoder type to construct a parametric
low-dimensional base model parameterization of complex binary geological media.
For inversion purposes, it has the attractive feature that random draws from an
uncorrelated standard normal distribution yield model realizations with spatial
characteristics that are in agreement with the training set. In comparison with
the most commonly used parametric representations in probabilistic inversion,
we find that our dimensionality reduction (DR) approach outperforms principle
component analysis (PCA), optimization-PCA (OPCA) and discrete cosine transform
(DCT) DR techniques for unconditional geostatistical simulation of a
channelized prior model. For the considered examples, important compression
ratios (200 - 500) are achieved. Given that the construction of our
parameterization requires a training set of several tens of thousands of prior
model realizations, our DR approach is more suited for probabilistic (or
deterministic) inversion than for unconditional (or point-conditioned)
geostatistical simulation. Probabilistic inversions of 2D steady-state and 3D
transient hydraulic tomography data are used to demonstrate the DR-based
inversion. For the 2D case study, the performance is superior compared to
current state-of-the-art multiple-point statistics inversion by sequential
geostatistical resampling (SGR). Inversion results for the 3D application are
also encouraging
Stopping Rules for Algebraic Iterative Reconstruction Methods in Computed Tomography
Algebraic models for the reconstruction problem in X-ray computed tomography
(CT) provide a flexible framework that applies to many measurement geometries.
For large-scale problems we need to use iterative solvers, and we need stopping
rules for these methods that terminate the iterations when we have computed a
satisfactory reconstruction that balances the reconstruction error and the
influence of noise from the measurements. Many such stopping rules are
developed in the inverse problems communities, but they have not attained much
attention in the CT world. The goal of this paper is to describe and illustrate
four stopping rules that are relevant for CT reconstructions.Comment: 11 pages, 10 figure
A New Statistical Reconstruction Method for the Computed Tomography Using an X-Ray Tube with Flying Focal Spot
Abstract
This paper presents a new image reconstruction method for spiral cone- beam tomography scanners in which an X-ray tube with a flying focal spot is used. The method is based on principles related to the statistical model-based iterative reconstruction (MBIR) methodology. The proposed approach is a continuous-to-continuous data model approach, and the forward model is formulated as a shift-invariant system. This allows for avoiding a nutating reconstruction-based approach, e.g. the advanced single slice rebinning methodology (ASSR) that is usually applied in computed tomography (CT) scanners with X-ray tubes with a flying focal spot. In turn, the proposed approach allows for significantly accelerating the reconstruction processing and, generally, for greatly simplifying the entire reconstruction procedure. Additionally, it improves the quality of the reconstructed images in comparison to the traditional algorithms, as confirmed by extensive simulations. It is worth noting that the main purpose of introducing statistical reconstruction methods to medical CT scanners is the reduction of the impact of measurement noise on the quality of tomography images and, consequently, the dose reduction of X-ray radiation absorbed by a patient. A series of computer simulations followed by doctor's assessments have been performed, which indicate how great a reduction of the absorbed dose can be achieved using the reconstruction approach presented here
Development and implementation of efficient noise suppression methods for emission computed tomography
In PET and SPECT imaging, iterative reconstruction is now widely used due to its capability of incorporating into the reconstruction process a physics model and Bayesian statistics involved in photon detection. Iterative reconstruction methods rely on regularization terms to suppress image noise and render radiotracer distribution with good image quality. The choice of regularization method substantially affects the appearances of reconstructed images, and is thus a critical aspect of the reconstruction process. Major contributions of this work include implementation and evaluation of various new regularization methods. Previously, our group developed a preconditioned alternating projection algorithm (PAPA) to optimize the emission computed tomography (ECT) objective function with the non-differentiable total variation (TV) regularizer. The algorithm was modified to optimize the proposed reconstruction objective functions.
First, two novel TV-based regularizersâhigh-order total variation (HOTV) and infimal convolution total variation (ICTV)âwere proposed as alternative choices to the customary TV regularizer in SPECT reconstruction, to reduce âstaircaseâ artifacts produced by TV. We have evaluated both proposed reconstruction methods (HOTV-PAPA and ICTV-PAPA), and compared them with the TV regularized reconstruction (TV-PAPA) and the clinical standard, Gaussian post-filtered, expectation-maximization reconstruction method (GPF-EM) using both Monte Carlo-simulated data and anonymized clinical data. Model-observer studies using Monte Carlo-simulated data indicate that ICTV-PAPA is able to reconstruct images with similar or better lesion detectability, compared with clinical standard GPF-EM methods, but at lower detected count levels. This implies that switching from GPF-EM to ICTV-PAPA can reduce patient dose while maintaining image quality for diagnostic use.
Second, the 1 norm of discrete cosine transform (DCT)-induced framelet regularization was studied. We decomposed the image into high and low spatial-frequency components, and then preferentially penalized the high spatial-frequency components. The DCT-induced framelet transform of the natural radiotracer distribution image is sparse. By using this property, we were able to effectively suppress image noise without overly compromising spatial resolution or image contrast.
Finally, the fractional norm of the first-order spatial gradient was introduced as a regularizer. We implemented 2/3 and 1/2 norms to suppress image spatial variability. Due to the strong penalty of small differences between neighboring pixels, fractional-norm regularizers suffer from similar cartoon-like artifacts as with the TV regularizer. However, when penalty weights are properly selected, fractional-norm regularizers outperform TV in terms of noise suppression and contrast recovery
Incorporating accurate statistical modeling in PET: reconstruction for whole-body imaging
Tese de doutoramento em BiofĂsica, apresentada Ă Universidade de Lisboa atravĂ©s da Faculdade de CiĂȘncias, 2007The thesis is devoted to image reconstruction in 3D whole-body PET imaging. OSEM ( Ordered Subsets Expectation maximization ) is a statistical algorithm that assumes Poisson data. However, corrections for physical effects (attenuation, scattered and random coincidences) and detector efficiency remove the Poisson characteristics of these data. The Fourier Rebinning (FORE), that combines 3D imaging with fast 2D reconstructions, requires corrected data. Thus, if it will be used or whenever data are corrected prior to OSEM, the need to restore the Poisson-like characteristics is present. Restoring Poisson-like data, i.e., making the variance equal to the mean, was achieved through the use of weighted OSEM algorithms. One of them is the NECOSEM, relying on the NEC weighting transformation. The distinctive feature of this algorithm is the NEC multiplicative factor, defined as the ratio between the mean and the variance. With real clinical data this is critical, since there is only one value collected for each bin the data value itself. For simulated data, if we keep track of the values for these two statistical moments, the exact values for the NEC weights can be calculated. We have compared the performance of five different weighted algorithms (FORE+AWOSEM, FORE+NECOSEM, ANWOSEM3D, SPOSEM3D and NECOSEM3D) on the basis of tumor detectablity. The comparison was done for simulated and clinical data. In the former case an analytical simulator was used. This is the ideal situation, since all the weighting factors can be exactly determined. For comparing the performance of the algorithms, we used the Non-Prewhitening Matched Filter (NPWMF) numerical observer. With some knowledge obtained from the simulation study we proceeded to the reconstruction of clinical data. In that case, it was necessary to devise a strategy for estimating the NEC weighting factors. The comparison between reconstructed images was done by a physician largely familiar with whole-body PET imaging
- âŠ