2,327 research outputs found
Robust Cardiac Motion Estimation using Ultrafast Ultrasound Data: A Low-Rank-Topology-Preserving Approach
Cardiac motion estimation is an important diagnostic tool to detect heart
diseases and it has been explored with modalities such as MRI and conventional
ultrasound (US) sequences. US cardiac motion estimation still presents
challenges because of the complex motion patterns and the presence of noise. In
this work, we propose a novel approach to estimate the cardiac motion using
ultrafast ultrasound data. -- Our solution is based on a variational
formulation characterized by the L2-regularized class. The displacement is
represented by a lattice of b-splines and we ensure robustness by applying a
maximum likelihood type estimator. While this is an important part of our
solution, the main highlight of this paper is to combine a low-rank data
representation with topology preservation. Low-rank data representation
(achieved by finding the k-dominant singular values of a Casorati Matrix
arranged from the data sequence) speeds up the global solution and achieves
noise reduction. On the other hand, topology preservation (achieved by
monitoring the Jacobian determinant) allows to radically rule out distortions
while carefully controlling the size of allowed expansions and contractions.
Our variational approach is carried out on a realistic dataset as well as on a
simulated one. We demonstrate how our proposed variational solution deals with
complex deformations through careful numerical experiments. While maintaining
the accuracy of the solution, the low-rank preprocessing is shown to speed up
the convergence of the variational problem. Beyond cardiac motion estimation,
our approach is promising for the analysis of other organs that experience
motion.Comment: 15 pages, 10 figures, Physics in Medicine and Biology, 201
Toward Feature-Preserving Vector Field Compression
The objective of this work is to develop error-bounded lossy compression methods to preserve topological features in 2D and 3D vector fields. Specifically, we explore the preservation of critical points in piecewise linear and bilinear vector fields. We define the preservation of critical points as, without any false positive, false negative, or false type in the decompressed data, (1) keeping each critical point in its original cell and (2) retaining the type of each critical point (e.g., saddle and attracting node). The key to our method is to adapt a vertex-wise error bound for each grid point and to compress input data together with the error bound field using a modified lossy compressor. Our compression algorithm can be also embarrassingly parallelized for large data handling and in situ processing. We benchmark our method by comparing it with existing lossy compressors in terms of false positive/negative/type rates, compression ratio, and various vector field visualizations with several scientific applications
Principal Geodesic Analysis of Merge Trees (and Persistence Diagrams)
This paper presents a computational framework for the Principal Geodesic
Analysis of merge trees (MT-PGA), a novel adaptation of the celebrated
Principal Component Analysis (PCA) framework [87] to the Wasserstein metric
space of merge trees [92]. We formulate MT-PGA computation as a constrained
optimization problem, aiming at adjusting a basis of orthogonal geodesic axes,
while minimizing a fitting energy. We introduce an efficient, iterative
algorithm which exploits shared-memory parallelism, as well as an analytic
expression of the fitting energy gradient, to ensure fast iterations. Our
approach also trivially extends to extremum persistence diagrams. Extensive
experiments on public ensembles demonstrate the efficiency of our approach -
with MT-PGA computations in the orders of minutes for the largest examples. We
show the utility of our contributions by extending to merge trees two typical
PCA applications. First, we apply MT-PGA to data reduction and reliably
compress merge trees by concisely representing them by their first coordinates
in the MT-PGA basis. Second, we present a dimensionality reduction framework
exploiting the first two directions of the MT-PGA basis to generate
two-dimensional layouts of the ensemble. We augment these layouts with
persistence correlation views, enabling global and local visual inspections of
the feature variability in the ensemble. In both applications, quantitative
experiments assess the relevance of our framework. Finally, we provide a
lightweight C++ implementation that can be used to reproduce our results
Constrained parameterization with applications to graphics and image processing.
Surface parameterization is to establish a transformation that maps the points on a surface to a specified parametric domain. It has been widely applied to computer graphics and image processing fields. The challenging issue is that the usual positional constraints always result in triangle flipping in parameterizations (also called foldovers). Additionally, distortion is inevitable in parameterizations. Thus the rigid constraint is always taken into account. In general, the constraints are application-dependent. This thesis thus focuses on the various constraints depended on applications and investigates the foldover-free constrained parameterization approaches individually. Such constraints usually include, simple positional constraints, tradeoff of positional constraints and rigid constraint, and rigid constraint. From the perspective of applications, we aim at the foldover-free parameterization methods with positional constraints, the as-rigid-as-possible parameterization with positional constraints, and the well-shaped well-spaced pre-processing procedure for low-distortion parameterizations in this thesis. The first contribution of this thesis is the development of a RBF-based re-parameterization algorithm for the application of the foldover-free constrained texture mapping. The basic idea is to split the usual parameterization procedure into two steps, 2D parameterization with the constraints of convex boundaries and 2D re-parameterization with the interior positional constraints. Moreover, we further extend the 2D re-parameterization approach with the interior positional constraints to high dimensional datasets, such as, volume data and polyhedrons. The second contribution is the development of a vector field based deformation algorithm for 2D mesh deformation and image warping. Many presented deformation approaches are used to employ the basis functions (including our proposed RBF-based re-parameterization algorithm here). The main problem is that such algorithms have infinite support, that is, any local deformation always leads to small changes over the whole domain. Our presented vector field based algorithm can effectively carry on the local deformation while reducing distortion as much as possible. The third contribution is the development of a pre-processing for surface parameterization. Except the developable surfaces, the current parameterization approaches inevitably incur large distortion. To reduce distortion, we proposed a pre-processing procedure in this thesis, including mesh partition and mesh smoothing. As a result, the resulting meshes are partitioned into a set of small patches with rectangle-like boundaries. Moreover, they are well-shaped and well-spaced. This pre-processing procedure can evidently improve the quality of meshes for low-distortion parameterizations
Doctor of Philosophy
dissertationWith modern computational resources rapidly advancing towards exascale, large-scale simulations useful for understanding natural and man-made phenomena are becoming in- creasingly accessible. As a result, the size and complexity of data representing such phenom- ena are also increasing, making the role of data analysis to propel science even more integral. This dissertation presents research on addressing some of the contemporary challenges in the analysis of vector fields--an important type of scientific data useful for representing a multitude of physical phenomena, such as wind flow and ocean currents. In particular, new theories and computational frameworks to enable consistent feature extraction from vector fields are presented. One of the most fundamental challenges in the analysis of vector fields is that their features are defined with respect to reference frames. Unfortunately, there is no single ""correct"" reference frame for analysis, and an unsuitable frame may cause features of interest to remain undetected, thus creating serious physical consequences. This work develops new reference frames that enable extraction of localized features that other techniques and frames fail to detect. As a result, these reference frames objectify the notion of ""correctness"" of features for certain goals by revealing the phenomena of importance from the underlying data. An important consequence of using these local frames is that the analysis of unsteady (time-varying) vector fields can be reduced to the analysis of sequences of steady (time- independent) vector fields, which can be performed using simpler and scalable techniques that allow better data management by accessing the data on a per-time-step basis. Nevertheless, the state-of-the-art analysis of steady vector fields is not robust, as most techniques are numerical in nature. The residing numerical errors can violate consistency with the underlying theory by breaching important fundamental laws, which may lead to serious physical consequences. This dissertation considers consistency as the most fundamental characteristic of computational analysis that must always be preserved, and presents a new discrete theory that uses combinatorial representations and algorithms to provide consistency guarantees during vector field analysis along with the uncertainty visualization of unavoidable discretization errors. Together, the two main contributions of this dissertation address two important concerns regarding feature extraction from scientific data: correctness and precision. The work presented here also opens new avenues for further research by exploring more-general reference frames and more-sophisticated domain discretizations
Topological shape optimization design of continuum structures via an effective level set method
© 2016 The Author(s). This open access article is distributed under a Creative Commons Attribution (CC-BY) 4.0 license. This paper proposes a new level set method for topological shape optimization of continuum structure using radial basis function (RBF) and discrete wavelet transform (DWT). The boundary of the structure is implicitly represented as the zero level set of a higher-dimensional level set function. The interpolation of the implicit surface using RBF is introduced to decouple the spatial and temporal dependence of the level set function. In doing so, the Hamilton-Jacobi partial differential equation (PDE) that defines the motion of the level set function is transformed into an explicit parametric form, without requiring the direct solution of the complicated PDE using the finite difference method. Therefore, many more efficient gradient-based optimization algorithms can be applied to solve the optimization problem, via updating the expansion coefficients of the interpolant and then evolving the level set function and the boundary. Furthermore, the DWT is employed to handle the full matrix arising from the globally supported RBF interpolation. Several high stiffness but lightweight designs with smooth and clear structural boundaries are optimized and presented. The numerical results show that the proposed method can remarkably increase the efficiency in the topology optimization design of both the 2D and 3D structures
- …