3 research outputs found

    Constrained parameterization with applications to graphics and image processing.

    Get PDF
    Surface parameterization is to establish a transformation that maps the points on a surface to a specified parametric domain. It has been widely applied to computer graphics and image processing fields. The challenging issue is that the usual positional constraints always result in triangle flipping in parameterizations (also called foldovers). Additionally, distortion is inevitable in parameterizations. Thus the rigid constraint is always taken into account. In general, the constraints are application-dependent. This thesis thus focuses on the various constraints depended on applications and investigates the foldover-free constrained parameterization approaches individually. Such constraints usually include, simple positional constraints, tradeoff of positional constraints and rigid constraint, and rigid constraint. From the perspective of applications, we aim at the foldover-free parameterization methods with positional constraints, the as-rigid-as-possible parameterization with positional constraints, and the well-shaped well-spaced pre-processing procedure for low-distortion parameterizations in this thesis. The first contribution of this thesis is the development of a RBF-based re-parameterization algorithm for the application of the foldover-free constrained texture mapping. The basic idea is to split the usual parameterization procedure into two steps, 2D parameterization with the constraints of convex boundaries and 2D re-parameterization with the interior positional constraints. Moreover, we further extend the 2D re-parameterization approach with the interior positional constraints to high dimensional datasets, such as, volume data and polyhedrons. The second contribution is the development of a vector field based deformation algorithm for 2D mesh deformation and image warping. Many presented deformation approaches are used to employ the basis functions (including our proposed RBF-based re-parameterization algorithm here). The main problem is that such algorithms have infinite support, that is, any local deformation always leads to small changes over the whole domain. Our presented vector field based algorithm can effectively carry on the local deformation while reducing distortion as much as possible. The third contribution is the development of a pre-processing for surface parameterization. Except the developable surfaces, the current parameterization approaches inevitably incur large distortion. To reduce distortion, we proposed a pre-processing procedure in this thesis, including mesh partition and mesh smoothing. As a result, the resulting meshes are partitioned into a set of small patches with rectangle-like boundaries. Moreover, they are well-shaped and well-spaced. This pre-processing procedure can evidently improve the quality of meshes for low-distortion parameterizations

    Doctor of Philosophy

    Get PDF
    dissertationWith modern computational resources rapidly advancing towards exascale, large-scale simulations useful for understanding natural and man-made phenomena are becoming in- creasingly accessible. As a result, the size and complexity of data representing such phenom- ena are also increasing, making the role of data analysis to propel science even more integral. This dissertation presents research on addressing some of the contemporary challenges in the analysis of vector fields--an important type of scientific data useful for representing a multitude of physical phenomena, such as wind flow and ocean currents. In particular, new theories and computational frameworks to enable consistent feature extraction from vector fields are presented. One of the most fundamental challenges in the analysis of vector fields is that their features are defined with respect to reference frames. Unfortunately, there is no single ""correct"" reference frame for analysis, and an unsuitable frame may cause features of interest to remain undetected, thus creating serious physical consequences. This work develops new reference frames that enable extraction of localized features that other techniques and frames fail to detect. As a result, these reference frames objectify the notion of ""correctness"" of features for certain goals by revealing the phenomena of importance from the underlying data. An important consequence of using these local frames is that the analysis of unsteady (time-varying) vector fields can be reduced to the analysis of sequences of steady (time- independent) vector fields, which can be performed using simpler and scalable techniques that allow better data management by accessing the data on a per-time-step basis. Nevertheless, the state-of-the-art analysis of steady vector fields is not robust, as most techniques are numerical in nature. The residing numerical errors can violate consistency with the underlying theory by breaching important fundamental laws, which may lead to serious physical consequences. This dissertation considers consistency as the most fundamental characteristic of computational analysis that must always be preserved, and presents a new discrete theory that uses combinatorial representations and algorithms to provide consistency guarantees during vector field analysis along with the uncertainty visualization of unavoidable discretization errors. Together, the two main contributions of this dissertation address two important concerns regarding feature extraction from scientific data: correctness and precision. The work presented here also opens new avenues for further research by exploring more-general reference frames and more-sophisticated domain discretizations

    Non-linear subdivision of univariate signals and discrete surfaces

    Get PDF
    During the last 20 years, the joint expansion of computing power, computer graphics, networking capabilities and multiresolution analysis have stimulated several research domains, and developed the need for new types of data such as 3D models, i.e. discrete surfaces. In the intersection between multiresolution analysis and computer graphics, subdivision methods, i.e. iterative refinement procedures of curves or surfaces, have a non-negligible place, since they are a basic component needed to adapt existing multiresolution techniques dedicated to signals and images to more complicated data such as discrete surfaces represented by polygonal meshes. Such representations are of great interest since they make polygonal meshes nearly as exible as higher level 3D model representations, such as piecewise polynomial based surfaces (e.g. NURBS, B-splines...). The generalization of subdivision methods from univariate data to polygonal meshes is relatively simple in case of a regular mesh but becomes less straightforward when handling irregularities. Moreover, in the linear univariate case, obtaining a smoother limit curve is achieved by increasing the size of the support of the subdivision scheme, which is not a trivial operation in the case of a surface subdivision scheme without a priori assumptions on the mesh. While many linear subdivision methods are available, the studies concerning more general non-linear methods are relatively sparse, whereas such techniques could be used to achieve better results without increasing the size support. The goal of this study is to propose and to analyze a binary non-linear interpolatory subdivision method. The proposed technique uses local polar coordinates to compute the positions of the newly inserted points. It is shown that the method converges toward continuous limit functions. The proposed univariate scheme is extended to triangular meshes, possibly with boundaries. In order to evaluate characteristics of the proposed scheme which are not proved analytically, numerical estimates to study convergence, regularity of the limit function and approximation order are studied and validated using known linear schemes of identical support. The convergence criterion is adapted to surface subdivision via a Hausdorff distance-based metric. The evolution of Gaussian and mean curvature of limit surfaces is also studied and compared against theoretical values when available. An application of surface subdivision to build a multiresolution representation of 3D models is also studied. In particular, the efficiency of such a representation for compression and in terms of rate-distortion of such a representation is shown. An alternate to the initial SPIHT-based encoding, based on the JPEG 2000 image compression standard method. This method makes possible partial decoding of the compressed model in both SNR-progressive and level-progressive ways, while adding only a minimal overhead when compared to SPIHT
    corecore