242 research outputs found

    Anisotropic Mesh Adaptation for Image Representation

    Get PDF
    Triangular meshes have gained much interest in image representation and have been widely used in image processing. This paper introduces a framework of anisotropic mesh adaptation (AMA) methods to image representation and proposes a GPRAMA method that is based on AMA and greedy-point removal (GPR) scheme. Different than many other methods that triangulate sample points to form the mesh, the AMA methods start directly with a triangular mesh and then adapt the mesh based on a user-defined metric tensor to represent the image. The AMA methods have clear mathematical framework and provides flexibility for both image representation and image reconstruction. A mesh patching technique is developed for the implementation of the GPRAMA method, which leads to an improved version of the popular GPRFS-ED method. The GPRAMA method can achieve better quality than the GPRFS-ED method but with lower computational cost.Comment: 25 pages, 15 figure

    Interval simplex splines for scientific databases

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Ocean Engineering, 1995.Includes bibliographical references (p. 130-138).by Jingfang Zhou.Ph.D

    Doctor of Philosophy

    Get PDF
    dissertationSmoothness-increasing accuracy-conserving (SIAC) filters were introduced as a class of postprocessing techniques to ameliorate the quality of numerical solutions of discontinuous Galerkin (DG) simulations. SIAC filtering works to eliminate the oscillations in the error by introducing smoothness back to the DG field and raises the accuracy in the L2-n o rm up to its natural superconvergent accuracy in the negative-order norm. The increased smoothness in the filtered DG solutions can then be exploited by simulation postprocessing tools such as streamline integrators where the absence of continuity in the data can lead to erroneous visualizations. However, lack of extension of this filtering technique, both theoretically and computationally, to nontrivial mesh structures along with the expensive core operators have been a hindrance towards the application of the SIAC filters to more realistic simulations. In this dissertation, we focus on the numerical solutions of linear hyperbolic equations solved with the discontinuous Galerkin scheme and provide a thorough analysis of SIAC filtering applied to such solution data. In particular, we investigate how the use of different quadrature techniques could mitigate the extensive processing required when filtering over the whole computational field. Moreover, we provide detailed and efficient algorithms that a numerical practitioner requires to know in order to implement this filtering technique effectively. In our first attempt to expand the application scope of this filtering technique, we demonstrate both mathematically and through numerical examples that it is indeed possible to observe SIAC filtering characteristics when applied to numerical solutions obtained over structured triangular meshes. We further provide a thorough investigation of the interplay between mesh geometry and filtering. Building upon these promising results, we present how SIAC filtering could be applied to gain higher accuracy and smoothness when dealing with totally unstructured triangular meshes. Lastly, we provide the extension of our filtering scheme to structured tetrahedral meshes. Guidelines and future work regarding the application of the SIAC filter in the visualization domain are also presented. We further note that throughout this document, the terms postprocessing and filtering will be used interchangeably

    Graph Wedgelets: Adaptive Data Compression on Graphs based on Binary Wedge Partitioning Trees and Geometric Wavelets

    Full text link
    We introduce graph wedgelets - a tool for data compression on graphs based on the representation of signals by piecewise constant functions on adaptively generated binary graph partitionings. The adaptivity of the partitionings, a key ingredient to obtain sparse representations of a graph signal, is realized in terms of recursive wedge splits adapted to the signal. For this, we transfer adaptive partitioning and compression techniques known for 2D images to general graph structures and develop discrete variants of continuous wedgelets and binary space partitionings. We prove that continuous results on best m-term approximation with geometric wavelets can be transferred to the discrete graph setting and show that our wedgelet representation of graph signals can be encoded and implemented in a simple way. Finally, we illustrate that this graph-based method can be applied for the compression of images as well.Comment: 12 pages, 10 figure

    Optimising Spatial and Tonal Data for PDE-based Inpainting

    Full text link
    Some recent methods for lossy signal and image compression store only a few selected pixels and fill in the missing structures by inpainting with a partial differential equation (PDE). Suitable operators include the Laplacian, the biharmonic operator, and edge-enhancing anisotropic diffusion (EED). The quality of such approaches depends substantially on the selection of the data that is kept. Optimising this data in the domain and codomain gives rise to challenging mathematical problems that shall be addressed in our work. In the 1D case, we prove results that provide insights into the difficulty of this problem, and we give evidence that a splitting into spatial and tonal (i.e. function value) optimisation does hardly deteriorate the results. In the 2D setting, we present generic algorithms that achieve a high reconstruction quality even if the specified data is very sparse. To optimise the spatial data, we use a probabilistic sparsification, followed by a nonlocal pixel exchange that avoids getting trapped in bad local optima. After this spatial optimisation we perform a tonal optimisation that modifies the function values in order to reduce the global reconstruction error. For homogeneous diffusion inpainting, this comes down to a least squares problem for which we prove that it has a unique solution. We demonstrate that it can be found efficiently with a gradient descent approach that is accelerated with fast explicit diffusion (FED) cycles. Our framework allows to specify the desired density of the inpainting mask a priori. Moreover, is more generic than other data optimisation approaches for the sparse inpainting problem, since it can also be extended to nonlinear inpainting operators such as EED. This is exploited to achieve reconstructions with state-of-the-art quality. We also give an extensive literature survey on PDE-based image compression methods

    Feature-sensitive and Adaptive Image Triangulation: A Super-pixel-based Scheme for Image Segmentation and Mesh Generation

    Get PDF
    With increasing utilization of various imaging techniques (such as CT, MRI and PET) in medical fields, it is often in great need to computationally extract the boundaries of objects of interest, a process commonly known as image segmentation. While numerous approaches have been proposed in literature on automatic/semi-automatic image segmentation, most of these approaches are based on image pixels. The number of pixels in an image can be huge, especially for 3D imaging volumes, which renders the pixel-based image segmentation process inevitably slow. On the other hand, 3D mesh generation from imaging data has become important not only for visualization and quantification but more critically for finite element based numerical simulation. Traditionally image-based mesh generation follows such a procedure as: (1) image boundary segmentation, (2) surface mesh generation from segmented boundaries, and (3) volumetric (e.g., tetrahedral) mesh generation from surface meshes. These three majors steps have been commonly treated as separate algorithms/steps and hence image information, once segmented, is not considered any more in mesh generation. In this thesis, we investigate a super-pixel based scheme that integrates both image segmentation and mesh generation into a single method, making mesh generation truly an image-incorporated approach. Our method, called image content-aware mesh generation, consists of several main steps. First, we generate a set of feature-sensitive, and adaptively distributed points from 2D grayscale images or 3D volumes. A novel image edge enhancement method via randomized shortest paths is introduced to be an optional choice to generate the features’ boundary map in mesh node generation step. Second, a Delaunay-triangulation generator (2D) or tetrahedral mesh generator (3D) is then utilized to generate a 2D triangulation or 3D tetrahedral mesh. The generated triangulation (or tetrahedralization) provides an adaptive partitioning of a given image (or volume). Each cluster of pixels within a triangle (or voxels within a tetrahedron) is called a super-pixel, which forms one of the nodes of a graph and adjacent super-pixels give an edge of the graph. A graph-cut method is then applied to the graph to define the boundary between two subsets of the graph, resulting in good boundary segmentations with high quality meshes. Thanks to the significantly reduced number of elements (super-pixels) as compared to that of pixels in an image, the super-pixel based segmentation method has tremendously improved the segmentation speed, making it feasible for real-time feature detection. In addition, the incorporation of image segmentation into mesh generation makes the generated mesh well adapted to image features, a desired property known as feature-preserving mesh generation

    Computations of Delaunay and Higher Order Triangulations, with Applications to Splines

    Get PDF
    Digital data that consist of discrete points are frequently captured and processed by scientific and engineering applications. Due to the rapid advance of new data gathering technologies, data set sizes are increasing, and the data distributions are becoming more irregular. These trends call for new computational tools that are both efficient enough to handle large data sets and flexible enough to accommodate irregularity. A mathematical foundation that is well-suited for developing such tools is triangulation, which can be defined for discrete point sets with little assumption about their distribution. The potential benefits from using triangulation are not fully exploited. The challenges fundamentally stem from the complexity of the triangulation structure, which generally takes more space to represent than the input points. This complexity makes developing a triangulation program a delicate task, particularly when it is important that the program runs fast and robustly over large data. This thesis addresses these challenges in two parts. The first part concentrates on techniques designed for efficiently and robustly computing Delaunay triangulations of three kinds of practical data: the terrain data from LIDAR sensors commonly found in GIS, the atom coordinate data used for biological applications, and the time varying volume data generated from from scientific simulations. The second part addresses the problem of defining spline spaces over triangulations in two dimensions. It does so by generalizing Delaunay configurations, defined as follows. For a given point set P in two dimensions, a Delaunay configuration is a pair of subsets (T, I) from P, where T, called the boundary set, is a triplet and I, called the interior set, is the set of points that fall in the circumcircle through T. The size of the interior set is the degree of the configuration. As recently discovered by Neamtu (2004), for a chosen point set, the set of all degree k Delaunay configurations can be associated with a set of degree k plus 1 splines that form the basis of a spline space. In particular, for the trivial case of k equals 0, the spline space coincides with the PL interpolation functions over the Delaunay triangulation. Neamtu’s definition of the spline space relies only on a few structural properties of the Delaunay configurations. This raises the question whether there exist other sets of configurations with identical structural properties. If there are, then these sets of configurations—let us call them generalized configurations from hereon—can be substituted for Delaunay configurations in Neamtu’s definition of spline space thereby yielding a family of splines over the same point set

    High-order adaptive methods for computing invariant manifolds of maps

    Get PDF
    The author presents efficient and accurate numerical methods for computing invariant manifolds of maps which arise in the study of dynamical systems. In order to decrease the number of points needed to compute a given curve/surface, he proposes using higher-order interpolation/approximation techniques from geometric modeling. He uses B´ezier curves/triangles, fundamental objects in curve/surface design, to create adaptive methods. The methods are based on tolerance conditions derived from properties of B´ezier curves/triangles. The author develops and tests the methods for an ordinary parametric curve; then he adapts these methods to invariant manifolds of planar maps. Next, he develops and tests the method for parametric surfaces and then he adapts this method to invariant manifolds of three-dimensional maps

    Extracting Geometric Structures in Images with Delaunay Point Processes

    Get PDF
    International audienceWe introduce Delaunay Point Processes, a framework for the extraction of geometric structures from images. Our approach simultaneously locates and groups geometric primitives (line segments, triangles) to form extended structures (line networks, polygons) for a variety of image analysis tasks. Similarly to traditional point processes, our approach uses Markov Chain Monte Carlo to minimize an energy that balances fidelity to the input image data with geometric priors on the output structures. However, while existing point processes struggle to model structures composed of interconnected components, we propose to embed the point process into a Delaunay triangulation, which provides high-quality connectivity by construction. We further leverage key properties of the Delaunay triangulation to devise a fast Markov Chain Monte Carlo sampler. We demonstrate the flexibility of our approach on a variety of applications, including line network extraction, object contouring, and mesh-based image compression
    • …
    corecore