951 research outputs found

    Deep Vectorization of Technical Drawings

    Full text link
    We present a new method for vectorization of technical line drawings, such as floor plans, architectural drawings, and 2D CAD images. Our method includes (1) a deep learning-based cleaning stage to eliminate the background and imperfections in the image and fill in missing parts, (2) a transformer-based network to estimate vector primitives, and (3) optimization procedure to obtain the final primitive configurations. We train the networks on synthetic data, renderings of vector line drawings, and manually vectorized scans of line drawings. Our method quantitatively and qualitatively outperforms a number of existing techniques on a collection of representative technical drawings

    Simulation of flows with violent free surface motion and moving objects using unstructured grids

    Get PDF
    This is the peer reviewed version of the following article: [Löhner, R. , Yang, C. and Oñate, E. (2007), Simulation of flows with violent free surface motion and moving objects using unstructured grids. Int. J. Numer. Meth. Fluids, 53: 1315-1338. doi:10.1002/fld.1244], which has been published in final form at https://doi.org/10.1002/fld.1244. This article may be used for non-commercial purposes in accordance with Wiley Terms and Conditions for Self-Archiving.A volume of fluid (VOF) technique has been developed and coupled with an incompressible Euler/Navier–Stokes solver operating on adaptive, unstructured grids to simulate the interactions of extreme waves and three-dimensional structures. The present implementation follows the classic VOF implementation for the liquid–gas system, considering only the liquid phase. Extrapolation algorithms are used to obtain velocities and pressure in the gas region near the free surface. The VOF technique is validated against the classic dam-break problem, as well as series of 2D sloshing experiments and results from SPH calculations. These and a series of other examples demonstrate that the ability of the present approach to simulate violent free surface flows with strong nonlinear behaviour.Peer ReviewedPostprint (author's final draft

    Supercomputer implementation of finite element algorithms for high speed compressible flows

    Get PDF
    Prediction of compressible flow phenomena using the finite element method is of recent origin and considerable interest. Two shock capturing finite element formulations for high speed compressible flows are described. A Taylor-Galerkin formulation uses a Taylor series expansion in time coupled with a Galerkin weighted residual statement. The Taylor-Galerkin algorithms use explicit artificial dissipation, and the performance of three dissipation models are compared. A Petrov-Galerkin algorithm has as its basis the concepts of streamline upwinding. Vectorization strategies are developed to implement the finite element formulations on the NASA Langley VPS-32. The vectorization scheme results in finite element programs that use vectors of length of the order of the number of nodes or elements. The use of the vectorization procedure speeds up processing rates by over two orders of magnitude. The Taylor-Galerkin and Petrov-Galerkin algorithms are evaluated for 2D inviscid flows on criteria such as solution accuracy, shock resolution, computational speed and storage requirements. The convergence rates for both algorithms are enhanced by local time-stepping schemes. Extension of the vectorization procedure for predicting 2D viscous and 3D inviscid flows are demonstrated. Conclusions are drawn regarding the applicability of the finite element procedures for realistic problems that require hundreds of thousands of nodes

    Photo2ClipArt: Image Abstraction and Vectorization Using Layered Linear Gradients

    Get PDF
    International audienceWe present a method to create vector cliparts from photographs. Our approach aims at reproducing two key properties of cliparts: they should be easily editable, and they should represent image content in a clean, simplified way. We observe that vector artists satisfy both of these properties by modeling cliparts with linear color gradients, which have a small number of parameters and approximate well smooth color variations. In addition, skilled artists produce intricate yet editable artworks by stacking multiple gradients using opaque and semi-transparent layers. Motivated by these observations, our goal is to decompose a bitmap photograph into a stack of layers, each layer containing a vector path filled with a linear color gradient. We cast this problem as an optimization that jointly assigns each pixel to one or more layer and finds the gradient parameters of each layer that best reproduce the input. Since a trivial solution would consist in assigning each pixel to a different, opaque layer, we complement our objective with a simplicity term that favors decompositions made of few, semi-transparent layers. However, this formulation results in a complex combinatorial problem combining discrete unknowns (the pixel assignments) and continuous unknowns (the layer parameters). We propose a Monte Carlo Tree Search algorithm that efficiently explores this solution space by leveraging layering cues at image junctions. We demonstrate the effectiveness of our method by reverse-engineering existing cliparts and by creating original cliparts from studio photographs

    Vectorizing Distributed Homology with Deep Set of Set Networks

    Get PDF
    Distributed homology, a topological invariant, holds potential as an instrument for uncov- ering insights into the structural characteristics of complex data. By considering both the density and connectivity of topological spaces, it offers the potential for a more detailed and stable understanding of the underlying structure of data sets. This is particularly beneficial when confronting noisy, real-world data. Despite its potential, the complexity and unstructured nature of distributed homology pose hurdles for practical use. This thesis tackles these issues by proposing a novel pipeline that fuses distributed homology and supervised learning techniques. The goal is to facilitate the effective incorporation of distributed homology into a wide array of supervised learning tasks. Our approach is anchored on the DeepSet network, an architecture adept at managing set inputs. Using this, we devise a comprehensive framework specifically designed to handle inputs composed of a set of sets. Furthermore, we present a dedicated architecture for distributed homology, designed to boost robustness to noise and overall performance. This approach shows marked improvements over full persistent homology methods for both synthetic and real data. While our results may not yet rival state-of-the-art performance on real data, they demonstrate the potential for distributed invariants to enhance the efficiency of topolog- ical approaches. This indicates a promising avenue for future research and development, contributing to the refinement of topological data analysis.Masteroppgave i informatikkINF399MAMN-PROGMAMN-IN
    corecore