886 research outputs found

    The cost of continuity: performance of iterative solvers on isogeometric finite elements

    Get PDF
    In this paper we study how the use of a more continuous set of basis functions affects the cost of solving systems of linear equations resulting from a discretized Galerkin weak form. Specifically, we compare performance of linear solvers when discretizing using C0C^0 B-splines, which span traditional finite element spaces, and Cp−1C^{p-1} B-splines, which represent maximum continuity. We provide theoretical estimates for the increase in cost of the matrix-vector product as well as for the construction and application of black-box preconditioners. We accompany these estimates with numerical results and study their sensitivity to various grid parameters such as element size hh and polynomial order of approximation pp. Finally, we present timing results for a range of preconditioning options for the Laplace problem. We conclude that the matrix-vector product operation is at most \slfrac{33p^2}{8} times more expensive for the more continuous space, although for moderately low pp, this number is significantly reduced. Moreover, if static condensation is not employed, this number further reduces to at most a value of 8, even for high pp. Preconditioning options can be up to p3p^3 times more expensive to setup, although this difference significantly decreases for some popular preconditioners such as Incomplete LU factorization

    On Stochastic Error and Computational Efficiency of the Markov Chain Monte Carlo Method

    Full text link
    In Markov Chain Monte Carlo (MCMC) simulations, the thermal equilibria quantities are estimated by ensemble average over a sample set containing a large number of correlated samples. These samples are selected in accordance with the probability distribution function, known from the partition function of equilibrium state. As the stochastic error of the simulation results is significant, it is desirable to understand the variance of the estimation by ensemble average, which depends on the sample size (i.e., the total number of samples in the set) and the sampling interval (i.e., cycle number between two consecutive samples). Although large sample sizes reduce the variance, they increase the computational cost of the simulation. For a given CPU time, the sample size can be reduced greatly by increasing the sampling interval, while having the corresponding increase in variance be negligible if the original sampling interval is very small. In this work, we report a few general rules that relate the variance with the sample size and the sampling interval. These results are observed and confirmed numerically. These variance rules are derived for the MCMC method but are also valid for the correlated samples obtained using other Monte Carlo methods. The main contribution of this work includes the theoretical proof of these numerical observations and the set of assumptions that lead to them

    Efficient mass and stiffness matrix assembly via weighted Gaussian quadrature rules for B-splines

    Get PDF
    Calabr{\`o} et al. [10] changed the paradigm of the mass and stiffness computation from the traditional element-wise assembly to a row-wise concept, showing that the latter one offers integration that may be orders of magnitude faster. Considering a B-spline basis function as a non-negative measure, each mass matrix row is integrated by its own quadrature rule with respect to that measure. Each rule is easy to compute as it leads to a linear system of equations, however, the quadrature rules are of the Newton-Cotes type, that is, they require a number of quadrature points that is equal to the dimension of the spline space. In this work, we propose weighted quadrature rules of Gaussian type which require the minimum number of quadrature points while guaranteeing exactness of integration with respect to the weight function. The weighted Gaussian rules arise as solutions of non-linear systems of equations. We derive rules for the mass and stiffness matrices for uniform C1C^1 quadratic and C2C^2 cubic isogeometric discretizations. In each parameter direction, our rules require locally only p+1p+1 quadrature points, pp being the polynomial degree. While the nodes cannot be reused for various weight functions as in [10], the computational cost of the mass and stiffness matrix assembly is comparable.RYC-2017-2264

    Goal-oriented adaptivity for a conforming residual minimization method in a dual discontinuous Galerkin norm

    Get PDF
    We propose a goal-oriented mesh-adaptive algorithm for a finite element method stabilized via residual minimization on dual discontinuous-Galerkin norms. By solving a saddle-point problem, this residual minimization delivers a stable continuous approximation to the solution on each mesh instance and a residual projection onto a broken polynomial space, which is a robust error estimator to minimize the discrete energy norm via automatic mesh refinement. In this work, we propose and analyze a goal-oriented adaptive algorithm for this stable residual minimization. We solve the primal and adjoint problems considering the same saddle-point formulation and different right-hand sides. By solving a third stable problem, we obtain two efficient error estimates to guide goal oriented adaptivity. We illustrate the performance of this goal-oriented adaptive strategy on advection-diffusion reaction problems

    Reducing spatial discretization error on coarse CFD simulations using an openFOAM-embedded deep learning framework

    Get PDF
    We propose a method for reducing the spatial discretization error of coarse computational fluid dynamics (CFD) problems by enhancing the quality of low-resolution simulations using deep learning. We feed the model with fine-grid data after projecting it to the coarse-grid discretization. We substitute the default differencing scheme for the convection term by a feed-forward neural network that interpolates velocities from cell centers to face values to produce velocities that approximate the down-sampled fine-grid data well. The deep learning framework incorporates the open-source CFD code OpenFOAM, resulting in an end-to-end differentiable model. We automatically differentiate the CFD physics using a discrete adjoint code version. We present a fast communication method between TensorFlow (Python) and OpenFOAM (c++) that accelerates the training process. We applied the model to the flow past a square cylinder problem, reducing the error from 120% to 25% in the velocity for simulations inside the training distribution compared to the traditional solver using an x8 coarser mesh. For simulations outside the training distribution, the error reduction in the velocities was about 50%. The training is affordable in terms of time and data samples since the architecture exploits the local features of the physics.PID2023-146678OB-I00 PRE2020-09309

    Numerical simulation of spheres moving and colliding close to bed streams, with a complete characterization of turbulence

    Get PDF
    River morphodynamics and sediment transportMechanics of sediment transpor

    Parallel refined Isogeometric Analysis in 3D

    Get PDF
    We study three-dimensional isogeometric analysis (IGA) and the solution of the resulting system of linear equations via a direct solver. IGA uses highly continuous Cp−1C^{p-1} basis functions, which provide multiple benefits in terms of stability and convergence properties. However, smooth basis significantly deteriorate the direct solver performance and its parallel scalability. As a partial remedy for this, refined Isogeometric Analysis (rIGA) method improves the sequential execution of direct solvers. The refinement strategy enriches traditional highly-continuous Cp−1C^{p-1} IGA spaces by introducing low-continuity C0C^{0} 0-hyperplanes along the boundaries of certain pre-defined macro-elements. In this work, propose a solution strategy for rIGA for parallel distributed memory machines and compare the computational costs of solving rIGA vs IGA discretizations. We verify our estimates with parallel numerical experiments. Results show that the weak parallel scalability of the direct solver improves approximately by a factor of p2p^{2} when considering rIGA discretizations rather than highly-continuous IGA spaces
    • …
    corecore