568 research outputs found

    An odyssey into local refinement and multilevel preconditioning III: Implementation and numerical experiments

    Get PDF
    In this paper, we examine a number of additive and multiplicative multilevel iterative methods and preconditioners in the setting of two-dimensional local mesh refinement. While standard multilevel methods are effective for uniform refinement-based discretizations of elliptic equations, they tend to be less effective for algebraic systems, which arise from discretizations on locally refined meshes, losing their optimal behavior in both storage and computational complexity. Our primary focus here is on Bramble, Pasciak, and Xu (BPX)-style additive and multiplicative multilevel preconditioners, and on various stabilizations of the additive and multiplicative hierarchical basis (HB) method, and their use in the local mesh refinement setting. In parts I and II of this trilogy, it was shown that both BPX and wavelet stabilizations of HB have uniformly bounded condition numbers on several classes of locally refined two- and three-dimensional meshes based on fairly standard (and easily implementable) red and red-green mesh refinement algorithms. In this third part of the trilogy, we describe in detail the implementation of these types of algorithms, including detailed discussions of the data structures and traversal algorithms we employ for obtaining optimal storage and computational complexity in our implementations. We show how each of the algorithms can be implemented using standard data types, available in languages such as C and FORTRAN, so that the resulting algorithms have optimal (linear) storage requirements, and so that the resulting multilevel method or preconditioner can be applied with optimal (linear) computational costs. We have successfully used these data structure ideas for both MATLAB and C implementations using the FEtk, an open source finite element software package. We finish the paper with a sequence of numerical experiments illustrating the effectiveness of a number of BPX and stabilized HB variants for several examples requiring local refinement

    Robust and parallel scalable iterative solutions for large-scale finite cell analyses

    Full text link
    The finite cell method is a highly flexible discretization technique for numerical analysis on domains with complex geometries. By using a non-boundary conforming computational domain that can be easily meshed, automatized computations on a wide range of geometrical models can be performed. Application of the finite cell method, and other immersed methods, to large real-life and industrial problems is often limited due to the conditioning problems associated with these methods. These conditioning problems have caused researchers to resort to direct solution methods, which signifi- cantly limit the maximum size of solvable systems. Iterative solvers are better suited for large-scale computations than their direct counterparts due to their lower memory requirements and suitability for parallel computing. These benefits can, however, only be exploited when systems are properly conditioned. In this contribution we present an Additive-Schwarz type preconditioner that enables efficient and parallel scalable iterative solutions of large-scale multi-level hp-refined finite cell analyses.Comment: 32 pages, 17 figure

    Optimal, scalable forward models for computing gravity anomalies

    Full text link
    We describe three approaches for computing a gravity signal from a density anomaly. The first approach consists of the classical "summation" technique, whilst the remaining two methods solve the Poisson problem for the gravitational potential using either a Finite Element (FE) discretization employing a multilevel preconditioner, or a Green's function evaluated with the Fast Multipole Method (FMM). The methods utilizing the PDE formulation described here differ from previously published approaches used in gravity modeling in that they are optimal, implying that both the memory and computational time required scale linearly with respect to the number of unknowns in the potential field. Additionally, all of the implementations presented here are developed such that the computations can be performed in a massively parallel, distributed memory computing environment. Through numerical experiments, we compare the methods on the basis of their discretization error, CPU time and parallel scalability. We demonstrate the parallel scalability of all these techniques by running forward models with up to 10810^8 voxels on 1000's of cores.Comment: 38 pages, 13 figures; accepted by Geophysical Journal Internationa
    • …
    corecore