4,116 research outputs found

    The cosmological simulation code GADGET-2

    Full text link
    We discuss the cosmological simulation code GADGET-2, a new massively parallel TreeSPH code, capable of following a collisionless fluid with the N-body method, and an ideal gas by means of smoothed particle hydrodynamics (SPH). Our implementation of SPH manifestly conserves energy and entropy in regions free of dissipation, while allowing for fully adaptive smoothing lengths. Gravitational forces are computed with a hierarchical multipole expansion, which can optionally be applied in the form of a TreePM algorithm, where only short-range forces are computed with the `tree'-method while long-range forces are determined with Fourier techniques. Time integration is based on a quasi-symplectic scheme where long-range and short-range forces can be integrated with different timesteps. Individual and adaptive short-range timesteps may also be employed. The domain decomposition used in the parallelisation algorithm is based on a space-filling curve, resulting in high flexibility and tree force errors that do not depend on the way the domains are cut. The code is efficient in terms of memory consumption and required communication bandwidth. It has been used to compute the first cosmological N-body simulation with more than 10^10 dark matter particles, reaching a homogeneous spatial dynamic range of 10^5 per dimension in a 3D box. It has also been used to carry out very large cosmological SPH simulations that account for radiative cooling and star formation, reaching total particle numbers of more than 250 million. We present the algorithms used by the code and discuss their accuracy and performance using a number of test problems. GADGET-2 is publicly released to the research community.Comment: submitted to MNRAS, 31 pages, 20 figures (reduced resolution), code available at http://www.mpa-garching.mpg.de/gadge

    Acceleration of a Full-scale Industrial CFD Application with OP2

    Get PDF

    Multiphase modelling of vascular tumour growth in two spatial dimensions

    Get PDF
    In this paper we present a continuum mathematical model of vascular tumour growth which is based on a multiphase framework in which the tissue is decomposed into four distinct phases and the principles of conservation of mass and momentum are applied to the normal/healthy cells, tumour cells, blood vessels and extracellular material. The inclusion of a diffusible nutrient, supplied by the blood vessels, allows the vasculature to have a nonlocal influence on the other phases. Two-dimensional computational simulations are carried out on unstructured, triangular meshes to allow a natural treatment of irregular geometries, and the tumour boundary is captured as a diffuse interface on this mesh, thereby obviating the need to explicitly track the (potentially highly irregular and ill-defined) tumour boundary. A hybrid finite volume/finite element algorithm is used to discretise the continuum model: the application of a conservative, upwind, finite volume scheme to the hyperbolic mass balance equations and a finite element scheme with a stable element pair to the generalised Stokes equations derived from momentum balance, leads to a robust algorithm which does not use any form of artificial stabilisation. The use of a matrix-free Newton iteration with a finite element scheme for the nutrient reaction-diffusion equations allows full nonlinearity in the source terms of the mathematical model. Numerical simulations reveal that this four-phase model reproduces the characteristic pattern of tumour growth in which a necrotic core forms behind an expanding rim of well-vascularised proliferating tumour cells. The simulations consistently predict linear tumour growth rates. The dependence of both the speed with which the tumour grows and the irregularity of the invading tumour front on the model parameters are investigated

    Simulation tools for biomechanical applications with PGD-based reduced order models

    Get PDF
    Cotutela Universitat Politècnica de Catalunya i Università degli Studi di PaviaNumerical simulation tools are generally used in all modern engineering fields, especially those having difficulties in performing large number of practical experiments, such as biomechanics. Among the computational methods, Finite Element (FE) is an essential tool. Nowadays, the fast-growing computational techniques, from the upgrading hardware to the emerging of novel algorithm, have already enabled extensive applications in biomechanics. For applications that require fast response and/or multiple queries, Reduced Order Modelling (ROM) methods have been developed based on existing methods such as FE, and have eventually enabled real-time numerical simulation for a large variety of engineering problems. In this thesis, several novel computational techniques are developed to explore the capability of Proper Generalised Decomposition (PGD), which is an important approach of ROM. To assess the usability of the PGD-based ROM for biomechanical applications, a real human femur bone is chosen to study its mechanical behaviour as an example. Standard image-based modelling procedure in biomechanics is performed to create an FE model which is then validated with in vitro experimental results. As a basis of this work, the medical image processing has to be performed, in order to generate an available FE model. This model is validated according to data collected from a previously performed \textit{in vitro} experimental test. The full procedure of image-based model generation and the validation of generated model is described in Chapter 2. As a major objective of this thesis, a non-intrusive scheme for the PGD framework is developed in Chapter 3. It is implemented using in-house developed Matlab (Mathworks, USA) code to conduct the PGD work flow, and calling Abaqus as an external solver for devised fictitious mechanical problems. The transformation of data from computed tomography (CT) image set to FE model including inhomogeneous material properties is subjected to some physical constraints, and when applying the load, there are also geometric constraints limiting the locations where load could be applied. These constraints will lead to a constrained parameter space, which possibly has difficulty to be separated in a Cartesian fashion. Therefore, a novel strategy to separate the parameters in a collective manner is proposed in Chapter 4. Chapter 5 details a comprehensive application in biomechanics, the methodologies proposed in Chapter 3 and 4 are applied on the practical model generated in Chapter 2. As a typical application of the PGD vademecum, a material property identification problem is discussed. Further PGD vademecum is generated using the identified material properties with variable loading locations, and with this vademecum, real-time mechanical response of the femur is available. In addition, for the purpose of extending the methodologies to orthotropic materials, which is commonly used in biomechanics, in Chapter 6 another linear elastic model is investigated with the non-intrusive PGD scheme. Nowadays, isogeometric analysis (IGA) is a very popular tool in computational mechanics. It is appealing to take advantage of non-uniform rational B-splines (NURBS) to discretise the model. For PGD, using B-splines for the discretisation of the parameter space could improve the quality of vademecum, especially for problems involving sensitivities with respect to the parameters during the online computations. It is important and necessary to extend the PGD framework to nonlinear solid mechanics, because most biological soft tissues have been observed nonlinear mechanical behaviours. Consequently, in Chapter 7 we have developed a PGD framework for the St.Venant-Kirchhoff constitutive model using the Picard linearisation which is consistent with the fixed-point iteration algorithm commonly used in PGD. In Chapter 8, conclusive remarks are addressed as well as forecasts of possible future works.Postprint (published version

    On a general implementation of hh- and pp-adaptive curl-conforming finite elements

    Get PDF
    Edge (or N\'ed\'elec) finite elements are theoretically sound and widely used by the computational electromagnetics community. However, its implementation, specially for high order methods, is not trivial, since it involves many technicalities that are not properly described in the literature. To fill this gap, we provide a comprehensive description of a general implementation of edge elements of first kind within the scientific software project FEMPAR. We cover into detail how to implement arbitrary order (i.e., pp-adaptive) elements on hexahedral and tetrahedral meshes. First, we set the three classical ingredients of the finite element definition by Ciarlet, both in the reference and the physical space: cell topologies, polynomial spaces and moments. With these ingredients, shape functions are automatically implemented by defining a judiciously chosen polynomial pre-basis that spans the local finite element space combined with a change of basis to automatically obtain a canonical basis with respect to the moments at hand. Next, we discuss global finite element spaces putting emphasis on the construction of global shape functions through oriented meshes, appropriate geometrical mappings, and equivalence classes of moments, in order to preserve the inter-element continuity of tangential components of the magnetic field. Finally, we extend the proposed methodology to generate global curl-conforming spaces on non-conforming hierarchically refined (i.e., hh-adaptive) meshes with arbitrary order finite elements. Numerical results include experimental convergence rates to test the proposed implementation

    Scalable domain decomposition methods for finite element approximations of transient and electromagnetic problems

    Get PDF
    The main object of study of this thesis is the development of scalable and robust solvers based on domain decomposition (DD) methods for the linear systems arising from the finite element (FE) discretization of transient and electromagnetic problems. The thesis commences with a theoretical review of the curl-conforming edge (or Nédélec) FEs of the first kind and a comprehensive description of a general implementation strategy for h- and p- adaptive elements of arbitrary order on tetrahedral and hexahedral non-conforming meshes. Then, a novel balancing domain decomposition by constraints (BDDC) preconditioner that is robust for multi-material and/or heterogeneous problems posed in curl-conforming spaces is presented. The new method, in contrast to existent approaches, is based on the definition of the ingredients of the preconditioner according to the physical coefficients of the problem and does not require spectral information. The result is a robust and highly scalable preconditioner that preserves the simplicity of the original BDDC method. When dealing with transient problems, the time direction offers itself an opportunity for further parallelization. Aiming to design scalable space-time solvers, first, parallel-in-time parallel methods for linear and non-linear ordinary differential equations (ODEs) are proposed, based on (non-linear) Schur complement efficient solvers of a multilevel partition of the time interval. Then, these ideas are combined with DD concepts in order to design a two-level preconditioner as an extension to space-time of the BDDC method. The key ingredients for these new methods are defined such that they preserve the time causality, i.e., information only travels from the past to the future. The proposed schemes are weakly scalable in time and space-time, i.e., one can efficiently exploit increasing computational resources to solve more time steps in (approximately) the same time-to-solution. All the developments presented herein are motivated by the driving application of the thesis, the 3D simulation of the low-frequency electromagnetic response of High Temperature Superconductors (HTS). Throughout the document, an exhaustive set of numerical experiments, which includes the simulation of a realistic 3D HTS problem, is performed in order to validate the suitability and assess the parallel performance of the High Performance Computing (HPC) implementation of the proposed algorithms.L’objecte principal d’estudi d’aquesta tesi és el desenvolupament de solucionadors escalables i robustos basats en mètodes de descomposició de dominis (DD) per a sistemes lineals que sorgeixen en la discretització mitjançant elements finits (FE) de problemes transitoris i electromagnètics. La tesi comença amb una revisió teòrica dels FE d’eix (o de Nédélec) de la primera família i una descripció exhaustiva d’una estratègia d’implementació general per a elements h- i p-adaptatius d’ordre arbitrari en malles de tetraedres i hexaedres noconformes. Llavors, es presenta un nou precondicionador de descomposició de dominis balancejats per restricció (BDDC) que és robust per a problemes amb múltiples materials i/o heterogenis definits en espais curl-conformes. El nou mètode, en contrast amb els enfocaments existents, està basat en la definició dels ingredients del precondicionador segons els coeficients físics del problema i no requereix informació espectral. El resultat és un precondicionador robust i escalable que preserva la simplicitat del mètode original BDDC. Quan tractem amb problemes transitoris, la direcció temporal ofereix ella mateixa l’oportunitat de seguir explotant paral·lelisme. Amb l’objectiu de dissenyar precondicionadors en espai-temps, primer, proposem solucionadors paral·lels en temps per equacions diferencials lineals i no-lineals, basats en un solucionador eficient del complement de Schur d’una partició multinivell de l’interval de temps. Seguidament, aquestes idees es combinen amb conceptes de DD amb l’objectiu de dissenyar precondicionadors com a extensió a espai-temps dels mètodes de BDDC. Els ingredients clau d’aquests nous mètodes es defineixen de tal manera que preserven la causalitat del temps, on la informació només viatja de temps passats a temps futurs. Els esquemes proposats són dèbilment escalables en temps i en espai-temps, és a dir, es poden explotar eficientment recursos computacionals creixents per resoldre més passos de temps en (aproximadament) el mateix temps transcorregut de càlcul. Tots els desenvolupaments presentats aquí són motivats pel problema d’aplicació de la tesi, la simulació de la resposta electromagnètica de baixa freqüència dels superconductors d’alta temperatura (HTS) en 3D. Al llarg del document, es realitza un conjunt exhaustiu d’experiments numèrics, els quals inclouen la simulació d’un problema de HTS realista en 3D, per validar la idoneïtat i el rendiment paral·lel de la implementació per a computació d’alt rendiment dels algorismes proposatsPostprint (published version

    Mesh adaptation for high-order flow simulations

    Get PDF
    Mesh adaptation has only been considered for high-order flow simulations in recent years and many techniques are still to be made more robust and efficient with curvilinear meshes required by these high-order methods. This thesis covers the developments made to improve the mesh generation and adaptation capabilities of the open-source spectral/hp element framework Nektar++ and its dedicated mesh utility NekMesh. This thesis first covers the generation of quality initial meshes typically required before an iterative adaptation procedure can be used. For optimal performance of the spectral/hp element method, quadrilateral and hexahedral meshes are preferred and two methods are presented to achieve this, either entirely or partially. The first method, inspired from cross field methods, solves a Laplace problem to obtain a guiding field from which a valid two-dimensional quadrilateral block decomposition can be automatically obtained. In turn, naturally curved meshes are generated. The second method takes advantage of the medial axis to generate structured partitions in the boundary layer region of three-dimensional domains. The method proves to be robust in generating hybrid high-order meshes with boundary layer aligned prismatic elements near boundaries and tetrahedral elements elsewhere. The thesis goes on to explore the adaptation of high-order meshes for the simulation of flows using a spectral/hp element formulation. First a new approach to moving meshes, referred to here as r-adaptation, based on a variational framework, is described. This new r-adaptation module is then enhanced by p-adaptation for the simulation of compressible inviscid flows with shocks. Where the flow is smooth, p-adaptation is used to benefit from the spectral convergence of the spectral/hp element methods. Where the flow is discontinuous, e.g. at shock waves, r-adaptation clusters nodes together to better capture these field discontinuities. The benefits of this dual, rp-adaptation approach are demonstrated through two-dimensional benchmark test cases.Open Acces
    • …
    corecore