5,453 research outputs found
A Parallel Tree-SPH code for Galaxy Formation
We describe a new implementation of a parallel Tree-SPH code with the aim to
simulate Galaxy Formation and Evolution. The code has been parallelized using
SHMEM, a Cray proprietary library to handle communications between the 256
processors of the Silicon Graphics T3E massively parallel supercomputer hosted
by the Cineca Supercomputing Center (Bologna, Italy). The code combines the
Smoothed Particle Hydrodynamics (SPH) method to solve hydro-dynamical equations
with the popular Barnes and Hut (1986) tree-code to perform gravity calculation
with a NlogN scaling, and it is based on the scalar Tree-SPH code developed by
Carraro et al(1998)[MNRAS 297, 1021]. Parallelization is achieved distributing
particles along processors according to a work-load criterion. Benchmarks, in
terms of load-balance and scalability, of the code are analyzed and critically
discussed against the adiabatic collapse of an isothermal gas sphere test using
20,000 particles on 8 processors. The code results balanced at more that 95%
level. Increasing the number of processors, the load-balance slightly worsens.
The deviation from perfect scalability at increasing number of processors is
almost negligible up to 32 processors. Finally we present a simulation of the
formation of an X-ray galaxy cluster in a flat cold dark matter cosmology,
using 200,000 particles and 32 processors, and compare our results with Evrard
(1988) P3M-SPH simulations. Additionaly we have incorporated radiative cooling,
star formation, feed-back from SNae of type II and Ia, stellar winds and UV
flux from massive stars, and an algorithm to follow the chemical enrichment of
the inter-stellar medium. Simulations with some of these ingredients are also
presented.Comment: 19 pages, 14 figures, accepted for publication in MNRA
AMR on the CM-2
We describe the development of a structured adaptive mesh algorithm (AMR) for the Connection Machine-2 (CM-2). We develop a data layout scheme that preserves locality even for communication between fine and coarse grids. On 8K of a 32K machine we achieve performance slightly less than 1 CPU of the Cray Y-MP. We apply our algorithm to an inviscid compressible flow problem
Large-Scale Forcing with Less Communication in Finite-Difference Simulations of Stationary Isotropic Turbulence
「気候変動に適応可能な環境探索のためのマルチスケールシミュレーション」プロジェク
Load management strategy for Particle-In-Cell simulations in high energy particle acceleration
In the wake of the intense effort made for the experimental CILEX project,
numerical simulation cam- paigns have been carried out in order to finalize the
design of the facility and to identify optimal laser and plasma parameters.
These simulations bring, of course, important insight into the fundamental
physics at play. As a by-product, they also characterize the quality of our
theoretical and numerical models. In this paper, we compare the results given
by different codes and point out algorithmic lim- itations both in terms of
physical accuracy and computational performances. These limitations are illu-
strated in the context of electron laser wakefield acceleration (LWFA). The
main limitation we identify in state-of-the-art Particle-In-Cell (PIC) codes is
computational load imbalance. We propose an innovative algorithm to deal with
this specific issue as well as milestones towards a modern, accurate high-per-
formance PIC code for high energy particle acceleration
- …