5,083 research outputs found
Plasma simulation using the massively parallel processor
Two dimensional electrostatic simulation codes using the particle-in-cell model are developed on the Massively Parallel Processor (MPP). The conventional plasma simulation procedure that computes electric fields at particle positions by means of a gridded system is found inefficient on the MPP. The MPP simulation code is thus based on the gridless system in which particles are assigned to processing elements and electric fields are computed directly via Discrete Fourier Transform. Currently, the gridless model on the MPP in two dimensions is about nine times slower that the gridded system on the CRAY X-MP without considering I/O time. However, the gridless system on the MPP can be improved by incorporating a faster I/O between the staging memory and Array Unit and a more efficient procedure for taking floating point sums over processing elements. The initial results suggest that the parallel processors have the potential for performing large scale plasma simulations
Message-Passing Multi-Cell Molecular Dynamics on the Connection Machine 5
We present a new scalable algorithm for short-range molecular dynamics
simulations on distributed memory MIMD multicomputer based on a message-passing
multi-cell approach. We have implemented the algorithm on the Connection
Machine 5 (CM-5) and demonstrate that meso-scale molecular dynamics with more
than particles is now possible on massively parallel MIMD computers.
Typical runs show single particle update-times of in 2 dimensions
(2D) and approximately in 3 dimensions (3D) on a 1024 node CM-5
without vector units, corresponding to more than 1.8 GFlops overall
performance. We also present a scaling equation which agrees well with actually
observed timings.Comment: 17 pages, Uuencoded compressed PostScript fil
A Parallel Tree-SPH code for Galaxy Formation
We describe a new implementation of a parallel Tree-SPH code with the aim to
simulate Galaxy Formation and Evolution. The code has been parallelized using
SHMEM, a Cray proprietary library to handle communications between the 256
processors of the Silicon Graphics T3E massively parallel supercomputer hosted
by the Cineca Supercomputing Center (Bologna, Italy). The code combines the
Smoothed Particle Hydrodynamics (SPH) method to solve hydro-dynamical equations
with the popular Barnes and Hut (1986) tree-code to perform gravity calculation
with a NlogN scaling, and it is based on the scalar Tree-SPH code developed by
Carraro et al(1998)[MNRAS 297, 1021]. Parallelization is achieved distributing
particles along processors according to a work-load criterion. Benchmarks, in
terms of load-balance and scalability, of the code are analyzed and critically
discussed against the adiabatic collapse of an isothermal gas sphere test using
20,000 particles on 8 processors. The code results balanced at more that 95%
level. Increasing the number of processors, the load-balance slightly worsens.
The deviation from perfect scalability at increasing number of processors is
almost negligible up to 32 processors. Finally we present a simulation of the
formation of an X-ray galaxy cluster in a flat cold dark matter cosmology,
using 200,000 particles and 32 processors, and compare our results with Evrard
(1988) P3M-SPH simulations. Additionaly we have incorporated radiative cooling,
star formation, feed-back from SNae of type II and Ia, stellar winds and UV
flux from massive stars, and an algorithm to follow the chemical enrichment of
the inter-stellar medium. Simulations with some of these ingredients are also
presented.Comment: 19 pages, 14 figures, accepted for publication in MNRA
A Parallel Adaptive P3M code with Hierarchical Particle Reordering
We discuss the design and implementation of HYDRA_OMP a parallel
implementation of the Smoothed Particle Hydrodynamics-Adaptive P3M (SPH-AP3M)
code HYDRA. The code is designed primarily for conducting cosmological
hydrodynamic simulations and is written in Fortran77+OpenMP. A number of
optimizations for RISC processors and SMP-NUMA architectures have been
implemented, the most important optimization being hierarchical reordering of
particles within chaining cells, which greatly improves data locality thereby
removing the cache misses typically associated with linked lists. Parallel
scaling is good, with a minimum parallel scaling of 73% achieved on 32 nodes
for a variety of modern SMP architectures. We give performance data in terms of
the number of particle updates per second, which is a more useful performance
metric than raw MFlops. A basic version of the code will be made available to
the community in the near future.Comment: 34 pages, 12 figures, accepted for publication in Computer Physics
Communication
GADGET: A code for collisionless and gasdynamical cosmological simulations
We describe the newly written code GADGET which is suitable both for
cosmological simulations of structure formation and for the simulation of
interacting galaxies. GADGET evolves self-gravitating collisionless fluids with
the traditional N-body approach, and a collisional gas by smoothed particle
hydrodynamics. Along with the serial version of the code, we discuss a parallel
version that has been designed to run on massively parallel supercomputers with
distributed memory. While both versions use a tree algorithm to compute
gravitational forces, the serial version of GADGET can optionally employ the
special-purpose hardware GRAPE instead of the tree. Periodic boundary
conditions are supported by means of an Ewald summation technique. The code
uses individual and adaptive timesteps for all particles, and it combines this
with a scheme for dynamic tree updates. Due to its Lagrangian nature, GADGET
thus allows a very large dynamic range to be bridged, both in space and time.
So far, GADGET has been successfully used to run simulations with up to 7.5e7
particles, including cosmological studies of large-scale structure formation,
high-resolution simulations of the formation of clusters of galaxies, as well
as workstation-sized problems of interacting galaxies. In this study, we detail
the numerical algorithms employed, and show various tests of the code. We
publically release both the serial and the massively parallel version of the
code.Comment: 32 pages, 14 figures, replaced to match published version in New
Astronomy. For download of the code, see
http://www.mpa-garching.mpg.de/gadget (new version 1.1 available
- …