3,421 research outputs found
Mixing multi-core CPUs and GPUs for scientific simulation software
Recent technological and economic developments have led to widespread availability of
multi-core CPUs and specialist accelerator processors such as graphical processing units
(GPUs). The accelerated computational performance possible from these devices can be very
high for some applications paradigms. Software languages and systems such as NVIDIA's
CUDA and Khronos consortium's open compute language (OpenCL) support a number of
individual parallel application programming paradigms. To scale up the performance of some
complex systems simulations, a hybrid of multi-core CPUs for coarse-grained parallelism and
very many core GPUs for data parallelism is necessary. We describe our use of hybrid applica-
tions using threading approaches and multi-core CPUs to control independent GPU devices.
We present speed-up data and discuss multi-threading software issues for the applications
level programmer and o er some suggested areas for language development and integration
between coarse-grained and ne-grained multi-thread systems. We discuss results from three
common simulation algorithmic areas including: partial di erential equations; graph cluster
metric calculations and random number generation. We report on programming experiences
and selected performance for these algorithms on: single and multiple GPUs; multi-core CPUs;
a CellBE; and using OpenCL. We discuss programmer usability issues and the outlook and
trends in multi-core programming for scienti c applications developers
Highly accelerated simulations of glassy dynamics using GPUs: caveats on limited floating-point precision
Modern graphics processing units (GPUs) provide impressive computing
resources, which can be accessed conveniently through the CUDA programming
interface. We describe how GPUs can be used to considerably speed up molecular
dynamics (MD) simulations for system sizes ranging up to about 1 million
particles. Particular emphasis is put on the numerical long-time stability in
terms of energy and momentum conservation, and caveats on limited
floating-point precision are issued. Strict energy conservation over 10^8 MD
steps is obtained by double-single emulation of the floating-point arithmetic
in accuracy-critical parts of the algorithm. For the slow dynamics of a
supercooled binary Lennard-Jones mixture, we demonstrate that the use of
single-floating point precision may result in quantitatively and even
physically wrong results. For simulations of a Lennard-Jones fluid, the
described implementation shows speedup factors of up to 80 compared to a serial
implementation for the CPU, and a single GPU was found to compare with a
parallelised MD simulation using 64 distributed cores.Comment: 12 pages, 7 figures, to appear in Comp. Phys. Comm., HALMD package
licensed under the GPL, see http://research.colberg.org/projects/halm
Three Dimensional Pseudo-Spectral Compressible Magnetohydrodynamic GPU Code for Astrophysical Plasma Simulation
This paper presents the benchmarking and scaling studies of a GPU accelerated
three dimensional compressible magnetohydrodynamic code. The code is developed
keeping an eye to explain the large and intermediate scale magnetic field
generation is cosmos as well as in nuclear fusion reactors in the light of the
theory given by Eugene Newman Parker. The spatial derivatives of the code are
pseudo-spectral method based and the time solvers are explicit. GPU
acceleration is achieved with minimal code changes through OpenACC
parallelization and use of NVIDIA CUDA Fast Fourier Transform library (cuFFT).
NVIDIAs unified memory is leveraged to enable over-subscription of the GPU
device memory for seamless out-of-core processing of large grids. Our
experimental results indicate that the GPU accelerated code is able to achieve
upto two orders of magnitude speedup over a corresponding OpenMP parallel, FFTW
library based code, on a NVIDIA Tesla P100 GPU. For large grids that require
out-of-core processing on the GPU, we see a 7x speedup over the OpenMP, FFTW
based code, on the Tesla P100 GPU. We also present performance analysis of the
GPU accelerated code on different GPU architectures - Kepler, Pascal and Volta
- …