254 research outputs found
On the Way to Future's High Energy Particle Physics Transport Code
High Energy Physics (HEP) needs a huge amount of computing resources. In
addition data acquisition, transfer, and analysis require a well developed
infrastructure too. In order to prove new physics disciplines it is required to
higher the luminosity of the accelerator facilities, which produce
more-and-more data in the experimental detectors. Both testing new theories and
detector R&D are based on complex simulations. Today have already reach that
level, the Monte Carlo detector simulation takes much more time than real data
collection. This is why speed up of the calculations and simulations became
important in the HEP community. The Geant Vector Prototype (GeantV) project
aims to optimize the most-used particle transport code applying parallel
computing and to exploit the capabilities of the modern CPU and GPU
architectures as well. With the maximized concurrency at multiple levels the
GeantV is intended to be the successor of the Geant4 particle transport code
that has been used since two decades successfully. Here we present our latest
result on the GeantV tests performances, comparing CPU/GPU based vectorized
GeantV geometrical code to the Geant4 version
Probabilistic structural mechanics research for parallel processing computers
Aerospace structures and spacecraft are a complex assemblage of structural components that are subjected to a variety of complex, cyclic, and transient loading conditions. Significant modeling uncertainties are present in these structures, in addition to the inherent randomness of material properties and loads. To properly account for these uncertainties in evaluating and assessing the reliability of these components and structures, probabilistic structural mechanics (PSM) procedures must be used. Much research has focused on basic theory development and the development of approximate analytic solution methods in random vibrations and structural reliability. Practical application of PSM methods was hampered by their computationally intense nature. Solution of PSM problems requires repeated analyses of structures that are often large, and exhibit nonlinear and/or dynamic response behavior. These methods are all inherently parallel and ideally suited to implementation on parallel processing computers. New hardware architectures and innovative control software and solution methodologies are needed to make solution of large scale PSM problems practical
Status of Vectorized Monte Carlo for Particle Transport Analysis
The conventional particle transport Monte Carlo algorithm is ill suited for modem vector supercomputers because the random nature of the particle transport process in the history based algorithm in hibits construction of vectors. An alterna tive, event-based algorithm is suitable for vectorization and has been used recently to achieve impressive gains in perfor mance on vector supercomputers. This re view describes the event-based algorithm and several variations of it Implementa tions of this algorithm for applications in particle transport are described, and their relative merits are discussed. The imple mentation of Monte Carlo methods on multiple vector parallel processors is con sidered, as is the potential of massively parallel processors for Monte Carlo par ticle transport simulations.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/67177/2/10.1177_109434208700100203.pd
A bibliography on parallel and vector numerical algorithms
This is a bibliography of numerical methods. It also includes a number of other references on machine architecture, programming language, and other topics of interest to scientific computing. Certain conference proceedings and anthologies which have been published in book form are listed also
HSMA: An O(N) electrostatics package implemented in LAMMPS
We implement two recently developed fast Coulomb solvers, HSMA3D [J. Chem.
Phys. 149 (8) (2018) 084111] and HSMA2D [J. Chem. Phys. 152 (13) (2020)
134109], into a new user package HSMA for molecular dynamics simulation engine
LAMMPS. The HSMA package is designed for efficient and accurate modeling of
electrostatic interactions in 3D and 2D periodic systems with dielectric
effects at the O(N) cost. The implementation is hybrid MPI and OpenMP
parallelized and compatible with existing LAMMPS functionalities. The
vectorization technique following AVX512 instructions is adopted for
acceleration. To establish the validity of our implementation, we have
presented extensive comparisons to the widely used particle-particle
particle-mesh (PPPM) algorithm in LAMMPS and other dielectric solvers. With the
proper choice of algorithm parameters and parallelization setup, the package
enables calculations of electrostatic interactions that outperform the standard
PPPM in speed for a wide range of particle numbers
Exploring Computational Chemistry on Emerging Architectures
Emerging architectures, such as next generation microprocessors, graphics processing units, and Intel MIC cards, are being used with increased popularity in high performance computing. Each of these architectures has advantages over previous generations of architectures including performance, programmability, and power efficiency. With the ever-increasing performance of these architectures, scientific computing applications are able to attack larger, more complicated problems. However, since applications perform differently on each of the architectures, it is difficult to determine the best tool for the job. This dissertation makes the following contributions to computer engineering and computational science. First, this work implements the computational chemistry variational path integral application, QSATS, on various architectures, ranging from microprocessors to GPUs to Intel MICs. Second, this work explores the use of analytical performance modeling to predict the runtime and scalability of the application on the architectures. This allows for a comparison of the architectures when determining which to use for a set of program input parameters. The models presented in this dissertation are accurate within 6%. This work combines novel approaches to this algorithm and exploration of the various architectural features to develop the application to perform at its peak. In addition, this expands the understanding of computational science applications and their implementation on emerging architectures while providing insight into the performance, scalability, and programmer productivity
Monte Carlo Photon Transport On Shared Memory and Distributed Memory Parallel Processors
Parallelized Monte Carlo algorithms for analyzing photon transport in an inertially confined fusion (ICF) plasma are consid ered. Algorithms were developed for shared memory (vector and scalar) and distributed memory (scalar) parallel pro cessors. The shared memory algorithm was implemented on the IBM 3090/400, and timing results are presented for dedi cated runs with two, three, and four pro cessors. Two alternative distributed memory algorithms (replication and dis patching) were implemented on a hyper cube parallel processor (1 through 64 nodes). The replication algorithm yields essentially full efficiency for all cube sizes; with the 64-node configuration, the absolute performance is nearly the same as with the CRAY X-MP The dispatching algorithm also yields efficiencies above 80% in a large simulation for the 64-pro cessor configuration.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/67146/2/10.1177_109434208700100306.pd
Path-tracing Monte Carlo Library for 3D Radiative Transfer in Highly Resolved Cloudy Atmospheres
Interactions between clouds and radiation are at the root of many
difficulties in numerically predicting future weather and climate and in
retrieving the state of the atmosphere from remote sensing observations. The
large range of issues related to these interactions, and in particular to
three-dimensional interactions, motivated the development of accurate radiative
tools able to compute all types of radiative metrics, from monochromatic, local
and directional observables, to integrated energetic quantities. In the
continuity of this community effort, we propose here an open-source library for
general use in Monte Carlo algorithms. This library is devoted to the
acceleration of path-tracing in complex data, typically high-resolution
large-domain grounds and clouds. The main algorithmic advances embedded in the
library are those related to the construction and traversal of hierarchical
grids accelerating the tracing of paths through heterogeneous fields in
null-collision (maximum cross-section) algorithms. We show that with these
hierarchical grids, the computing time is only weakly sensitivive to the
refinement of the volumetric data. The library is tested with a rendering
algorithm that produces synthetic images of cloud radiances. Two other examples
are given as illustrations, that are respectively used to analyse the
transmission of solar radiation under a cloud together with its sensitivity to
an optical parameter, and to assess a parametrization of 3D radiative effects
of clouds.Comment: Submitted to JAMES, revised and submitted again (this is v2
ASCR/HEP Exascale Requirements Review Report
This draft report summarizes and details the findings, results, and
recommendations derived from the ASCR/HEP Exascale Requirements Review meeting
held in June, 2015. The main conclusions are as follows. 1) Larger, more
capable computing and data facilities are needed to support HEP science goals
in all three frontiers: Energy, Intensity, and Cosmic. The expected scale of
the demand at the 2025 timescale is at least two orders of magnitude -- and in
some cases greater -- than that available currently. 2) The growth rate of data
produced by simulations is overwhelming the current ability, of both facilities
and researchers, to store and analyze it. Additional resources and new
techniques for data analysis are urgently needed. 3) Data rates and volumes
from HEP experimental facilities are also straining the ability to store and
analyze large and complex data volumes. Appropriately configured
leadership-class facilities can play a transformational role in enabling
scientific discovery from these datasets. 4) A close integration of HPC
simulation and data analysis will aid greatly in interpreting results from HEP
experiments. Such an integration will minimize data movement and facilitate
interdependent workflows. 5) Long-range planning between HEP and ASCR will be
required to meet HEP's research needs. To best use ASCR HPC resources the
experimental HEP program needs a) an established long-term plan for access to
ASCR computational and data resources, b) an ability to map workflows onto HPC
resources, c) the ability for ASCR facilities to accommodate workflows run by
collaborations that can have thousands of individual members, d) to transition
codes to the next-generation HPC platforms that will be available at ASCR
facilities, e) to build up and train a workforce capable of developing and
using simulations and analysis to support HEP scientific research on
next-generation systems.Comment: 77 pages, 13 Figures; draft report, subject to further revisio
Computational Methods in Science and Engineering : Proceedings of the Workshop SimLabs@KIT, November 29 - 30, 2010, Karlsruhe, Germany
In this proceedings volume we provide a compilation of article contributions equally covering applications from different research fields and ranging from capacity up to capability computing. Besides classical computing aspects such as parallelization, the focus of these proceedings is on multi-scale approaches and methods for tackling algorithm and data complexity. Also practical aspects regarding the usage of the HPC infrastructure and available tools and software at the SCC are presented
- …