911 research outputs found

    Off-line computing for experimental high-energy physics

    Get PDF
    The needs of experimental high-energy physics for large-scale computing and data handling are explained in terms of the complexity of individual collisions and the need for high statistics to study quantum mechanical processes. The prevalence of university-dominated collaborations adds a requirement for high-performance wide-area networks. The data handling and computational needs of the different types of large experiment, now running or under construction, are evaluated. Software for experimental high-energy physics is reviewed briefly with particular attention to the success of packages written within the discipline. It is argued that workstations and graphics are important in ensuring that analysis codes are correct, and the worldwide networks which support the involvement of remote physicists are described. Computing and data handling are reviewed showing how workstations and RISC processors are rising in importance but have not supplanted traditional mainframe processing. Examples of computing systems constructed within high-energy physics are examined and evaluated

    The 1998 Center for Simulation of Dynamic Response in Materials Annual Technical Report

    Get PDF
    Introduction: This annual report describes research accomplishments for FY 98 of the Center for Simulation of Dynamic Response of Materials. The Center is constructing a virtual shock physics facility in which the full three dimensional response of a variety of target materials can be computed for a wide range of compressive, tensional, and shear loadings, including those produced by detonation of energetic materials. The goals are to facilitate computation of a variety of experiments in which strong shock and detonation waves are made to impinge on targets consisting of various combinations of materials, compute the subsequent dynamic response of the target materials, and validate these computations against experimental data

    Semi‐automatic porting of a large‐scale CFD code to multi‐graphics processing unit clusters

    Get PDF
    A typical large‐scale CFD code based on adaptive, edge‐based finite‐element formulations for the solution of compressible and incompressible flow is taken as a test bed to port such codes to graphics hardware (graphics processing units, GPUs) using semi‐automatic techniques. In previous work, a GPU version of this code was presented, in which, for many run configurations, all mesh‐sized loops required throughout time stepping were ported. This approach simultaneously achieves the fine‐grained parallelism required to fully exploit the capabilities of many‐core GPUs, completely avoids the crippling bottleneck of GPU–CPU data transfer, and uses a transposed memory layout to meet the distinct memory access requirements posed by GPUs. The present work describes the next step of this porting effort, namely to integrate GPU‐based, fine‐grained parallelism with Message‐Passing‐Interface‐based, coarse‐grained parallelism, in order to achieve a code capable of running on multi‐GPU clusters. This is carried out in a semi‐automated fashion: the existing Fortran–Message Passing Interface code is preserved, with the translator inserting data transfer calls as required. Performance benchmarks indicate up to a factor of 2 performance advantage of the NVIDIA Tesla M2050 GPU (Santa Clara, CA, USA) over the six‐core Intel Xeon X5670 CPU (Santa Clara, CA, USA), for certain run configurations. In addition, good scalability is observed when running across multiple GPUs. The approach should be of general interest, as how best to run on GPUs is being presently considered for many so‐called legacy codes

    Semi‐automatic porting of a large‐scale Fortran CFD code to GPUs

    Get PDF
    The development of automatic techniques to port a substantial portion of FEFLO, a general‐purpose legacy CFD code operating on unstructured grids, to run on GPUs is described. FEFLO is a typical adaptive, edge‐based finite element code for the solution of compressible and incompressible flows, which is primarily written in Fortran 77 and has previously been ported to vector, shared memory parallel and distributed memory parallel machines. Owing to the large size of FEFLO and the likelihood of human error in porting, as well as the desire for continued development within a single codebase, a specialized Python script, based on FParser (Int. J. Comput. Sci. Eng. 2009; 4 :296–305), was written to perform automated translation from the OpenMP‐parallelized edge and point loops to GPU kernels implemented in CUDA, along with GPU memory management. The results of verification benchmarks and performance indicate that performances achieved by such a translator can rival those of codes rewritten by specialists. The approach should be of general interest, as how best to run on GPUs is being presently considered for many so‐called legacy codes

    HPCCP/CAS Workshop Proceedings 1998

    Get PDF
    This publication is a collection of extended abstracts of presentations given at the HPCCP/CAS (High Performance Computing and Communications Program/Computational Aerosciences Project) Workshop held on August 24-26, 1998, at NASA Ames Research Center, Moffett Field, California. The objective of the Workshop was to bring together the aerospace high performance computing community, consisting of airframe and propulsion companies, independent software vendors, university researchers, and government scientists and engineers. The Workshop was sponsored by the HPCCP Office at NASA Ames Research Center. The Workshop consisted of over 40 presentations, including an overview of NASA's High Performance Computing and Communications Program and the Computational Aerosciences Project; ten sessions of papers representative of the high performance computing research conducted within the Program by the aerospace industry, academia, NASA, and other government laboratories; two panel sessions; and a special presentation by Mr. James Bailey

    Shared-Memory parallelization of consistent particle method for violent wave impact problems

    Get PDF
    A shared-memory parallelization is implemented to the recently developed Consistent Particle Method (CPM) for violent wave impact problems. The advantages of this relatively new particle method lie in four key aspects: (1) accurate computation of Laplacian and gradient operators based on Taylor series expansion, alleviating spurious pressure fluctuation and being able to model two-phase flows characterized by large density difference, (2) a thermodynamics-based compressible solver for modelling compressible air that eliminates the need of determining artificial sound speed, (3) seamless coupling of the compressible air solver and incompressible water solver, and (4) parallelization of the numerical model based on Open Multi-Processing (OpenMP) and a parallel direct sparse solver (Pardiso) to significantly improve computational efficiency. Strong and weak scaling analyses of the parallelized CPM are conducted, showing an efficiency speedup of 100 times or more depending on the size of simulated problem. To demonstrate the accuracy of the developed numerical model, three numerical examples are studied including the benchmark study of wave impact on seawall, and our experimental studies of violent water sloshing under rotational excitations and sloshing impact with entrapped air pocket. CPM is shown to accurately capture highly deformed breaking waves and violent wave impact pressure including pressure oscillation induced by air cushion effect
    corecore