10 research outputs found

    Performance analysis and acceleration of nuclear physics application on high-performance computing platforms using GPGPUs and topology-aware mapping techniques

    Get PDF
    The number of nodes on current generation of high performance computing (HPC) platforms increases with a steady rate, and nodes of these computing platforms support multiple/many core hardware designs. As the number of cores per node increase, either CPU or accelerator based, we need to make use of all those cores. Thus, one has to use the accelerators as much as possible inside scientific applications. Furthermore, with the increase of the number of nodes, the communication time between nodes is likely to increase, which necessitates application specific network topology-aware mapping techniques for efficient utilization of these platforms. In addition, one also needs to construct network models in order to study the benefits of specific network mapping. The specific topology-aware mapping techniques will help to distribute the computational tasks so that the communication patterns make optimal use of the underlying network hardware. This research will mainly focus on the Many Fermion Dynamics nuclear (MFDn) application developed at Iowa State University, a computational tool for low-energy nuclear physics, which utilizes the so-called Lanczos algorithm (LA), an algorithm for diagonalization of sparse matrices that is widely used in the scientific parallel computing domain. We present techniques applied to this application which enhance its performance with the utilization of general purpose graphics processing units (GPGPUs). Additionally, we compare the performance of the sparse matrix vector multiplication (SpMVM), the main computationally intensive kernel in the LA, with other efficient approaches presented in the literature. We compare results for the total HPC platforms\u27 resources needed for different SpMVM implementations, present and analyze the implementation of communication and computation overlapping method, and extend a model for the analysis of network topology presented in the literature. Finally, we present network topology-aware mapping techniques, focused at the LA stage, for IBM Blue Gene/Q (BG/Q) supercomputers, which enhance the performance as compared to the default mapping, and validate the results of our test using the network model

    Ab Initio No Core Shell Model with Leadership-Class Supercomputers

    Full text link
    Nuclear structure and reaction theory is undergoing a major renaissance with advances in many-body methods, strong interactions with greatly improved links to Quantum Chromodynamics (QCD), the advent of high performance computing, and improved computational algorithms. Predictive power, with well-quantified uncertainty, is emerging from non-perturbative approaches along with the potential for guiding experiments to new discoveries. We present an overview of some of our recent developments and discuss challenges that lie ahead. Our foci include: (1) strong interactions derived from chiral effective field theory; (2) advances in solving the large sparse matrix eigenvalue problem on leadership-class supercomputers; (3) selected observables in light nuclei with the JISP16 interaction; (4) effective electroweak operators consistent with the Hamiltonian; and, (5) discussion of A=48 system as an opportunity for the no-core approach with the reintroduction of the core.Comment: 23 pages, 7 figures, Conference Proceedings online at http://ntse.khb.ru/files/uploads/2016/proceedings/Vary.pd

    Ab Initio No Core Shell Model - Recent Results and Further Prospects

    Full text link
    There has been significant recent progress in solving the long-standing problems of how nuclear shell structure and collective motion emerge from underlying microscopic inter-nucleon interactions. We review a selection of recent significant results within the ab initio No Core Shell Model (NCSM) closely tied to three major factors enabling this progress: (1) improved nuclear interactions that accurately describe the experimental two-nucleon and three-nucleon interaction data; (2) advances in algorithms to simulate the quantum many-body problem with strong interactions; and (3) continued rapid development of high-performance computers now capable of performing 20×101520 \times 10^{15} floating point operations per second. We also comment on prospects for further developments.Comment: Invited paper presented at NTSE-2014 and published online in the proceedings (see footnote on p.1

    Performance analysis and acceleration of nuclear physics application on high-performance computing platforms using GPGPUs and topology-aware mapping techniques

    No full text
    The number of nodes on current generation of high performance computing (HPC) platforms increases with a steady rate, and nodes of these computing platforms support multiple/many core hardware designs. As the number of cores per node increase, either CPU or accelerator based, we need to make use of all those cores. Thus, one has to use the accelerators as much as possible inside scientific applications. Furthermore, with the increase of the number of nodes, the communication time between nodes is likely to increase, which necessitates application specific network topology-aware mapping techniques for efficient utilization of these platforms. In addition, one also needs to construct network models in order to study the benefits of specific network mapping. The specific topology-aware mapping techniques will help to distribute the computational tasks so that the communication patterns make optimal use of the underlying network hardware. This research will mainly focus on the Many Fermion Dynamics nuclear (MFDn) application developed at Iowa State University, a computational tool for low-energy nuclear physics, which utilizes the so-called Lanczos algorithm (LA), an algorithm for diagonalization of sparse matrices that is widely used in the scientific parallel computing domain. We present techniques applied to this application which enhance its performance with the utilization of general purpose graphics processing units (GPGPUs). Additionally, we compare the performance of the sparse matrix vector multiplication (SpMVM), the main computationally intensive kernel in the LA, with other efficient approaches presented in the literature. We compare results for the total HPC platforms' resources needed for different SpMVM implementations, present and analyze the implementation of communication and computation overlapping method, and extend a model for the analysis of network topology presented in the literature. Finally, we present network topology-aware mapping techniques, focused at the LA stage, for IBM Blue Gene/Q (BG/Q) supercomputers, which enhance the performance as compared to the default mapping, and validate the results of our test using the network model.</p

    Ab Initio No Core Shell Model - Recent Resultsand Further Prospects

    No full text
    There has been significant recent progress in solving the long-standing problems of how nuclear shell structure and collective motion emerge from underlying microscopic inter-nucleon interactions. We review a selection of recent significant results within the ab initio No Core Shell Model (NCSM) closely tied to three major factors enabling this progress: (1) improved nuclear interactions that accurately describe the experimental two-nucleon and three-nucleon interaction data; (2) advances in algorithms to simulate the quantum manybody problem with strong interactions; and (3) continued rapid development of high-performance computers now capable of performing 20×1015 floating point operations per second. We also comment on prospects for further developments
    corecore