7 research outputs found

    Equipping Sparse Solvers for Exascale (ESSEX / ESSEX II)

    Get PDF
    The ESSEX project is funded by the German DFG priority programme 1648 "Software for Exascale Computing" (SPPEXA). In 2016 it has entered its second funding phase, ESSEX-II. ESSEX investigated programming concepts and numerical algorithms for scalable, efficient and robust iterative sparse matrix applications on exascale systems. Starting with successful blueprints and prototype solutions identified in ESSEX-I, the second phase project ESSEX-II developed a collection of broadly usable and scalable sparse eigenvalue solvers with high hardware efficiency for the computer architectures to come. Project activities were organized along the traditional software layers of low-level parallel building blocks (kernels), algorithm implementations, and applications. The classic abstraction boundaries separating these layers were broken in ESSEX by strongly integrating objectives: scalability, numerical reliability, fault tolerance, and holistic performance and power engineering. The basic building block library supports an elaborate MPI+X approach that is able to fully exploit hardware heterogeneity while exposing functional parallelism and data parallelism to all other software layers in a flexible way. In addition, facilities for fully asynchronous checkpointing, silent data corruption detection and correction, performance assessment, performance model validation, and energy measurements are provided transparently. The advanced building blocks were defined and employed by the developments at the algorithms layer. Here, ESSEX-II provides state-of-the-art library implementations of classic linear sparse eigenvalue solvers including block Jacobi-Davidson, Kernel Polynomial Method (KPM), and Chebyshev filter diagonalization (ChebFD) that are ready to use for production on modern heterogeneous compute nodes with best performance and numerical accuracy. Research in this direction included the development of appropriate parallel adaptive AMG software for the block Jacobi-Davidson method. Contour integral-based approaches were also covered in ESSEX-II and were extended in two directions: The FEAST method was further developed for improved scalability, and the Sakurai-Sugiura method (SSM) method was extended to nonlinear sparse eigenvalue problems. These developments were strongly supported by additional Japanese project partners from University of Tokyo, Computer Science, and University of Tsukuba, Applied Mathematics. The applications layer delivers scalable solutions for conservative (Hermitian) and dissipative (non- Hermitian) quantum systems with strong links to optics and biology and to novel materials such as graphene and topological insulators. This talk gives a survey on latest results of the ESSEX-II project

    Equipping Sparse Solvers for Exascale - A Survey of the DFG Project ESSEX

    Get PDF
    The ESSEX project investigates computational issues arising at exascale for large-scale sparse eigenvalue problems and develops programming concepts and numerical methods for their solution. The project pursues a coherent co-design of all software layers where a holistic performance engineering process guides code development across the classic boundaries of application, numerical method and basic kernel library. Within ESSEX the numerical methods cover both widely applicable solvers such as classic Krylov, Jacobi-Davidson or recent FEAST methods as well as domain specific iterative schemes relevant for the ESSEX quantum physics application. This presentation introduces the project structure and presents selected results which demonstrate the potential impact of ESSEX for efficient sparse solvers on highly scalable heterogeneous supercomputers. In the second project phase from 2016 to 2018, the ESSEX consortium will include partners from the Universities of Tokyo and of Tsukuba. Extensions of existing work will regard numerically reliable computing methods, scalability improvements by leveraging functional parallelism in asynchronous preconditioners, hiding and reducing communication cost, improving load balancing by advanced partitioning schemes, as well as the treatment of non-Hermitian matrix problems

    GHOST: Building blocks for high performance sparse linear algebra on heterogeneous systems

    Get PDF
    While many of the architectural details of future exascale-class high performance computer systems are still a matter of intense research, there appears to be a general consensus that they will be strongly heterogeneous, featuring "standard" as well as "accelerated" resources. Today, such resources are available as multicore processors, graphics processing units (GPUs), and other accelerators such as the Intel Xeon Phi. Any software infrastructure that claims usefulness for such environments must be able to meet their inherent challenges: massive multi-level parallelism, topology, asynchronicity, and abstraction. The "General, Hybrid, and Optimized Sparse Toolkit" (GHOST) is a collection of building blocks that targets algorithms dealing with sparse matrix representations on current and future large-scale systems. It implements the "MPI+X" paradigm, has a pure C interface, and provides hybrid-parallel numerical kernels, intelligent resource management, and truly heterogeneous parallelism for multicore CPUs, Nvidia GPUs, and the Intel Xeon Phi. We describe the details of its design with respect to the challenges posed by modern heterogeneous supercomputers and recent algorithmic developments. Implementation details which are indispensable for achieving high efficiency are pointed out and their necessity is justified by performance measurements or predictions based on performance models. The library code and several applications are available as open source. We also provide instructions on how to make use of GHOST in existing software packages, together with a case study which demonstrates the applicability and performance of GHOST as a component within a larger software stack.Comment: 32 pages, 11 figure

    Benefits from using mixed precision computations in the ELPA-AEO and ESSEX-II eigensolver projects

    Get PDF
    We first briefly report on the status and recent achievements of the ELPA-AEO (Eigenvalue Solvers for Petaflop Applications - Algorithmic Extensions and Optimizations) and ESSEX II (Equipping Sparse Solvers for Exascale) projects. In both collaboratory efforts, scientists from the application areas, mathematicians, and computer scientists work together to develop and make available efficient highly parallel methods for the solution of eigenvalue problems. Then we focus on a topic addressed in both projects, the use of mixed precision computations to enhance efficiency. We give a more detailed description of our approaches for benefiting from either lower or higher precision in three selected contexts and of the results thus obtained

    Increasing the performance of the Jacobi-Davidson method by blocking

    Get PDF
    Block variants of the Jacobi-Davidson method for computing a few eigenpairs of a large sparse matrix are known to improve the robustness of the standard algorithm, but are generally shunned because the total number of floating-point operations increases. In this paper we present the implementation of a block Jacobi-Davidson solver. By detailed performance engineering and numerical experiments we demonstrate that the increase in operations is typically more than compensated by performance gains on modern architectures, giving a method that is both more efficient and robust than its single vector counterpart
    corecore