90 research outputs found

    Lattice QCD Production on Commodity Clusters at Fermilab

    Full text link
    We describe the construction and results to date of Fermilab's three Myrinet-networked lattice QCD production clusters (an 80-node dual Pentium III cluster, a 48-node dual Xeon cluster, and a 128-node dual Xeon cluster). We examine a number of aspects of performance of the MILC lattice QCD code running on these clusters.Comment: Talk from the 2003 Computing in High Energy and Nuclear Physics (CHEP03), La Jolla, Ca, USA, March 2003, 6 pages, LaTeX, 8 eps figures. PSN TUIT00

    ASCR/HEP Exascale Requirements Review Report

    Full text link
    This draft report summarizes and details the findings, results, and recommendations derived from the ASCR/HEP Exascale Requirements Review meeting held in June, 2015. The main conclusions are as follows. 1) Larger, more capable computing and data facilities are needed to support HEP science goals in all three frontiers: Energy, Intensity, and Cosmic. The expected scale of the demand at the 2025 timescale is at least two orders of magnitude -- and in some cases greater -- than that available currently. 2) The growth rate of data produced by simulations is overwhelming the current ability, of both facilities and researchers, to store and analyze it. Additional resources and new techniques for data analysis are urgently needed. 3) Data rates and volumes from HEP experimental facilities are also straining the ability to store and analyze large and complex data volumes. Appropriately configured leadership-class facilities can play a transformational role in enabling scientific discovery from these datasets. 4) A close integration of HPC simulation and data analysis will aid greatly in interpreting results from HEP experiments. Such an integration will minimize data movement and facilitate interdependent workflows. 5) Long-range planning between HEP and ASCR will be required to meet HEP's research needs. To best use ASCR HPC resources the experimental HEP program needs a) an established long-term plan for access to ASCR computational and data resources, b) an ability to map workflows onto HPC resources, c) the ability for ASCR facilities to accommodate workflows run by collaborations that can have thousands of individual members, d) to transition codes to the next-generation HPC platforms that will be available at ASCR facilities, e) to build up and train a workforce capable of developing and using simulations and analysis to support HEP scientific research on next-generation systems.Comment: 77 pages, 13 Figures; draft report, subject to further revisio

    National Computational Infrastructure for Lattice Gauge Theory

    Full text link

    Summary: Working Group on QCD and Strong Interactions

    Full text link
    In this summary of the considerations of the QCD working group at Snowmass 2001, the roles of quantum chromodynamics in the Standard Model and in the search for new physics are reviewed, with empahsis on frontier areas in the field. We discuss the importance of, and prospects for, precision QCD in perturbative and lattice calculations. We describe new ideas in the analysis of parton distribution functions and jet structure, and review progress in small-xx and in polarization.Comment: Snowmass 2001. Revtex4, 34 pages, 4 figures, revised to include additional references on jets and lattice QC

    An Algorithmic Approach to Quantum Field Theory

    Full text link
    The lattice formulation provides a way to regularize, define and compute the Path Integral in a Quantum Field Theory. In this paper we review the theoretical foundations and the most basic algorithms required to implement a typical lattice computation, including the Metropolis, the Gibbs sampling, the Minimal Residual, and the Stabilized Biconjugate inverters. The main emphasis is on gauge theories with fermions such as QCD. We also provide examples of typical results from lattice QCD computations for quantities of phenomenological interest.Comment: 44 pages, to be published in IJMP
    corecore