14 research outputs found
ASCR/HEP Exascale Requirements Review Report
This draft report summarizes and details the findings, results, and
recommendations derived from the ASCR/HEP Exascale Requirements Review meeting
held in June, 2015. The main conclusions are as follows. 1) Larger, more
capable computing and data facilities are needed to support HEP science goals
in all three frontiers: Energy, Intensity, and Cosmic. The expected scale of
the demand at the 2025 timescale is at least two orders of magnitude -- and in
some cases greater -- than that available currently. 2) The growth rate of data
produced by simulations is overwhelming the current ability, of both facilities
and researchers, to store and analyze it. Additional resources and new
techniques for data analysis are urgently needed. 3) Data rates and volumes
from HEP experimental facilities are also straining the ability to store and
analyze large and complex data volumes. Appropriately configured
leadership-class facilities can play a transformational role in enabling
scientific discovery from these datasets. 4) A close integration of HPC
simulation and data analysis will aid greatly in interpreting results from HEP
experiments. Such an integration will minimize data movement and facilitate
interdependent workflows. 5) Long-range planning between HEP and ASCR will be
required to meet HEP's research needs. To best use ASCR HPC resources the
experimental HEP program needs a) an established long-term plan for access to
ASCR computational and data resources, b) an ability to map workflows onto HPC
resources, c) the ability for ASCR facilities to accommodate workflows run by
collaborations that can have thousands of individual members, d) to transition
codes to the next-generation HPC platforms that will be available at ASCR
facilities, e) to build up and train a workforce capable of developing and
using simulations and analysis to support HEP scientific research on
next-generation systems.Comment: 77 pages, 13 Figures; draft report, subject to further revisio
A new CDF model for data movement based on SRM
Being a large international collaboration established well before the full development of the Grid as the main computing tool for High Energy Physics, CDF has recently changed and improved its computing model, decentralizing some parts of it in order to be able to exploit the rising number of distributed resources available nowadays. Despite those efforts, while the large majority of CDF Monte Carlo production has moved to the Grid, data processing is still mainly performed in dedicated farms hosted at FNAL, requiring a centralized management of data and Monte Carlo samples needed for physics analysis. This rises the question on how to manage the transfer of produced Monte Carlo samples from remote Grid sites to FNAL in an efficient way; up to now CDF has relied on a non scalable centralized solution based on dedicated data servers accessed through rcp protocol, which has proven to be unsatisfactory. A new data transfer model has been designed that uses SRMs as local caches for remote Monte Carlo production sites, interfaces them with SAM, the experiment data catalog, and finally realizes the file movement exploiting the features provided by the data catalog transfer layer. We describe here the model and its integration within the current CDF computing architecture
Belle II Technical Design Report
The Belle detector at the KEKB electron-positron collider has collected
almost 1 billion Y(4S) events in its decade of operation. Super-KEKB, an
upgrade of KEKB is under construction, to increase the luminosity by two orders
of magnitude during a three-year shutdown, with an ultimate goal of 8E35 /cm^2
/s luminosity. To exploit the increased luminosity, an upgrade of the Belle
detector has been proposed. A new international collaboration Belle-II, is
being formed. The Technical Design Report presents physics motivation, basic
methods of the accelerator upgrade, as well as key improvements of the
detector.Comment: Edited by: Z. Dole\v{z}al and S. Un
Virtualisation of Grid Resources and Prospects of the Measurement of Z Boson Production in Association with Jets at the LHC
At the Large Hadron Collider, a large number of events containing Z bosons will be available enabling the calibration of the absolute jet energy scale for the first time. In this thesis, such a calibration is deduced within the CMS experiment including the investigation of effects from the underlying event and the jet size parameter. In addition, virtualisation of operating systems is applied to increase the load, stability and maintainability of local grid computing infrastructures
An Approach to Grid Scheduling by Using Condor-G Matchmaking Mechanism
Grid is a distributed environment that integrates computing, storage and other resources in order to enable execution of applications that cannot be run on a single resource. Such environment requires advanced scheduling system in order to efficiently execute usersâ applications. In this paper, we give an overview of issues related to grid scheduling. We describe in details one of the most mature solutions â Condor-G Matchmaking mechanism. Furthermore, we propose our own approach to building grid scheduling system based on Condor-G Matchmaking
Data processing model for the CDF experiment
The data processing model for the CDF experiment is described. Data
processing reconstructs events from parallel data streams taken with different
combinations of physics event triggers and further splits the events into
datasets of specialized physics datasets. The design of the processing control
system faces strict requirements on bookkeeping records, which trace the status
of data files and event contents during processing and storage. The computing
architecture was updated to meet the mass data flow of the Run II data
collection, recently upgraded to a maximum rate of 40 MByte/sec. The data
processing facility consists of a large cluster of Linux computers with data
movement managed by the CDF data handling system to a multi-petaByte Enstore
tape library. The latest processing cycle has achieved a stable speed of 35
MByte/sec (3 TByte/day). It can be readily scaled by increasing CPU and
data-handling capacity as required.Comment: 12 pages, 10 figures, submitted to IEEE-TN
D0 Top Quark Results and their Dependence on Successful Grid Computing
The heaviest known Fermion particle -- the top quark -- was discovered at
Fermilab in the first run of the Tevatron in 1995. However, besides its mere
existence one needs to study its properties precisely in order to verify or
falsify the predictions of the Standard Model. With the top quark's extremely
high mass and short lifetime such measurements probe yet unexplored regions of
the theory and bring us closer to solving the open fundamental questions of our
universe of elementary particles such as why three families of quarks and
leptons exist and why their masses differ so dramatically.
To perform these measurements hundreds of millions of recorded
proton-antiproton collisions must be reconstructed and filtered to extract the
few top quarks produced. Simulated background and signal events with full
detector response need to be generated and reconstructed to validate and
understand the results. Since the start of the second run of the Tevatron the
D0 collaboration has brought Grid computing to its aid for the production of
simulated events. Data processing on the Grid has recently been added and
thereby enabled us to effectively triple the amount of data available with the
highest quality reconstruction methods.
We will present recent top quark results D0 obtained from these improved data
and explain how they benefited from the availability of computing resources on
the Grid.Comment: 10 pages, 8 figures, Invited talk at SciDAC2005, San Francisco, CA,
26-30 June 200