825 research outputs found
CMS software deployment on OSG
A set of software deployment tools has been developed for the installation,
verification, and removal of a CMS software release. The tools that are mainly targeted for the
deployment on the OSG have the features of instant release deployment, corrective
resubmission of the initial installation job, and an independent web-based deployment portal
with Grid security infrastructure login mechanism. We have been deploying over 500
installations and found the tools are reliable and adaptable to cope with problems with changes
in the Grid computing environment and the software releases. We present the design of the
tools, statistics that we gathered during the operation of the tools, and our experience with the
CMS software deployment on the OSG Grid computing environment
Designing Computing System Architecture and Models for the HL-LHC era
This paper describes a programme to study the computing model in CMS after
the next long shutdown near the end of the decade.Comment: Submitted to proceedings of the 21st International Conference on
Computing in High Energy and Nuclear Physics (CHEP2015), Okinawa, Japa
Characterizing network paths in and out of the clouds
Commercial Cloud computing is becoming mainstream, with funding agencies
moving beyond prototyping and starting to fund production campaigns, too. An
important aspect of any scientific computing production campaign is data
movement, both incoming and outgoing. And while the performance and cost of VMs
is relatively well understood, the network performance and cost is not. This
paper provides a characterization of networking in various regions of Amazon
Web Services, Microsoft Azure and Google Cloud Platform, both between Cloud
resources and major DTNs in the Pacific Research Platform, including OSG data
federation caches in the network backbone, and inside the clouds themselves.
The paper contains both a qualitative analysis of the results as well as
latency and throughput measurements. It also includes an analysis of the costs
involved with Cloud-based networking.Comment: 7 pages, 1 figure, 5 tables, to be published in CHEP19 proceeding
Running a Pre-Exascale, Geographically Distributed, Multi-Cloud Scientific Simulation
As we approach the Exascale era, it is important to verify that the existing
frameworks and tools will still work at that scale. Moreover, public Cloud
computing has been emerging as a viable solution for both prototyping and
urgent computing. Using the elasticity of the Cloud, we have thus put in place
a pre-exascale HTCondor setup for running a scientific simulation in the Cloud,
with the chosen application being IceCube's photon propagation simulation. I.e.
this was not a purely demonstration run, but it was also used to produce
valuable and much needed scientific results for the IceCube collaboration. In
order to reach the desired scale, we aggregated GPU resources across 8 GPU
models from many geographic regions across Amazon Web Services, Microsoft
Azure, and the Google Cloud Platform. Using this setup, we reached a peak of
over 51k GPUs corresponding to almost 380 PFLOP32s, for a total integrated
compute of about 100k GPU hours. In this paper we provide the description of
the setup, the problems that were discovered and overcome, as well as a short
description of the actual science output of the exercise.Comment: 18 pages, 5 figures, 4 tables, to be published in Proceedings of ISC
High Performance 202
The evolution of bits and bottlenecks in a scientific workflow trying to keep up with technology: Accelerating 4D image segmentation applied to nasa data
In 2016, a team of earth scientists directly engaged a team of computer scientists to identify cyberinfrastructure (CI) approaches that would speed up an earth science workflow. This paper describes the evolution of that workflow as the two teams bridged CI and an image segmentation algorithm to do large scale earth science research. The Pacific Research Platform (PRP) and The Cognitive Hardware and Software Ecosystem Community Infrastructure (CHASE-CI) resources were used to significantly decreased the earth science workflow's wall-clock time from 19.5 days to 53 minutes. The improvement in wall-clock time comes from the use of network appliances, improved image segmentation, deployment of a containerized workflow, and the increase in CI experience and training for the earth scientists. This paper presents a description of the evolving innovations used to improve the workflow, bottlenecks identified within each workflow version, and improvements made within each version of the workflow, over a three-year time period
The Scalable Systems Laboratory: a Platform for Software Innovation for HEP
The Scalable Systems Laboratory (SSL), part of the IRIS-HEP Software
Institute, provides Institute participants and HEP software developers
generally with a means to transition their R&D from conceptual toys to testbeds
to production-scale prototypes. The SSL enables tooling, infrastructure, and
services supporting the innovation of novel analysis and data architectures,
development of software elements and tool-chains, reproducible functional and
scalability testing of service components, and foundational systems R&D for
accelerated services developed by the Institute. The SSL is constructed with a
core team having expertise in scale testing and deployment of services across a
wide range of cyberinfrastructure. The core team embeds and partners with other
areas in the Institute, and with LHC and other HEP development and operations
teams as appropriate, to define investigations and required service deployment
patterns. We describe the approach and experiences with early application
deployments, including analysis platforms and intelligent data delivery
systems
The Gluonic Decay of the --Quark and tne --Meson
The observed inclusive decay of B-mesons into eta' + X is interpreted as the
consequence of the gluonic decay of the b-quark into an s-quark. As a result of
the QCD anomaly this decay proceeds partly as the decay b ---> s + eta',
similar to b ---> s + J/psi. The hadronic recoiling system is found to have a
relatively large mass. Analogously one expects a decay of the type b ---> s +
sigma. The branching ratios for these decays are large (of the order of 10%).
The results indicate that there is no room for an anomalously large
chromomagnetic decay mode of the b-quark. Gluon jets are expected to exhibit an
anomalously large tendency to fragment into eta'- and sigma-mesons.Comment: 13 latex page
Parallelized and Vectorized Tracking Using Kalman Filters with CMS Detector Geometry and Events
The High-Luminosity Large Hadron Collider at CERN will be characterized by
greater pileup of events and higher occupancy, making the track reconstruction
even more computationally demanding. Existing algorithms at the LHC are based
on Kalman filter techniques with proven excellent physics performance under a
variety of conditions. Starting in 2014, we have been developing
Kalman-filter-based methods for track finding and fitting adapted for many-core
SIMD processors that are becoming dominant in high-performance systems.
This paper summarizes the latest extensions to our software that allow it to
run on the realistic CMS-2017 tracker geometry using CMSSW-generated events,
including pileup. The reconstructed tracks can be validated against either the
CMSSW simulation that generated the hits, or the CMSSW reconstruction of the
tracks. In general, the code's computational performance has continued to
improve while the above capabilities were being added. We demonstrate that the
present Kalman filter implementation is able to reconstruct events with
comparable physics performance to CMSSW, while providing generally better
computational performance. Further plans for advancing the software are
discussed
- …