4 research outputs found
Integration of NEMO into an existing particle physics environment through virtualization
With the ever-growing amount of data collected with the experiments at the
Large Hadron Collider (LHC) (Evans et al., 2008), the need for computing resources
that can handle the analysis of this data is also rapidly increasing. This
increase will even be amplified after upgrading to the High Luminosity LHC
(Apollinari et al., 2017). High-Performance Computing (HPC) and other cluster
computing resources provided by universities can be useful supplements to the
resources dedicated to the experiment as part of the Worldwide LHC Computing
Grid (WLCG) (Eck et al., 2005) for data analysis and production of simulated
event samples. Computing resources in the WLCG are structured in four layers
– so-called Tiers. The first layer comprises two Tier-0 computing centres
located at CERN in Geneva, Switzerland and at the Wigner Research Centre
for Physics in Budapest, Hungary. The second layer consists of thirteen Tier-1
centres, followed by 160 Tier-2 sites, which are typically universities and other
scientific institutes. The final layer are Tier-3 sites which are directly used by
local users. The University of Freiburg is operating a combined Tier-2/Tier-3,
the ATLAS-BFG (Backofen et al., 2006). The shared HPC cluster »NEMO« at
the University of Freiburg has been made available to local ATLAS (Aad et al.,
2008) users through the provisioning of virtual machines incorporating the ATLAS software environment analogously to the bare metal system at the Tier-3. In
addition to the provisioning of the virtual environment, the on-demand integration
of these resources into the Tier-3 scheduler in a dynamic way is described. In
order to provide the external NEMO resources to the user in a transparent way,
an intermediate layer connecting the two batch systems is put into place. This
resource scheduler monitors requirements on the user-facing system and requests
resources on the backend-system
AUDITOR: Accounting for opportunistic resources
The increasing computational demand in High Energy Physics (HEP) as well as increasing concerns about energy efficiency in highperformance/high-throughput computing are driving forces in the search for more efficient ways to utilise available resources. Since avoiding idle resources is key in achieving high efficiency, an appropriate measure is sharing of idle resources of underutilised sites with fully occupied sites. The software COBalD/TARDIS can automatically, transparently, and dynamically (dis)integrate such resources in an opportunistic manner. Sharing resources however also requires accounting.
In this work we introduce AUDITOR (AccoUnting DatahandlIng Toolbox for Opportunistic Resources), a flexible and extensible accounting system that is able to cover a wide range of use cases and infrastructures. AUDITOR gathers accounting data via so-called collectors which are designed to monitor batch systems, COBalD/TARDIS, cloud schedulers, or other sources of information. The data is stored in a database and provided to so-called plugins, which act based on accounting records. An action could for instance be creating a bill of utilised resources, computing the CO2 footprint, adjusting parameters of a service, or forwarding accounting information to other accounting systems. Depending on the use case, a suitable collector and plugin can be chosen from a growing ecosystem of collectors and plugins. Libraries for interacting with AUDITOR are provided to facilitate the development of collectors and plugins by the community