645 research outputs found

    Scalable Declarative HEP Analysis Workflows for Containerised Compute Clouds

    Get PDF
    We describe a novel approach for experimental High-Energy Physics (HEP) data analyses that is centred around the declarative rather than imperative paradigm when describing analysis computational tasks. The analysis process can be structured in the form of a Directed Acyclic Graph (DAG), where each graph vertex represents a unit of computation with its inputs and outputs, and the graph edges describe the interconnection of various computational steps. We have developed REANA, a platform for reproducible data analyses, that supports several such DAG workflow specifications. The REANA platform parses the analysis workflow and dispatches its computational steps to various supported computing backends (Kubernetes, HTCondor, Slurm). The focus on declarative rather than imperative programming enables researchers to concentrate on the problem domain at hand without having to think about implementation details such as scalable job orchestration. The declarative programming approach is further exemplified by a multi-level job cascading paradigm that was implemented in the Yadage workflow specification language. We present two recent LHC particle physics analyses, ATLAS searches for dark matter and CMS jet energy correction pipelines, where the declarative approach was successfully applied. We argue that the declarative approach to data analyses, combined with recent advancements in container technology, facilitates the portability of computational data analyses to various compute backends, enhancing the reproducibility and the knowledge preservation behind particle physics data analyses.Peer reviewe

    Open data provenance and reproducibility : a case study from publishing CMS open data

    Get PDF
    In this paper we present the latest CMS open data release published on the CERN Oopen Data portal. Samples of collision and simulated datasets were released together with detailed information about the data provenance. The associated data production chains cover the necessary computing environments, the configuration files and the computational procedures used in each data production step. We describe data curation techniques used to obtain and publish the data provenance information and we study the possibility of reproducing parts of the released data using the publicly available information. The present work demonstrates the usefulness of releasing selected samples of raw and primary data in order to fully ensure the completeness of information about the data production chain for the attention of general data scientists and other non-specialists interested in using particle physics data for education or research purposes.In this paper we present the latest CMS open data release published on the CERN Oopen Data portal. Samples of collision and simulated datasets were released together with detailed information about the data provenance. The associated data production chains cover the necessary computing environments, the configuration files and the computational procedures used in each data production step. We describe data curation techniques used to obtain and publish the data provenance information and we study the possibility of reproducing parts of the released data using the publicly available information. The present work demonstrates the usefulness of releasing selected samples of raw and primary data in order to fully ensure the completeness of information about the data production chain for the attention of general data scientists and other non-specialists interested in using particle physics data for education or research purposes.Peer reviewe

    Studier av jet energi korrigeringar på CMS experimentet samt automatiseringen av dem

    Get PDF
    At the Compact Muon Solenoid (CMS) experiment at CERN (European Organization for Nuclear Research), the building blocks of the Universe are investigated by analysing the observed final-state particles resulting from high-energy proton-proton collisions. However, direct detection of final-state quarks and gluons is not possible due to a phenomenon known as colour confinement. Instead, event properties with a close correspondence with their distributions are studied. These event properties are known as jets. Jets are central to particle physics analysis and our understanding of them, and hence of our Universe, is dependent upon our ability to accurately measure their energy. Unfortunately, current detector technology is imprecise, necessitating downstream correction of measurement discrepancies. To achieve this, the CMS experiment employs a sequential multi-step jet calibration process. The process is performed several times per year, and more often during periods of data collection. Automating the jet calibration would increase the efficiency of the CMS experiment. By automating the code execution, the workflow could be performed independently of the analyst. This in turn, would speed up the analysis and reduce analyst workload. In addition, automation facilitates higher levels of reproducibility. In this thesis, a novel method for automating the derivation of jet energy corrections from simulation is presented. To achieve automation, the methodology utilises declarative programming. The analyst is simply required to express what should be executed, and no longer needs to determine how to execute it. To successfully automate the computation of jet energy corrections, it is necessary to capture detailed information concerning both the computational steps and the computational environment. The former is achieved with a computational workflow, and the latter using container technology. This allows a portable and scalable workflow to be achieved, which is easy to maintain and compare to previous runs. The results of this thesis strongly suggest that capturing complex experimental particle physics analyses with declarative workflow languages is both achievable and advantageous. The productivity of the analyst was improved, and reproducibility facilitated. However, the method is not without its challenges. Declarative programming requires the analyst to think differently about the problem at hand. As a result there are some sociological challenges to methodological uptake. However, once the extensive benefits are understood, we anticipate widespread adoption of this approach.På experimentet Compact Muon Solenoid (CMS) vid CERN (Europeiska organisationen för kärnforskning) undersöks universums byggstenar genom att analysera de observerade sluttillstånds partiklarna som härrör från proton-proton kollisioner med hög energi. Direkt detektering av sluttillstånds kvarkar och gluoner är dock inte möjligt på grund av ett fenomen känt som färg inneslutning. Istället studeras händelse egenskaper med en nära överensstämmelse med deras distributioner. Dessa händelse egenskaper är kända som jets. Jets är centrala för partikelfysik analys och vår förståelse av dem, och därmed förståelsen av vårt universum, är beroende av vår förmåga att noggrant mäta deras energi. Tyvärr är nuvarande detektorteknik inte exakt, vilket kräver korrigering av mät avvikelser. För att uppnå detta använder CMS-experimentet en sekventiell flerstegs jet kalibreringsprocess. Processen utförs flera gånger per år, och oftare under datainsamling perioder. Att automatisera jet kalibreringen skulle öka effektiviteten i CMS-experimentet. Genom att automatisera kod körningen kunde arbetsflödet utföras oberoende av analytikerns närvarande. Detta skulle i sin tur påskynda analysen och minska analytikerns arbetsbelastning. Dessutom möjliggör automatisering högre nivåer av reproducerbarhet. I denna avhandling presenteras en ny metod för att automatisera härledningen av jet energi korrigeringar från simulering. För att uppnå automatisering använder metoden deklarativ programmering. Av analytikern krävs enbart att uttrycka vad som ska utföras och hen behöver inte längre avgöra hur den ska utföras. För att framgångsrikt automatisera beräkningen av jet korrigeringar är det nödvändigt att fånga detaljerad information om både beräknings stegen och beräknings miljön. Det förstnämnda uppnås med ett beräknings arbetsflöde och det senare med container teknologi. Detta möjliggör ett bärbart och skalbart arbetsflöde,, vilket är enkelt att förvalta och jämföra med tidigare körningar. Resultaten av denna avhandling tyder starkt på att det är både uppnåeligt och fördelaktigt att fånga komplexa experimentella partikelfysik analyser med deklarativa arbetsflödes språk. Produktiviteten hos analytikern förbättrades och reproducerbarheten underlättades. Metoden är dock inte utan sina utmaningar. Deklarativ programmering kräver att analytikern tänker annorlunda om det aktuella problemet. Som ett resultat finns det ett par sociologiska utmaningar för metodens upptagningen. Men när de omfattande fördelarna har förståtts, förväntar vi oss en bred tillämpning av detta tillvägagångssätt

    Measurement of the double-differential inclusive jet cross section in proton-proton collisions at s\sqrt{s} = 5.02 TeV

    No full text
    International audienceThe inclusive jet cross section is measured as a function of jet transverse momentum pTp_\mathrm{T} and rapidity yy. The measurement is performed using proton-proton collision data at s\sqrt{s} = 5.02 TeV, recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 27.4 pb1^{-1}. The jets are reconstructed with the anti-kTk_\mathrm{T} algorithm using a distance parameter of RR = 0.4, within the rapidity interval y\lvert y\rvert<\lt 2, and across the kinematic range 0.06 <\ltpTp_\mathrm{T}<\lt 1 TeV. The jet cross section is unfolded from detector to particle level using the determined jet response and resolution. The results are compared to predictions of perturbative quantum chromodynamics, calculated at both next-to-leading order and next-to-next-to-leading order. The predictions are corrected for nonperturbative effects, and presented for a variety of parton distribution functions and choices of the renormalization/factorization scales and the strong coupling αS\alpha_\mathrm{S}

    Measurement of the double-differential inclusive jet cross section in proton-proton collisions at s= \sqrt{s} = 5.02 TeV

    No full text
    The inclusive jet cross section is measured as a function of jet transverse momentum pT p_{\mathrm{T}} and rapidity y y . The measurement is performed using proton-proton collision data at s= \sqrt{s} = 5.02 TeV, recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 27.4pb1\,\text{pb}^{-1}. The jets are reconstructed with the anti-kT k_{\mathrm{T}} algorithm using a distance parameter of R= R= 0.4, within the rapidity interval y< |y| < 2, and across the kinematic range 0.06 <pT< < p_{\mathrm{T}} < 1 TeV. The jet cross section is unfolded from detector to particle level using the determined jet response and resolution. The results are compared to predictions of perturbative quantum chromodynamics, calculated at both next-to-leading order and next-to-next-to-leading order. The predictions are corrected for nonperturbative effects, and presented for a variety of parton distribution functions and choices of the renormalization/factorization scales and the strong coupling αS \alpha_\mathrm{S} .The inclusive jet cross section is measured as a function of jet transverse momentum pTp_\mathrm{T} and rapidity yy. The measurement is performed using proton-proton collision data at s\sqrt{s} = 5.02 TeV, recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 27.4 pb1^{-1}. The jets are reconstructed with the anti-kTk_\mathrm{T} algorithm using a distance parameter of RR = 0.4, within the rapidity interval y\lvert y\rvert<\lt 2, and across the kinematic range 0.06 <\ltpTp_\mathrm{T}<\lt 1 TeV. The jet cross section is unfolded from detector to particle level using the determined jet response and resolution. The results are compared to predictions of perturbative quantum chromodynamics, calculated at both next-to-leading order and next-to-next-to-leading order. The predictions are corrected for nonperturbative effects, and presented for a variety of parton distribution functions and choices of the renormalization/factorization scales and the strong coupling αS\alpha_\mathrm{S}

    Measurement of the double-differential inclusive jet cross section in proton-proton collisions at s\sqrt{s} = 5.02 TeV

    No full text
    International audienceThe inclusive jet cross section is measured as a function of jet transverse momentum pTp_\mathrm{T} and rapidity yy. The measurement is performed using proton-proton collision data at s\sqrt{s} = 5.02 TeV, recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 27.4 pb1^{-1}. The jets are reconstructed with the anti-kTk_\mathrm{T} algorithm using a distance parameter of RR = 0.4, within the rapidity interval y\lvert y\rvert<\lt 2, and across the kinematic range 0.06 <\ltpTp_\mathrm{T}<\lt 1 TeV. The jet cross section is unfolded from detector to particle level using the determined jet response and resolution. The results are compared to predictions of perturbative quantum chromodynamics, calculated at both next-to-leading order and next-to-next-to-leading order. The predictions are corrected for nonperturbative effects, and presented for a variety of parton distribution functions and choices of the renormalization/factorization scales and the strong coupling αS\alpha_\mathrm{S}

    Measurement of the double-differential inclusive jet cross section in proton-proton collisions at s\sqrt{s} = 5.02 TeV

    No full text
    International audienceThe inclusive jet cross section is measured as a function of jet transverse momentum pTp_\mathrm{T} and rapidity yy. The measurement is performed using proton-proton collision data at s\sqrt{s} = 5.02 TeV, recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 27.4 pb1^{-1}. The jets are reconstructed with the anti-kTk_\mathrm{T} algorithm using a distance parameter of RR = 0.4, within the rapidity interval y\lvert y\rvert<\lt 2, and across the kinematic range 0.06 <\ltpTp_\mathrm{T}<\lt 1 TeV. The jet cross section is unfolded from detector to particle level using the determined jet response and resolution. The results are compared to predictions of perturbative quantum chromodynamics, calculated at both next-to-leading order and next-to-next-to-leading order. The predictions are corrected for nonperturbative effects, and presented for a variety of parton distribution functions and choices of the renormalization/factorization scales and the strong coupling αS\alpha_\mathrm{S}

    Measurement of the double-differential inclusive jet cross section in proton-proton collisions at s\sqrt{s} = 5.02 TeV

    No full text
    International audienceThe inclusive jet cross section is measured as a function of jet transverse momentum pTp_\mathrm{T} and rapidity yy. The measurement is performed using proton-proton collision data at s\sqrt{s} = 5.02 TeV, recorded by the CMS experiment at the LHC, corresponding to an integrated luminosity of 27.4 pb1^{-1}. The jets are reconstructed with the anti-kTk_\mathrm{T} algorithm using a distance parameter of RR = 0.4, within the rapidity interval y\lvert y\rvert<\lt 2, and across the kinematic range 0.06 <\ltpTp_\mathrm{T}<\lt 1 TeV. The jet cross section is unfolded from detector to particle level using the determined jet response and resolution. The results are compared to predictions of perturbative quantum chromodynamics, calculated at both next-to-leading order and next-to-next-to-leading order. The predictions are corrected for nonperturbative effects, and presented for a variety of parton distribution functions and choices of the renormalization/factorization scales and the strong coupling αS\alpha_\mathrm{S}

    Inclusive nonresonant multilepton probes of new phenomena at s\sqrt{s} = 13 TeV

    No full text
    An inclusive search for nonresonant signatures of beyond the standard model (SM) phenomena in events with three or more charged leptons, including hadronically decaying τ\tau leptons, is presented. The analysis is based on a data sample corresponding to an integrated luminosity of 138 fb1^{-1} of proton-proton collisions at s\sqrt{s} = 13 TeV, collected by the CMS experiment at the LHC in 2016-2018. Events are categorized based on the lepton and b-tagged jet multiplicities and various kinematic variables. Three scenarios of physics beyond the SM are probed, and signal-specific boosted decision trees are used for enhancing sensitivity. No significant deviations from the background expectations are observed. Lower limits are set at 95% confidence level on the mass of type-III seesaw heavy fermions in the range 845-1065 GeV for various decay branching fraction combinations to SM leptons. Doublet and singlet vector-like τ\tau lepton extensions of the SM are excluded for masses below 1045 GeV and in the mass range 125-150 GeV, respectively. Scalar leptoquarks decaying exclusively to a top quark and a lepton are excluded below 1.12-1.42 TeV, depending on the lepton flavor. For the type-III seesaw as well as the vector-like doublet model, these constraints are the most stringent to date. For the vector-like singlet model, these are the first constraints from the LHC experiments. Detailed results are also presented to facilitate alternative theoretical interpretations
    corecore