1,250 research outputs found
Numerical optimization for Artificial Retina Algorithm
High-energy physics experiments rely on reconstruction of the trajectories of
particles produced at the interaction point. This is a challenging task,
especially in the high track multiplicity environment generated by p-p
collisions at the LHC energies. A typical event includes hundreds of signal
examples (interesting decays) and a significant amount of noise (uninteresting
examples).
This work describes a modification of the Artificial Retina algorithm for
fast track finding: numerical optimization methods were adopted for fast local
track search. This approach allows for considerable reduction of the total
computational time per event. Test results on simplified simulated model of
LHCb VELO (VErtex LOcator) detector are presented. Also this approach is
well-suited for implementation of paralleled computations as GPGPU which look
very attractive in the context of upcoming detector upgrades
GRID Storage Optimization in Transparent and User-Friendly Way for LHCb Datasets
The LHCb collaboration is one of the four major experiments at the Large
Hadron Collider at CERN. Many petabytes of data are produced by the detectors
and Monte-Carlo simulations. The LHCb Grid interware LHCbDIRAC is used to make
data available to all collaboration members around the world. The data is
replicated to the Grid sites in different locations. However the Grid disk
storage is limited and does not allow keeping replicas of each file at all
sites. Thus it is essential to optimize number of replicas to achieve a better
Grid performance.
In this study, we present a new approach of data replication and distribution
strategy based on data popularity prediction. The popularity is performed based
on the data access history and metadata, and uses machine learning techniques
and time series analysis methods
Disk storage management for LHCb based on Data Popularity estimator
This paper presents an algorithm providing recommendations for optimizing the
LHCb data storage. The LHCb data storage system is a hybrid system. All
datasets are kept as archives on magnetic tapes. The most popular datasets are
kept on disks. The algorithm takes the dataset usage history and metadata
(size, type, configuration etc.) to generate a recommendation report. This
article presents how we use machine learning algorithms to predict future data
popularity. Using these predictions it is possible to estimate which datasets
should be removed from disk. We use regression algorithms and time series
analysis to find the optimal number of replicas for datasets that are kept on
disk. Based on the data popularity and the number of replicas optimization, the
algorithm minimizes a loss function to find the optimal data distribution. The
loss function represents all requirements for data distribution in the data
storage system. We demonstrate how our algorithm helps to save disk space and
to reduce waiting times for jobs using this data
Cherenkov Detectors Fast Simulation Using Neural Networks
We propose a way to simulate Cherenkov detector response using a generative
adversarial neural network to bypass low-level details. This network is trained
to reproduce high level features of the simulated detector events based on
input observables of incident particles. This allows the dramatic increase of
simulation speed. We demonstrate that this approach provides simulation
precision which is consistent with the baseline and discuss possible
implications of these results.Comment: In proceedings of 10th International Workshop on Ring Imaging
Cherenkov Detector
Reproducible Experiment Platform
Data analysis in fundamental sciences nowadays is an essential process that
pushes frontiers of our knowledge and leads to new discoveries. At the same
time we can see that complexity of those analyses increases fast due to
a)~enormous volumes of datasets being analyzed, b)~variety of techniques and
algorithms one have to check inside a single analysis, c)~distributed nature of
research teams that requires special communication media for knowledge and
information exchange between individual researchers. There is a lot of
resemblance between techniques and problems arising in the areas of industrial
information retrieval and particle physics. To address those problems we
propose Reproducible Experiment Platform (REP), a software infrastructure to
support collaborative ecosystem for computational science. It is a Python based
solution for research teams that allows running computational experiments on
shared datasets, obtaining repeatable results, and consistent comparisons of
the obtained results. We present some key features of REP based on case studies
which include trigger optimization and physics analysis studies at the LHCb
experiment.Comment: 21st International Conference on Computing in High Energy Physics
(CHEP2015), 6 page
LHCb Topological Trigger Reoptimization
The main b-physics trigger algorithm used by the LHCb experiment is the
so-called topological trigger. The topological trigger selects vertices which
are a) detached from the primary proton-proton collision and b) compatible with
coming from the decay of a b-hadron. In the LHC Run 1, this trigger, which
utilized a custom boosted decision tree algorithm, selected a nearly 100% pure
sample of b-hadrons with a typical efficiency of 60-70%; its output was used in
about 60% of LHCb papers. This talk presents studies carried out to optimize
the topological trigger for LHC Run 2. In particular, we have carried out a
detailed comparison of various machine learning classifier algorithms, e.g.,
AdaBoost, MatrixNet and neural networks. The topological trigger algorithm is
designed to select all "interesting" decays of b-hadrons, but cannot be trained
on every such decay. Studies have therefore been performed to determine how to
optimize the performance of the classification algorithm on decays not used in
the training. Methods studied include cascading, ensembling and blending
techniques. Furthermore, novel boosting techniques have been implemented that
will help reduce systematic uncertainties in Run 2 measurements. We demonstrate
that the reoptimized topological trigger is expected to significantly improve
on the Run 1 performance for a wide range of b-hadron decays.Comment: 21st International Conference on Computing in High Energy Physics
(CHEP2015
- …
