41 research outputs found
On combining multi-normalization and ancillary measures for the optimal score level fusion of fingerprint and voice biometrics
Identification and reconstruction of low-energy electrons in the ProtoDUNE-SP detector
Measurements of electrons from interactions are crucial for the Deep
Underground Neutrino Experiment (DUNE) neutrino oscillation program, as well as
searches for physics beyond the standard model, supernova neutrino detection,
and solar neutrino measurements. This article describes the selection and
reconstruction of low-energy (Michel) electrons in the ProtoDUNE-SP detector.
ProtoDUNE-SP is one of the prototypes for the DUNE far detector, built and
operated at CERN as a charged particle test beam experiment. A sample of
low-energy electrons produced by the decay of cosmic muons is selected with a
purity of 95%. This sample is used to calibrate the low-energy electron energy
scale with two techniques. An electron energy calibration based on a cosmic ray
muon sample uses calibration constants derived from measured and simulated
cosmic ray muon events. Another calibration technique makes use of the
theoretically well-understood Michel electron energy spectrum to convert
reconstructed charge to electron energy. In addition, the effects of detector
response to low-energy electron energy scale and its resolution including
readout electronics threshold effects are quantified. Finally, the relation
between the theoretical and reconstructed low-energy electron energy spectrum
is derived and the energy resolution is characterized. The low-energy electron
selection presented here accounts for about 75% of the total electron deposited
energy. After the addition of lost energy using a Monte Carlo simulation, the
energy resolution improves from about 40% to 25% at 50~MeV. These results are
used to validate the expected capabilities of the DUNE far detector to
reconstruct low-energy electrons.Comment: 19 pages, 10 figure
Impact of cross-section uncertainties on supernova neutrino spectral parameter fitting in the Deep Underground Neutrino Experiment
A primary goal of the upcoming Deep Underground Neutrino Experiment (DUNE) is
to measure the MeV neutrinos produced by a Galactic
core-collapse supernova if one should occur during the lifetime of the
experiment. The liquid-argon-based detectors planned for DUNE are expected to
be uniquely sensitive to the component of the supernova flux, enabling
a wide variety of physics and astrophysics measurements. A key requirement for
a correct interpretation of these measurements is a good understanding of the
energy-dependent total cross section for charged-current
absorption on argon. In the context of a simulated extraction of
supernova spectral parameters from a toy analysis, we investigate the
impact of modeling uncertainties on DUNE's supernova neutrino
physics sensitivity for the first time. We find that the currently large
theoretical uncertainties on must be substantially reduced
before the flux parameters can be extracted reliably: in the absence of
external constraints, a measurement of the integrated neutrino luminosity with
less than 10\% bias with DUNE requires to be known to about 5%.
The neutrino spectral shape parameters can be known to better than 10% for a
20% uncertainty on the cross-section scale, although they will be sensitive to
uncertainties on the shape of . A direct measurement of
low-energy -argon scattering would be invaluable for improving the
theoretical precision to the needed level.Comment: 25 pages, 21 figure
Highly-parallelized simulation of a pixelated LArTPC on a GPU
The rapid development of general-purpose computing on graphics processing units (GPGPU) is allowing the implementation of highly-parallelized Monte Carlo simulation chains for particle physics experiments. This technique is particularly suitable for the simulation of a pixelated charge readout for time projection chambers, given the large number of channels that this technology employs. Here we present the first implementation of a full microphysical simulator of a liquid argon time projection chamber (LArTPC) equipped with light readout and pixelated charge readout, developed for the DUNE Near Detector. The software is implemented with an end-to-end set of GPU-optimized algorithms. The algorithms have been written in Python and translated into CUDA kernels using Numba, a just-in-time compiler for a subset of Python and NumPy instructions. The GPU implementation achieves a speed up of four orders of magnitude compared with the equivalent CPU version. The simulation of the current induced on 10^3 pixels takes around 1 ms on the GPU, compared with approximately 10 s on the CPU. The results of the simulation are compared against data from a pixel-readout LArTPC prototype
Design and Deployment of Identity Recognition Systems
Modern Identity Recognition Systems (IRSs) can benefit from the simultaneous use of different biometric parameters but their activity cannot be modeled using a rigid sequential path because they have to take into account outcomes from different data as well as from different algorithms at different times. This paper is about the representation of an ad hoc composition of elementary recognition service in order to form an integrated recognition facilities, to be deployed within the scope of energy-aware systems. We aim to investigate the types of useful collaboration among IRSs components and to present a tool to assist designers both in the design process and in the system deployment phase. The tool define a catalogue of elementary recognition services that are typically associated with single trait recognition modules. They are considered as prototypical services that can be differentiated by proper parameter settings. Those basic building blocks can form compound collaborative patterns: composition rules are expressed both as logical association and causal relation (what-if rules). The recognition activity can be organized through different levels and feedback is allowed. Partition let to evaluate the overall QoS from performance figures that are associated with the elementary blocks, in order to certify expected Service Level Agreements (SLA)s. Deployment represents an important step before system release. The proposed approach is an interactive one: system administrators are asked to allocate all the components on actual machines and secure communication channels are automatically established. Scalability, redundancy and fault tolerance and energy issues can be addressed at this level