267 research outputs found
Inclusive Flavour Tagging Algorithm
Identifying the flavour of neutral mesons production is one of the most
important components needed in the study of time-dependent violation. The
harsh environment of the Large Hadron Collider makes it particularly hard to
succeed in this task. We present an inclusive flavour-tagging algorithm as an
upgrade of the algorithms currently used by the LHCb experiment. Specifically,
a probabilistic model which efficiently combines information from reconstructed
vertices and tracks using machine learning is proposed. The algorithm does not
use information about underlying physics process. It reduces the dependence on
the performance of lower level identification capacities and thus increases the
overall performance. The proposed inclusive flavour-tagging algorithm is
applicable to tag the flavour of mesons in any proton-proton experiment.Comment: 5 pages, 5 figures, 17th International workshop on Advanced Computing
and Analysis Techniques in physics research (ACAT-2016
Reproducible Experiment Platform
Data analysis in fundamental sciences nowadays is an essential process that
pushes frontiers of our knowledge and leads to new discoveries. At the same
time we can see that complexity of those analyses increases fast due to
a)~enormous volumes of datasets being analyzed, b)~variety of techniques and
algorithms one have to check inside a single analysis, c)~distributed nature of
research teams that requires special communication media for knowledge and
information exchange between individual researchers. There is a lot of
resemblance between techniques and problems arising in the areas of industrial
information retrieval and particle physics. To address those problems we
propose Reproducible Experiment Platform (REP), a software infrastructure to
support collaborative ecosystem for computational science. It is a Python based
solution for research teams that allows running computational experiments on
shared datasets, obtaining repeatable results, and consistent comparisons of
the obtained results. We present some key features of REP based on case studies
which include trigger optimization and physics analysis studies at the LHCb
experiment.Comment: 21st International Conference on Computing in High Energy Physics
(CHEP2015), 6 page
Unsupervised ASR via Cross-Lingual Pseudo-Labeling
Recent work has shown that it is possible to train an
automatic speech recognition (ASR) system using only unpaired audio and text.
Existing unsupervised ASR methods assume that no labeled data can be used for
training. We argue that even if one does not have any labeled audio for a given
language, there is labeled data available for other
languages. We show that it is possible to use character-level acoustic models
(AMs) from other languages to bootstrap an AM in a new
language. Here, "unsupervised" means no labeled audio is available for the
language. Our approach is based on two key ingredients: (i)
generating pseudo-labels (PLs) of the language using some
language AM and (ii) constraining these PLs with a
. Our approach is effective on Common Voice:
e.g. transfer of English AM to Swahili achieves 18% WER. It also outperforms
character-based wav2vec-U 2.0 by 15% absolute WER on LJSpeech with 800h of
labeled German data instead of 60k hours of unlabeled English data.Comment: under revie
LHCb Topological Trigger Reoptimization
The main b-physics trigger algorithm used by the LHCb experiment is the
so-called topological trigger. The topological trigger selects vertices which
are a) detached from the primary proton-proton collision and b) compatible with
coming from the decay of a b-hadron. In the LHC Run 1, this trigger, which
utilized a custom boosted decision tree algorithm, selected a nearly 100% pure
sample of b-hadrons with a typical efficiency of 60-70%; its output was used in
about 60% of LHCb papers. This talk presents studies carried out to optimize
the topological trigger for LHC Run 2. In particular, we have carried out a
detailed comparison of various machine learning classifier algorithms, e.g.,
AdaBoost, MatrixNet and neural networks. The topological trigger algorithm is
designed to select all "interesting" decays of b-hadrons, but cannot be trained
on every such decay. Studies have therefore been performed to determine how to
optimize the performance of the classification algorithm on decays not used in
the training. Methods studied include cascading, ensembling and blending
techniques. Furthermore, novel boosting techniques have been implemented that
will help reduce systematic uncertainties in Run 2 measurements. We demonstrate
that the reoptimized topological trigger is expected to significantly improve
on the Run 1 performance for a wide range of b-hadron decays.Comment: 21st International Conference on Computing in High Energy Physics
(CHEP2015
More Speaking or More Speakers?
Self-training (ST) and self-supervised learning (SSL) methods have
demonstrated strong improvements in automatic speech recognition (ASR). In
spite of these advances, to the best of our knowledge, there is no analysis of
how the composition of the labelled and unlabelled datasets used in these
methods affects the results. In this work we aim to analyse the effect of
numbers of speakers in the training data on a recent SSL algorithm (wav2vec
2.0), and a recent ST algorithm (slimIPL). We perform a systematic analysis on
both labeled and unlabeled data by varying the number of speakers while keeping
the number of hours fixed and vice versa. Our findings suggest that SSL
requires a large amount of unlabeled data to produce high accuracy results,
while ST requires a sufficient number of speakers in the labelled data,
especially in the low-regime setting. In this manner these two approaches
improve supervised learning in different regimes of dataset composition
Continuous Pseudo-Labeling from the Start
Self-training (ST), or pseudo-labeling has sparked significant interest in
the automatic speech recognition (ASR) community recently because of its
success in harnessing unlabeled data. Unlike prior semi-supervised learning
approaches that relied on iteratively regenerating pseudo-labels (PLs) from a
trained model and using them to train a new model, recent state-of-the-art
methods perform `continuous training' where PLs are generated using a very
recent version of the model being trained. Nevertheless, these approaches still
rely on bootstrapping the ST using an initial supervised learning phase where
the model is trained on labeled data alone. We believe this has the potential
for over-fitting to the labeled dataset in low resource settings and that ST
from the start of training should reduce over-fitting. In this paper we show
how we can do this by dynamically controlling the evolution of PLs during the
training process in ASR. To the best of our knowledge, this is the first study
that shows the feasibility of generating PLs from the very start of the
training. We are able to achieve this using two techniques that avoid
instabilities which lead to degenerate models that do not generalize. Firstly,
we control the evolution of PLs through a curriculum that uses the online
changes in PLs to control the membership of the cache of PLs and improve
generalization. Secondly, we find that by sampling transcriptions from the
predictive distribution, rather than only using the best transcription, we can
stabilize training further. With these techniques, our ST models match prior
works without an external language model.Comment: To appear in ICLR 202
Importance of Smoothness Induced by Optimizers in FL4ASR: Towards Understanding Federated Learning for End-to-End ASR
In this paper, we start by training End-to-End Automatic Speech Recognition
(ASR) models using Federated Learning (FL) and examining the fundamental
considerations that can be pivotal in minimizing the performance gap in terms
of word error rate between models trained using FL versus their centralized
counterpart. Specifically, we study the effect of (i) adaptive optimizers, (ii)
loss characteristics via altering Connectionist Temporal Classification (CTC)
weight, (iii) model initialization through seed start, (iv) carrying over
modeling setup from experiences in centralized training to FL, e.g., pre-layer
or post-layer normalization, and (v) FL-specific hyperparameters, such as
number of local epochs, client sampling size, and learning rate scheduler,
specifically for ASR under heterogeneous data distribution. We shed light on
how some optimizers work better than others via inducing smoothness. We also
summarize the applicability of algorithms, trends, and propose best practices
from prior works in FL (in general) toward End-to-End ASR models.Comment: In Proceedings of the IEEE Automatic Speech Recognition and
Understanding Workshop (ASRU) 202
- …