15,802 research outputs found
An Analysis Tool for Push-Sum Based Distributed Optimization
The push-sum algorithm is probably the most important distributed averaging
approach over directed graphs, which has been applied to various problems
including distributed optimization. This paper establishes the explicit
absolute probability sequence for the push-sum algorithm, and based on which,
constructs quadratic Lyapunov functions for push-sum based distributed
optimization algorithms. As illustrative examples, the proposed novel analysis
tool can improve the convergence rates of the subgradient-push and stochastic
gradient-push, two important algorithms for distributed convex optimization
over unbalanced directed graphs. Specifically, the paper proves that the
subgradient-push algorithm converges at a rate of for general
convex functions and stochastic gradient-push algorithm converges at a rate of
for strongly convex functions, over time-varying unbalanced directed
graphs. Both rates are respectively the same as the state-of-the-art rates of
their single-agent counterparts and thus optimal, which closes the theoretical
gap between the centralized and push-sum based (sub)gradient methods. The paper
further proposes a heterogeneous push-sum based subgradient algorithm in which
each agent can arbitrarily switch between subgradient-push and
push-subgradient. The heterogeneous algorithm thus subsumes both
subgradient-push and push-subgradient as special cases, and still converges to
an optimal point at an optimal rate. The proposed tool can also be extended to
analyze distributed weighted averaging.Comment: arXiv admin note: substantial text overlap with arXiv:2203.16623,
arXiv:2303.1706
Towards Advantages of Parameterized Quantum Pulses
The advantages of quantum pulses over quantum gates have attracted increasing
attention from researchers. Quantum pulses offer benefits such as flexibility,
high fidelity, scalability, and real-time tuning. However, while there are
established workflows and processes to evaluate the performance of quantum
gates, there has been limited research on profiling parameterized pulses and
providing guidance for pulse circuit design. To address this gap, our study
proposes a set of design spaces for parameterized pulses, evaluating these
pulses based on metrics such as expressivity, entanglement capability, and
effective parameter dimension. Using these design spaces, we demonstrate the
advantages of parameterized pulses over gate circuits in the aspect of duration
and performance at the same time thus enabling high-performance quantum
computing. Our proposed design space for parameterized pulse circuits has shown
promising results in quantum chemistry benchmarks.Comment: 11 Figures, 4 Table
Sensitivity analysis for ReaxFF reparameterization using the Hilbert-Schmidt independence criterion
We apply a global sensitivity method, the Hilbert-Schmidt independence
criterion (HSIC), to the reparameterization of a Zn/S/H ReaxFF force field to
identify the most appropriate parameters for reparameterization. Parameter
selection remains a challenge in this context as high dimensional optimizations
are prone to overfitting and take a long time, but selecting too few parameters
leads to poor quality force fields. We show that the HSIC correctly and quickly
identifies the most sensitive parameters, and that optimizations done using a
small number of sensitive parameters outperform those done using a higher
dimensional reasonable-user parameter selection. Optimizations using only
sensitive parameters: 1) converge faster, 2) have loss values comparable to
those found with the naive selection, 3) have similar accuracy in validation
tests, and 4) do not suffer from problems of overfitting. We demonstrate that
an HSIC global sensitivity is a cheap optimization pre-processing step that has
both qualitative and quantitative benefits which can substantially simplify and
speedup ReaxFF reparameterizations.Comment: author accepted manuscrip
Conditional Adapters: Parameter-efficient Transfer Learning with Fast Inference
We propose Conditional Adapter (CoDA), a parameter-efficient transfer
learning method that also improves inference efficiency. CoDA generalizes
beyond standard adapter approaches to enable a new way of balancing speed and
accuracy using conditional computation. Starting with an existing dense
pretrained model, CoDA adds sparse activation together with a small number of
new parameters and a light-weight training phase. Our experiments demonstrate
that the CoDA approach provides an unexpectedly efficient way to transfer
knowledge. Across a variety of language, vision, and speech tasks, CoDA
achieves a 2x to 8x inference speed-up compared to the state-of-the-art Adapter
approach with moderate to no accuracy loss and the same parameter efficiency
ADS_UNet: A Nested UNet for Histopathology Image Segmentation
The UNet model consists of fully convolutional network (FCN) layers arranged
as contracting encoder and upsampling decoder maps. Nested arrangements of
these encoder and decoder maps give rise to extensions of the UNet model, such
as UNete and UNet++. Other refinements include constraining the outputs of the
convolutional layers to discriminate between segment labels when trained end to
end, a property called deep supervision. This reduces feature diversity in
these nested UNet models despite their large parameter space. Furthermore, for
texture segmentation, pixel correlations at multiple scales contribute to the
classification task; hence, explicit deep supervision of shallower layers is
likely to enhance performance. In this paper, we propose ADS UNet, a stage-wise
additive training algorithm that incorporates resource-efficient deep
supervision in shallower layers and takes performance-weighted combinations of
the sub-UNets to create the segmentation model. We provide empirical evidence
on three histopathology datasets to support the claim that the proposed ADS
UNet reduces correlations between constituent features and improves performance
while being more resource efficient. We demonstrate that ADS_UNet outperforms
state-of-the-art Transformer-based models by 1.08 and 0.6 points on CRAG and
BCSS datasets, and yet requires only 37% of GPU consumption and 34% of training
time as that required by Transformers.Comment: To be published in Expert Systems With Application
Soliton Gas: Theory, Numerics and Experiments
The concept of soliton gas was introduced in 1971 by V. Zakharov as an
infinite collection of weakly interacting solitons in the framework of
Korteweg-de Vries (KdV) equation. In this theoretical construction of a diluted
soliton gas, solitons with random parameters are almost non-overlapping. More
recently, the concept has been extended to dense gases in which solitons
strongly and continuously interact. The notion of soliton gas is inherently
associated with integrable wave systems described by nonlinear partial
differential equations like the KdV equation or the one-dimensional nonlinear
Schr\"odinger equation that can be solved using the inverse scattering
transform. Over the last few years, the field of soliton gases has received a
rapidly growing interest from both the theoretical and experimental points of
view. In particular, it has been realized that the soliton gas dynamics
underlies some fundamental nonlinear wave phenomena such as spontaneous
modulation instability and the formation of rogue waves. The recently
discovered deep connections of soliton gas theory with generalized
hydrodynamics have broadened the field and opened new fundamental questions
related to the soliton gas statistics and thermodynamics. We review the main
recent theoretical and experimental results in the field of soliton gas. The
key conceptual tools of the field, such as the inverse scattering transform,
the thermodynamic limit of finite-gap potentials and the Generalized Gibbs
Ensembles are introduced and various open questions and future challenges are
discussed.Comment: 35 pages, 8 figure
ENABLING EFFICIENT FLEET COMPOSITION SELECTION THROUGH THE DEVELOPMENT OF A RANK HEURISTIC FOR A BRANCH AND BOUND METHOD
In the foreseeable future, autonomous mobile robots (AMRs) will become a key enabler
for increasing productivity and flexibility in material handling in warehousing facilities,
distribution centers and manufacturing systems.
The objective of this research is to develop and validate parametric models of AMRs,
develop ranking heuristic using a physics-based algorithm within the framework of the
Branch and Bound method, integrate the ranking algorithm into a Fleet Composition
Optimization (FCO) tool, and finally conduct simulations under various scenarios to
verify the suitability and robustness of the developed tool in a factory equipped with
AMRs. Kinematic-based equations are used for computing both energy and time
consumption. Multivariate linear regression, a data-driven method, is used for designing
the ranking heuristic. The results indicate that the unique physical structures and
parameters of each robot are the main factors contributing to differences in energy and
time consumption. improvement on reducing computation time was achieved by
comparing heuristic-based search and non-heuristic-based search. This research is
expected to significantly improve the current nested fleet composition optimization tool
by reducing computation time without sacrificing optimality. From a practical
perspective, greater efficiency in reducing energy and time costs can be achieved.Ford Motor CompanyNo embargoAcademic Major: Aerospace Engineerin
Neural Architecture Search: Insights from 1000 Papers
In the past decade, advances in deep learning have resulted in breakthroughs
in a variety of areas, including computer vision, natural language
understanding, speech recognition, and reinforcement learning. Specialized,
high-performing neural architectures are crucial to the success of deep
learning in these areas. Neural architecture search (NAS), the process of
automating the design of neural architectures for a given task, is an
inevitable next step in automating machine learning and has already outpaced
the best human-designed architectures on many tasks. In the past few years,
research in NAS has been progressing rapidly, with over 1000 papers released
since 2020 (Deng and Lindauer, 2021). In this survey, we provide an organized
and comprehensive guide to neural architecture search. We give a taxonomy of
search spaces, algorithms, and speedup techniques, and we discuss resources
such as benchmarks, best practices, other surveys, and open-source libraries
Rational-approximation-based model order reduction of Helmholtz frequency response problems with adaptive finite element snapshots
We introduce several spatially adaptive model order reduction approaches tailored to non-coercive elliptic boundary value problems, specifically, parametric-in-frequency Helmholtz problems. The offline information is computed by means of adaptive finite elements, so that each snapshot lives in a different discrete space that resolves the local singularities of the analytical solution and is adjusted to the considered frequency value. A rational surrogate is then assembled adopting either a least squares or an interpolatory approach, yielding a function-valued version of the standard rational interpolation method (V-SRI) and the minimal rational interpolation method (MRI). In the context of building an approximation for linear or quadratic functionals of the Helmholtz solution, we perform several numerical experiments to compare the proposed methodologies. Our simulations show that, for interior resonant problems (whose singularities are encoded by poles on the V-SRI and MRI work comparably well. Instead, when dealing with exterior scattering problems, whose frequency response is mostly smooth, the V-SRI method seems to be the best performing one
A spatio-temporal framework for modelling wastewater concentration during the COVID-19 pandemic
The potential utility of wastewater-based epidemiology as an early warning tool has been explored widely across the globe during the current COVID-19 pandemic. Methods to detect the presence of SARS-CoV-2 RNA in wastewater were developed early in the pandemic, and extensive work has been conducted to evaluate the relationship between viral concentration and COVID-19 case numbers at the catchment areas of sewage treatment works (STWs) over time. However, no attempt has been made to develop a model that predicts wastewater concentration at fine spatio-temporal resolutions covering an entire country, a necessary step towards using wastewater monitoring for the early detection of local outbreaks. We consider weekly averages of flow-normalised viral concentration, reported as the number of SARS-CoV-2N1 gene copies per litre (gc/L) of wastewater available at 303 STWs over the period between 1 June 2021 and 30 March 2022. We specify a spatially continuous statistical model that quantifies the relationship between weekly viral concentration and a collection of covariates covering socio-demographics, land cover and virus associated genomic characteristics at STW catchment areas while accounting for spatial and temporal correlation. We evaluate the model’s predictive performance at the catchment level through 10-fold cross-validation. We predict the weekly viral concentration at the population-weighted centroid of the 32,844 lower super output areas (LSOAs) in England, then aggregate these LSOA predictions to the Lower Tier Local Authority level (LTLA), a geography that is more relevant to public health policy-making. We also use the model outputs to quantify the probability of local changes of direction (increases or decreases) in viral concentration over short periods (e.g. two consecutive weeks). The proposed statistical framework can predict SARS-CoV-2 viral concentration in wastewater at high spatio-temporal resolution across England. Additionally, the probabilistic quantification of local changes can be used as an early warning tool for public health surveillance
- …