12,551 research outputs found
An Accelerated Decentralized Stochastic Proximal Algorithm for Finite Sums
Modern large-scale finite-sum optimization relies on two key aspects:
distribution and stochastic updates. For smooth and strongly convex problems,
existing decentralized algorithms are slower than modern accelerated
variance-reduced stochastic algorithms when run on a single machine, and are
therefore not efficient. Centralized algorithms are fast, but their scaling is
limited by global aggregation steps that result in communication bottlenecks.
In this work, we propose an efficient \textbf{A}ccelerated
\textbf{D}ecentralized stochastic algorithm for \textbf{F}inite \textbf{S}ums
named ADFS, which uses local stochastic proximal updates and randomized
pairwise communications between nodes. On machines, ADFS learns from
samples in the same time it takes optimal algorithms to learn from samples
on one machine. This scaling holds until a critical network size is reached,
which depends on communication delays, on the number of samples , and on the
network topology. We provide a theoretical analysis based on a novel augmented
graph approach combined with a precise evaluation of synchronization times and
an extension of the accelerated proximal coordinate gradient algorithm to
arbitrary sampling. We illustrate the improvement of ADFS over state-of-the-art
decentralized approaches with experiments.Comment: Code available in source files. arXiv admin note: substantial text
overlap with arXiv:1901.0986
Cloud-Based Centralized/Decentralized Multi-Agent Optimization with Communication Delays
We present and analyze a computational hybrid architecture for performing
multi-agent optimization. The optimization problems under consideration have
convex objective and constraint functions with mild smoothness conditions
imposed on them. For such problems, we provide a primal-dual algorithm
implemented in the hybrid architecture, which consists of a decentralized
network of agents into which centralized information is occasionally injected,
and we establish its convergence properties. To accomplish this, a central
cloud computer aggregates global information, carries out computations of the
dual variables based on this information, and then distributes the updated dual
variables to the agents. The agents update their (primal) state variables and
also communicate among themselves with each agent sharing and receiving state
information with some number of its neighbors. Throughout, communications with
the cloud are not assumed to be synchronous or instantaneous, and communication
delays are explicitly accounted for in the modeling and analysis of the system.
Experimental results are presented to support the theoretical developments
made.Comment: 8 pages, 4 figure
- …