38,279 research outputs found
Logarithmic Communication for Distributed Optimization in Multi-Agent Systems
Classically, the design of multi-agent systems is approached using techniques from distributed optimization such as dual descent and consensus algorithms. Such algorithms depend on convergence to global consensus before any individual agent can determine its local action. This leads to challenges with respect to communication overhead and robustness, and improving algorithms with respect to these measures has been a focus of the community for decades.
This paper presents a new approach for multi-agent system design based on ideas from the emerging field of local computation algorithms. The framework we develop, LOcal Convex Optimization (LOCO), is the first local computation algorithm for convex optimization problems and can be applied in a wide-variety of settings. We demonstrate the generality of the framework via applications to Network Utility Maximization (NUM) and the distributed training of Support Vector Machines (SVMs), providing numerical results illustrating the improvement compared to classical distributed optimization approaches in each case
Consensus-Based Transfer Linear Support Vector Machines for Decentralized Multi-Task Multi-Agent Learning
Transfer learning has been developed to improve the performances of different
but related tasks in machine learning. However, such processes become less
efficient with the increase of the size of training data and the number of
tasks. Moreover, privacy can be violated as some tasks may contain sensitive
and private data, which are communicated between nodes and tasks. We propose a
consensus-based distributed transfer learning framework, where several tasks
aim to find the best linear support vector machine (SVM) classifiers in a
distributed network. With alternating direction method of multipliers, tasks
can achieve better classification accuracies more efficiently and privately, as
each node and each task train with their own data, and only decision variables
are transferred between different tasks and nodes. Numerical experiments on
MNIST datasets show that the knowledge transferred from the source tasks can be
used to decrease the risks of the target tasks that lack training data or have
unbalanced training labels. We show that the risks of the target tasks in the
nodes without the data of the source tasks can also be reduced using the
information transferred from the nodes who contain the data of the source
tasks. We also show that the target tasks can enter and leave in real-time
without rerunning the whole algorithm
- …