79,014 research outputs found
Distributed Optimization with Application to Power Systems and Control
In many engineering domains, systems are composed of partially independent subsystems—power systems are composed of distribution and transmission systems, teams of robots are composed of individual robots, and chemical process systems are composed of vessels, heat exchangers and reactors. Often, these subsystems should reach a common goal such as satisfying a power demand with minimum cost, flying in a formation, or reaching an optimal set-point. At the same time, limited information exchange is desirable—for confidentiality reasons but also due to communication constraints. Moreover, a fast and reliable decision process is key as applications might be safety-critical.
Mathematical optimization techniques are among the most successful tools for controlling systems optimally with feasibility guarantees. Yet, they are often centralized—all data has to be collected in one central and computationally powerful entity. Methods from distributed optimization control the subsystems in a distributed or decentralized fashion, reducing or avoiding central coordination. These methods have a long and successful history. Classical distributed optimization algorithms, however, are typically designed for convex problems. Hence, they are only partially applicable in the above domains since many of them lead to optimization problems with non-convex constraints. This thesis develops one of the first frameworks for distributed and decentralized optimization with non-convex constraints.
Based on the Augmented Lagrangian Alternating Direction Inexact Newton (ALADIN) algorithm, a bi-level distributed ALADIN framework is presented, solving the coordination step of ALADIN in a decentralized fashion. This framework can handle various decentralized inner algorithms, two of which we develop here: a decentralized variant of the Alternating Direction Method of Multipliers (ADMM) and a novel decentralized Conjugate Gradient algorithm. Decentralized conjugate gradient is to the best of our knowledge the first decentralized algorithm with a guarantee of convergence to the exact solution in a finite number of iterates. Sufficient conditions for fast local convergence of bi-level ALADIN are derived. Bi-level ALADIN strongly reduces the communication and coordination effort of ALADIN and preserves its fast convergence guarantees. We illustrate these properties on challenging problems from power systems and control, and compare performance to the widely used ADMM.
The developed methods are implemented in the open-source MATLAB toolbox ALADIN-—one of the first toolboxes for decentralized non-convex optimization. ALADIN- comes with a rich set of application examples from different domains showing its broad applicability. As an additional contribution, this thesis provides new insights why state-of-the-art distributed algorithms might encounter issues for constrained problems
Recommended from our members
Distributed learning under resource constraints
The enormous amount of data encountered in control, signal processing, and machine learning applications presents numerous computational challenges. In many settings, data is distributed across numerous devices and needs to be processed, stored and communicated, naturally leading to distributed optimization problems. It is often the case that the communication networks connecting the devices are characterized by unidirectional and time-varying communication links, and that the devices face resource constraints. Therefore, it is of importance to design learning algorithms that can achieve required accuracy and convergence properties while satisfying the computation and communication constraints. This dissertation first focuses on distributed convex optimization tasks deployed over time-varying networks with communication constraints. For this setting, we developed provably converging communication-efficient algorithms that rely on sparsification to reduce the communication cost; efficacy of the developed framework was demonstrated on computer vision and natural language processing tasks. Then we turn our attention to more general, challenging decentralized non-convex problems over directed time-varying networks with stochastic first-order oracle and local computation constraints, for which we develop a provably fast converging algorithm achieving highly accurate performance. The final part of the dissertation focuses on the practical scenario of distributed learning applications where the clients are heterogeneous and operate under constraints on the local computation power, memory footprint and communication bandwidth and propose Federated Quantized Self Supervised Learning (Fed-QSSL) algorithm, an effective framework for federated learning under bitwidth constraints and data heterogeneity. We theoretically analyze the impact of low-bit training on the convergence and robustness of federated learning, and experimentally demonstrate that Fed-QSSL achieves more robust and personalized performance than the competing methods.Electrical and Computer Engineerin
Decentralized Proximal Method of Multipliers for Convex Optimization with Coupled Constraints
In this paper, a decentralized proximal method of multipliers (DPMM) is
proposed to solve constrained convex optimization problems over multi-agent
networks, where the local objective of each agent is a general closed convex
function, and the constraints are coupled equalities and inequalities. This
algorithm strategically integrates the dual decomposition method and the
proximal point algorithm. One advantage of DPMM is that subproblems can be
solved inexactly and in parallel by agents at each iteration, which relaxes the
restriction of requiring exact solutions to subproblems in many distributed
constrained optimization algorithms. We show that the first-order optimality
residual of the proposed algorithm decays to at a rate of under
general convexity. Furthermore, if a structural assumption for the considered
optimization problem is satisfied, the sequence generated by DPMM converges
linearly to an optimal solution. In numerical simulations, we compare DPMM with
several existing algorithms using two examples to demonstrate its
effectiveness
Distributed Spatiotemporal Motion Planning for Spacecraft Swarms in Cluttered Environments
This paper focuses on trajectory planning for spacecraft swarms in cluttered environments, like debris fields or the asteroid belt. Our objective is to reconfigure the spacecraft swarm to a desired formation in a distributed manner while minimizing fuel and avoiding collisions among themselves and with the obstacles. In our prior work we proposed a novel distributed guidance algorithm for spacecraft swarms in static environments. In this paper, we present the Multi-Agent Moving-Obstacles Spherical Expansion and Sequential Convex Programming (MAMO SE-SCP) algorithm that extends our prior work to include spatiotemporal constraints such as time-varying, moving obstacles and desired time-varying terminal positions. In the MAMO SE-SCP algorithm, each agent uses a spherical-expansion-based sampling algorithm to cooperatively explore the time-varying environment, a distributed assignment algorithm to agree on the terminal position for each agent, and a sequential-convex-programming-based optimization step to compute the locally-optimal trajectories from the current location to the assigned time-varying terminal
position while avoiding collision with other agent and the moving obstacles. Simulations results demonstrate that the proposed distributed algorithm can be used by a spacecraft
swarm to achieve a time-varying, desired formation around an object of interest in a dynamic environment with many moving and tumbling obstacles
Fast, Distributed Optimization Strategies for Resource Allocation in Networks
Many challenges in network science and engineering today arise from systems composed of many individual agents interacting over a network. Such problems range from humans interacting with each other in social networks to computers processing and exchanging information over wired or wireless networks. In any application where information is spread out spatially, solutions must address information aggregation in addition to the decision process itself. Intelligently addressing the trade off between information aggregation and decision accuracy is fundamental to finding solutions quickly and accurately. Network optimization challenges such as these have generated a lot of interest in distributed optimization methods. The field of distributed optimization deals with iterative methods which perform calculations using locally available information. Early methods such as subgradient descent suffer very slow convergence rates because the underlying optimization method is a first order method. My work addresses problems in the area of network optimization and control with an emphasis on accelerating the rate of convergence by using a faster underlying optimization method. In the case of convex network flow optimization, the problem is transformed to the dual domain, moving the equality constraints which guarantee flow conservation into the objective. The Newton direction can be computed locally by using a consensus iteration to solve a Poisson equation, but this requires a lot of communication between neighboring nodes. Accelerated Dual Descent (ADD) is an approximate Newton method, which significantly reduces the communication requirement. Defining a stochastic version of the convex network flow problem with edge capacities yields a problem equivalent to the queue stability problem studied in the backpressure literature. Accelerated Backpressure (ABP) is developed to solve the queue stabilization problem. A queue reduction method is introduced by merging ideas from integral control and momentum based optimization
Distributed averaging over communication networks:Fragility, robustness and opportunities
Distributed averaging, a canonical operation among many natural interconnected systems, has found applications in
a tremendous variety of applied fields, including statistical physics, signal processing, systems and control, communication and social science. As information exchange is a central part of distributed averaging systems, it is of practical as well as theoretical importance to understand various properties/limitations of those systems in
the presence of communication constraints and devise new algorithms
to alleviate those limitations.
We study the fragility of a popular distributed averaging algorithm
when the information exchange among the nodes is limited
by communication delays, fading connections and additive noise.
We show that the otherwise well studied and benign
multi-agent system can generate a collective global complex behavior.
We characterize this behavior, common to many natural and human-made interconnected systems, as a collective hyper-jump diffusion process and as a L\\u27{e}vy flights process in a special case. We further describe the mechanism for its emergence and predict its occurrence, under standard assumptions, by checking the Mean Square instability of a certain part of the system. We show that the strong connectivity property of the network topology guarantees that the complex behavior is global and manifested by all
the agents in the network, even though the source of uncertainty is localized. We provide novel computational analysis of the MS stability index under spatially invariant structures and gain certain qualitative as well as quantitative insights of the system.
We then focus on design of agents\u27 dynamics to increase the robustness of distributed averaging system to topology variations. We provide a general structure of distributed averaging systems where individual agents are modeled by LTI systems. We show the problem of designing agents\u27 dynamics for distributed averaging is equivalent to an minimization problem. In this way, we could use tools from robust control theory to build the distributed averaging system where the design is fully distributed and scalable with the size of the network. It is also shown that the agents could be used in different fixed networks and networks with speical time varying interconnections.
We develop new iterative distributed averaging algorithms which allow
agents to compute the average quantity in the presence of
additive noise and random changing interconnections.
The algorithm relaxes several previous restrictive
assumptions on distributed averaging under
uncertainties, such as diminishing step size rule, doubly
stochastic weights, symmetric link switching styles, etc, and
introduces novel mechanism of network feedback to mitigate effects
of communication uncertainties on information aggregation.
Based on the robust distributed averaging algorithm, we propose continuous as well as discrete time computation models
to solve the distributed optimization problem where the objective function is formed by the summation of convex functions of the same variable.
The algorithm shows faster convergence speed than existing ones and
exhibits robustness to additive noise, which is the main source of limitation on algorithms based on convex mixing.
It is shown that agents with simple dynamics and gradient sensing abilities could collectively solve complicated convex optimization problems. Finally, we generalize this algorithm to build a general framework forconstrained convex optimization problems. This framework is shown to be particularly effective to derive solutions for distributed decision making problems and lead to a systems perspective for convex optimization
- …