4,942 research outputs found
Approximate Convex Optimization by Online Game Playing
Lagrangian relaxation and approximate optimization algorithms have received
much attention in the last two decades. Typically, the running time of these
methods to obtain a approximate solution is proportional to
. Recently, Bienstock and Iyengar, following Nesterov,
gave an algorithm for fractional packing linear programs which runs in
iterations. The latter algorithm requires to solve a
convex quadratic program every iteration - an optimization subroutine which
dominates the theoretical running time.
We give an algorithm for convex programs with strictly convex constraints
which runs in time proportional to . The algorithm does NOT
require to solve any quadratic program, but uses gradient steps and elementary
operations only. Problems which have strictly convex constraints include
maximum entropy frequency estimation, portfolio optimization with loss risk
constraints, and various computational problems in signal processing.
As a side product, we also obtain a simpler version of Bienstock and
Iyengar's result for general linear programming, with similar running time.
We derive these algorithms using a new framework for deriving convex
optimization algorithms from online game playing algorithms, which may be of
independent interest
Convex Relaxations for Gas Expansion Planning
Expansion of natural gas networks is a critical process involving substantial
capital expenditures with complex decision-support requirements. Given the
non-convex nature of gas transmission constraints, global optimality and
infeasibility guarantees can only be offered by global optimisation approaches.
Unfortunately, state-of-the-art global optimisation solvers are unable to scale
up to real-world size instances. In this study, we present a convex
mixed-integer second-order cone relaxation for the gas expansion planning
problem under steady-state conditions. The underlying model offers tight lower
bounds with high computational efficiency. In addition, the optimal solution of
the relaxation can often be used to derive high-quality solutions to the
original problem, leading to provably tight optimality gaps and, in some cases,
global optimal soluutions. The convex relaxation is based on a few key ideas,
including the introduction of flux direction variables, exact McCormick
relaxations, on/off constraints, and integer cuts. Numerical experiments are
conducted on the traditional Belgian gas network, as well as other real larger
networks. The results demonstrate both the accuracy and computational speed of
the relaxation and its ability to produce high-quality solutions
A Distributed Approach for the Optimal Power Flow Problem Based on ADMM and Sequential Convex Approximations
The optimal power flow (OPF) problem, which plays a central role in operating
electrical networks is considered. The problem is nonconvex and is in fact NP
hard. Therefore, designing efficient algorithms of practical relevance is
crucial, though their global optimality is not guaranteed. Existing
semi-definite programming relaxation based approaches are restricted to OPF
problems where zero duality holds. In this paper, an efficient novel method to
address the general nonconvex OPF problem is investigated. The proposed method
is based on alternating direction method of multipliers combined with
sequential convex approximations. The global OPF problem is decomposed into
smaller problems associated to each bus of the network, the solutions of which
are coordinated via a light communication protocol. Therefore, the proposed
method is highly scalable. The convergence properties of the proposed algorithm
are mathematically substantiated. Finally, the proposed algorithm is evaluated
on a number of test examples, where the convergence properties of the proposed
algorithm are numerically substantiated and the performance is compared with a
global optimal method.Comment: 14 page
A Unified Successive Pseudo-Convex Approximation Framework
In this paper, we propose a successive pseudo-convex approximation algorithm
to efficiently compute stationary points for a large class of possibly
nonconvex optimization problems. The stationary points are obtained by solving
a sequence of successively refined approximate problems, each of which is much
easier to solve than the original problem. To achieve convergence, the
approximate problem only needs to exhibit a weak form of convexity, namely,
pseudo-convexity. We show that the proposed framework not only includes as
special cases a number of existing methods, for example, the gradient method
and the Jacobi algorithm, but also leads to new algorithms which enjoy easier
implementation and faster convergence speed. We also propose a novel line
search method for nondifferentiable optimization problems, which is carried out
over a properly constructed differentiable function with the benefit of a
simplified implementation as compared to state-of-the-art line search
techniques that directly operate on the original nondifferentiable objective
function. The advantages of the proposed algorithm are shown, both
theoretically and numerically, by several example applications, namely, MIMO
broadcast channel capacity computation, energy efficiency maximization in
massive MIMO systems and LASSO in sparse signal recovery.Comment: submitted to IEEE Transactions on Signal Processing; original title:
A Novel Iterative Convex Approximation Metho
- …