11 research outputs found
On Convergence of Heuristics Based on Douglas-Rachford Splitting and ADMM to Minimize Convex Functions over Nonconvex Sets
Recently, heuristics based on the Douglas-Rachford splitting algorithm and
the alternating direction method of multipliers (ADMM) have found empirical
success in minimizing convex functions over nonconvex sets, but not much has
been done to improve the theoretical understanding of them. In this paper, we
investigate convergence of these heuristics. First, we characterize optimal
solutions of minimization problems involving convex cost functions over
nonconvex constraint sets. We show that these optimal solutions are related to
the fixed point set of the underlying nonconvex Douglas-Rachford operator.
Next, we establish sufficient conditions under which the Douglas-Rachford
splitting heuristic either converges to a point or its cluster points form a
nonempty compact connected set. In the case where the heuristic converges to a
point, we establish sufficient conditions for that point to be an optimal
solution. Then, we discuss how the ADMM heuristic can be constructed from the
Douglas-Rachford splitting algorithm. We show that, unlike in the convex case,
the algorithms in our nonconvex setup are not equivalent to each other and have
a rather involved relationship between them. Finally, we comment on convergence
of the ADMM heuristic and compare it with the Douglas-Rachford splitting
heuristic.Comment: 11 page
Exterior-point Optimization for Nonconvex Learning
In this paper we present the nonconvex exterior-point optimization solver
(NExOS) -- a novel first-order algorithm tailored to constrained nonconvex
learning problems. We consider the problem of minimizing a convex function over
nonconvex constraints, where the projection onto the constraint set is
single-valued around local minima. A wide range of nonconvex learning problems
have this structure including (but not limited to) sparse and low-rank
optimization problems. By exploiting the underlying geometry of the constraint
set, NExOS finds a locally optimal point by solving a sequence of penalized
problems with strictly decreasing penalty parameters. NExOS solves each
penalized problem by applying a first-order algorithm, which converges linearly
to a local minimum of the corresponding penalized formulation under regularity
conditions. Furthermore, the local minima of the penalized problems converge to
a local minimum of the original problem as the penalty parameter goes to zero.
We implement NExOS in the open-source Julia package NExOS.jl, which has been
extensively tested on many instances from a wide variety of learning problems.
We demonstrate that our algorithm, in spite of being general purpose,
outperforms specialized methods on several examples of well-known nonconvex
learning problems involving sparse and low-rank optimization. For sparse
regression problems, NExOS finds locally optimal solutions which dominate
glmnet in terms of support recovery, yet its training loss is smaller by an
order of magnitude. For low-rank optimization with real-world data, NExOS
recovers solutions with 3 fold training loss reduction, but with a proportion
of explained variance that is 2 times better compared to the nuclear norm
heuristic.Comment: 40 pages, 6 figure
Nonlinear conjugate gradient methods: worst-case convergence rates via computer-assisted analyses
We propose a computer-assisted approach to the analysis of the worst-case
convergence of nonlinear conjugate gradient methods (NCGMs). Those methods are
known for their generally good empirical performances for large-scale
optimization, while having relatively incomplete analyses. Using our
computer-assisted approach, we establish novel complexity bounds for the
Polak-Ribi\`ere-Polyak (PRP) and the Fletcher-Reeves (FR) NCGMs for smooth
strongly convex minimization. Conversely, we provide examples showing that
those methods might behave worse than the regular steepest descent on the same
class of problems.Comment: 29 pages, 6 figures, 4 table
Energy-optimal Timetable Design for Sustainable Metro Railway Networks
We present our collaboration with Thales Canada Inc, the largest provider of
communication-based train control (CBTC) systems worldwide. We study the
problem of designing energy-optimal timetables in metro railway networks to
minimize the effective energy consumption of the network, which corresponds to
simultaneously minimizing total energy consumed by all the trains and
maximizing the transfer of regenerative braking energy from suitable braking
trains to accelerating trains. We propose a novel data-driven linear
programming model that minimizes the total effective energy consumption in a
metro railway network, capable of computing the optimal timetable in real-time,
even for some of the largest CBTC systems in the world. In contrast with
existing works, which are either NP-hard or involve multiple stages requiring
extensive simulation, our model is a single linear programming model capable of
computing the energy-optimal timetable subject to the constraints present in
the railway network. Furthermore, our model can predict the total energy
consumption of the network without requiring time-consuming simulations, making
it suitable for widespread use in managerial settings. We apply our model to
Shanghai Railway Network's Metro Line 8 -- one of the largest and busiest
railway services in the world -- and empirically demonstrate that our model
computes energy-optimal timetables for thousands of active trains spanning an
entire service period of one day in real-time (solution time less than one
second on a standard desktop), achieving energy savings between approximately
20.93% and 28.68%. Given the compelling advantages, our model is in the process
of being integrated into Thales Canada Inc's industrial timetable compiler.Comment: 28 pages, 8 figures, 2 table
Optimization Models for Energy-efficient Railway Timetables
This thesis presents two novel optimization models to calculate energy-efficient railway timetables in a railway network. The first optimization model is a mixed integer programming one, which saves energy by maximizing the total overlapping time between the braking and accelerating phases of suitable train pairs. However, it suffers from some limitations associated with NP-hard computational complexity and modeling of energy saving strategy. To overcome the limitations of the first model, we propose a second optimization model consisting of two stages. The first stage of this model minimizes the total energy consumed by all trains and the second stage maximizes the transfer of regenerative braking energy between suitable train pairs. Both of these stages are solvable in polynomial time, compared to other existing models, which are NP-hard. The two-stage model has proven to be very effective in practice and has been incorporated into an industrial railway timetable compiler.M.A.S
A two-step linear programming model for energy-efficient timetables in metro railway networks
In this paper we propose a novel two-step linear optimization model to calculate energy-efficient timetables in metro railway networks. The resultant timetable minimizes the total energy consumed by all trains and maximizes the utilization of regenerative energy produced by braking trains, subject to the constraints in the railway network. In contrast to other existing models, which are NP-hard, our model is computationally the most tractable one being a linear program. We apply our optimization model to different instances of service PES2-SFM2 of line 8 of Shanghai Metro network spanning a full service period of one day (18 h) with thousands of active trains. For every instance, our model finds an optimal timetable very quickly (largest runtime being less than 13 s) with significant reduction in effective energy consumption (the worst case being 19.27%). Code based on the model has been integrated with Thales Timetable Compiler - the industrial timetable compiler of Thales Inc that has the largest installed base of communication-based train control systems worldwide
Branch-and-Bound Performance Estimation Programming: A Unified Methodology for Constructing Optimal Optimization Methods
We present the Branch-and-Bound Performance Estimation Programming (BnB-PEP),
a unified methodology for constructing optimal first-order methods for convex
and nonconvex optimization. BnB-PEP poses the problem of finding the optimal
optimization method as a nonconvex but practically tractable quadratically
constrained quadratic optimization problem and solves it to certifiable global
optimality using a customized branch-and-bound algorithm. By directly
confronting the nonconvexity, BnB-PEP offers significantly more flexibility and
removes the many limitations of the prior methodologies. Our customized
branch-and-bound algorithm, through exploiting specific problem structures,
outperforms the latest off-the-shelf implementations by orders of magnitude,
accelerating the solution time from hours to seconds and weeks to minutes. We
apply BnB-PEP to several setups for which the prior methodologies do not apply
and obtain methods with bounds that improve upon prior state-of-the-art
results. Finally, we use the BnB-PEP methodology to find proofs with potential
function structures, thereby systematically generating analytical convergence
proofs.Comment: 65 pages, 7 figures, 17 table