1,062 research outputs found
Optimal web-scale tiering as a flow problem
We present a fast online solver for large scale parametric max-flow problems as they occur in portfolio optimization, inventory management, computer vision, and logistics. Our algorithm solves an integer linear program in an online fashion. It exploits total unimodularity of the constraint matrix and a Lagrangian relaxation to solve the problem as a convex online game. The algorithm generates approximate solutions of max-flow problems by performing stochastic gradient descent on a set of flows. We apply the algorithm to optimize tier arrangement of over 84 million web pages on a layered set of caches to serve an incoming query stream optimally
Submodular relaxation for inference in Markov random fields
In this paper we address the problem of finding the most probable state of a
discrete Markov random field (MRF), also known as the MRF energy minimization
problem. The task is known to be NP-hard in general and its practical
importance motivates numerous approximate algorithms. We propose a submodular
relaxation approach (SMR) based on a Lagrangian relaxation of the initial
problem. Unlike the dual decomposition approach of Komodakis et al., 2011 SMR
does not decompose the graph structure of the initial problem but constructs a
submodular energy that is minimized within the Lagrangian relaxation. Our
approach is applicable to both pairwise and high-order MRFs and allows to take
into account global potentials of certain types. We study theoretical
properties of the proposed approach and evaluate it experimentally.Comment: This paper is accepted for publication in IEEE Transactions on
Pattern Analysis and Machine Intelligenc
A New Dantzig-Wolfe Reformulation And Branch-And-Price Algorithm For The Capacitated Lot Sizing Problem With Set Up Times
The textbook Dantzig-Wolfe decomposition for the Capacitated LotSizing Problem (CLSP),as already proposed by Manne in 1958, has animportant structural deficiency. Imposingintegrality constraints onthe variables in the full blown master will not necessarily givetheoptimal IP solution as only production plans which satisfy theWagner-Whitin condition canbe selected. It is well known that theoptimal solution to a capacitated lot sizing problem willnotnecessarily have this Wagner-Whitin property. The columns of thetraditionaldecomposition model include both the integer set up andcontinuous production quantitydecisions. Choosing a specific set upschedule implies also taking the associated Wagner-Whitin productionquantities. We propose the correct Dantzig-Wolfedecompositionreformulation separating the set up and productiondecisions. This formulation gives the samelower bound as Manne'sreformulation and allows for branch-and-price. We use theCapacitatedLot Sizing Problem with Set Up Times to illustrate our approach.Computationalexperiments are presented on data sets available from theliterature. Column generation isspeeded up by a combination of simplexand subgradient optimization for finding the dualprices. The resultsshow that branch-and-price is computationally tractable andcompetitivewith other approaches. Finally, we briefly discuss how thisnew Dantzig-Wolfe reformulationcan be generalized to other mixedinteger programming problems, whereas in theliterature,branch-and-price algorithms are almost exclusivelydeveloped for pure integer programmingproblems.branch-and-price;Lagrange relaxation;Dantzig-Wolfe decomposition;lot sizing;mixed-integer programming
Large-scale Binary Quadratic Optimization Using Semidefinite Relaxation and Applications
In computer vision, many problems such as image segmentation, pixel
labelling, and scene parsing can be formulated as binary quadratic programs
(BQPs). For submodular problems, cuts based methods can be employed to
efficiently solve large-scale problems. However, general nonsubmodular problems
are significantly more challenging to solve. Finding a solution when the
problem is of large size to be of practical interest, however, typically
requires relaxation. Two standard relaxation methods are widely used for
solving general BQPs--spectral methods and semidefinite programming (SDP), each
with their own advantages and disadvantages. Spectral relaxation is simple and
easy to implement, but its bound is loose. Semidefinite relaxation has a
tighter bound, but its computational complexity is high, especially for large
scale problems. In this work, we present a new SDP formulation for BQPs, with
two desirable properties. First, it has a similar relaxation bound to
conventional SDP formulations. Second, compared with conventional SDP methods,
the new SDP formulation leads to a significantly more efficient and scalable
dual optimization approach, which has the same degree of complexity as spectral
methods. We then propose two solvers, namely, quasi-Newton and smoothing Newton
methods, for the dual problem. Both of them are significantly more efficiently
than standard interior-point methods. In practice, the smoothing Newton solver
is faster than the quasi-Newton solver for dense or medium-sized problems,
while the quasi-Newton solver is preferable for large sparse/structured
problems. Our experiments on a few computer vision applications including
clustering, image segmentation, co-segmentation and registration show the
potential of our SDP formulation for solving large-scale BQPs.Comment: Fixed some typos. 18 pages. Accepted to IEEE Transactions on Pattern
Analysis and Machine Intelligenc
Hierarchy construction schemes within the Scale set framework
Segmentation algorithms based on an energy minimisation framework often
depend on a scale parameter which balances a fit to data and a regularising
term. Irregular pyramids are defined as a stack of graphs successively reduced.
Within this framework, the scale is often defined implicitly as the height in
the pyramid. However, each level of an irregular pyramid can not usually be
readily associated to the global optimum of an energy or a global criterion on
the base level graph. This last drawback is addressed by the scale set
framework designed by Guigues. The methods designed by this author allow to
build a hierarchy and to design cuts within this hierarchy which globally
minimise an energy. This paper studies the influence of the construction scheme
of the initial hierarchy on the resulting optimal cuts. We propose one
sequential and one parallel method with two variations within both. Our
sequential methods provide partitions near the global optima while parallel
methods require less execution times than the sequential method of Guigues even
on sequential machines
Convex and Network Flow Optimization for Structured Sparsity
We consider a class of learning problems regularized by a structured
sparsity-inducing norm defined as the sum of l_2- or l_infinity-norms over
groups of variables. Whereas much effort has been put in developing fast
optimization techniques when the groups are disjoint or embedded in a
hierarchy, we address here the case of general overlapping groups. To this end,
we present two different strategies: On the one hand, we show that the proximal
operator associated with a sum of l_infinity-norms can be computed exactly in
polynomial time by solving a quadratic min-cost flow problem, allowing the use
of accelerated proximal gradient methods. On the other hand, we use proximal
splitting techniques, and address an equivalent formulation with
non-overlapping groups, but in higher dimension and with additional
constraints. We propose efficient and scalable algorithms exploiting these two
strategies, which are significantly faster than alternative approaches. We
illustrate these methods with several problems such as CUR matrix
factorization, multi-task learning of tree-structured dictionaries, background
subtraction in video sequences, image denoising with wavelets, and topographic
dictionary learning of natural image patches.Comment: to appear in the Journal of Machine Learning Research (JMLR
End-to-End Decision Focused Learning using Learned Solvers
Achieving fusion of deep learning with combinatorial algorithms promises transformativechanges to AI. Creating an impact in a real-world setting requires AI techniques to span a pipeline from data, to predictive models, to decisions. Aligning these components together requires careful consideration, as having these components trained separately does not account for the end goal of the model. This work surveys general frameworks for melding these components, we focus on the integration of optimization methods with machine learning architectures. We address some challenges and limitations associated with these methods and propose a novel approach to address some of the bottlenecks that arise
Additive versus multiplicative clause weighting for SAT
This paper examines the relative performance of additive and multiplicative clause weighting schemes for propositional satisfiability testing. Starting with one of the most recently developed multiplicative algorithms (SAPS), an experimental study was constructed to isolate the effects of multiplicative in comparison to additive weighting, while controlling other key features of the two approaches, namely the use of random versus flat moves, deterministic versus probabilistic weight smoothing and multiple versus single inclusion of literals in the local search neighborhood. As a result of this investigation we developed a pure additive weighting scheme (PAWS) which can outperform multiplicative weighting on a range of difficult problems, while requiring considerably less effort in terms of parameter tuning. W
A dual ascent framework for Lagrangean decomposition of combinatorial problems
We propose a general dual ascent framework for Lagrangean decomposition of combinatorial problems. Although methods of this type have shown their efficiency for a number of problems, so far there was no general algorithm applicable to multiple problem types. In this work, we propose such a general algorithm. It depends on several parameters, which can be used to optimize its performance in each particular setting. We demonstrate efficacy of our method on graph matching and multicut problems, where it outperforms state-of-the-art solvers including those based on subgradient optimization and off-the-shelf linear programming solvers
- …