3,863 research outputs found
A note on Probably Certifiably Correct algorithms
Many optimization problems of interest are known to be intractable, and while
there are often heuristics that are known to work on typical instances, it is
usually not easy to determine a posteriori whether the optimal solution was
found. In this short note, we discuss algorithms that not only solve the
problem on typical instances, but also provide a posteriori certificates of
optimality, probably certifiably correct (PCC) algorithms. As an illustrative
example, we present a fast PCC algorithm for minimum bisection under the
stochastic block model and briefly discuss other examples
Submodular Optimization with Submodular Cover and Submodular Knapsack Constraints
We investigate two new optimization problems -- minimizing a submodular
function subject to a submodular lower bound constraint (submodular cover) and
maximizing a submodular function subject to a submodular upper bound constraint
(submodular knapsack). We are motivated by a number of real-world applications
in machine learning including sensor placement and data subset selection, which
require maximizing a certain submodular function (like coverage or diversity)
while simultaneously minimizing another (like cooperative cost). These problems
are often posed as minimizing the difference between submodular functions [14,
35] which is in the worst case inapproximable. We show, however, that by
phrasing these problems as constrained optimization, which is more natural for
many applications, we achieve a number of bounded approximation guarantees. We
also show that both these problems are closely related and an approximation
algorithm solving one can be used to obtain an approximation guarantee for the
other. We provide hardness results for both problems thus showing that our
approximation factors are tight up to log-factors. Finally, we empirically
demonstrate the performance and good scalability properties of our algorithms.Comment: 23 pages. A short version of this appeared in Advances of NIPS-201
SURROGATE SEARCH: A SIMULATION OPTIMIZATION METHODOLOGY FOR LARGE-SCALE SYSTEMS
For certain settings in which system performance cannot be evaluated by analytical methods, simulation models are widely utilized. This is especially for complex systems. To try to optimize these models, simulation optimization techniques have been developed. These attempt to identify the system designs and parameters that result in (near) optimal system performance. Although more realistic results can be provided by simulation, the computational time for simulator execution, and consequently, simulation optimization may be very long. Hence, the major challenge in determining improved system designs by incorporating simulation and search methodologies is to develop more efficient simulation optimization heuristics or algorithms. This dissertation develops a new approach, Surrogate Search, to determine near optimal system designs for large-scale simulation problems that contain combinatorial decision variables. First, surrogate objective functions are identified by analyzing simulation results to observe system behavior. Multiple linear regression is utilized to examine simulation results and construct surrogate objective functions. The identified surrogate objective functions, which can be quickly executed, are then utilized as simulator replacements in the search methodologies. For multiple problems containing different settings of the same simulation model, only one surrogate objective function needs to be identified. The development of surrogate objective functions benefits the optimization process by reducing the number of simulation iterations. Surrogate Search approaches are developed for two combinatorial problems, operator assignment and task sequencing, using a large-scale sortation system simulation model. The experimental results demonstrate that Surrogate Search can be applied to such large-scale simulation problems and outperform recognized simulation optimization methodology, Scatter Search (SS). This dissertation provides a systematic methodology to perform simulation optimization for complex operations research problems and contributes to the simulation optimization field
Recommended from our members
Revisiting lagrange relaxation (LR) for processing large-scale mixed integer programming (MIP) problems
Lagrangean Relaxation has been successfully applied to process many well known
instances of NP-hard Mixed Integer Programming problems. In this paper we present
a Lagrangean Relaxation based generic solver for processing Mixed Integer
Programming problems. We choose the constraints, which are relaxed using a
constraint classification scheme. The tactical issue of updating the Lagrange
multiplier is addressed through sub-gradient optimisation; alternative rules for
updating their values are investigated. The Lagrangean relaxation provides a lower
bound to the original problem and the upper bound is calculated using a heuristic
technique. The bounds obtained by the Lagrangean Relaxation based generic solver
were used to warm-start the Branch and Bound algorithm; the performance of the
generic solver and the effect of the alternative control settings are reported for a wide
class of benchmark models. Finally, we present an alternative technique to calculate
the upper bound, using a genetic algorithm that benefits from the mathematical
structure of the constraints. The performance of the genetic algorithm is also
presented
An Efficient Local Search for Partial Latin Square Extension Problem
A partial Latin square (PLS) is a partial assignment of n symbols to an nxn
grid such that, in each row and in each column, each symbol appears at most
once. The partial Latin square extension problem is an NP-hard problem that
asks for a largest extension of a given PLS. In this paper we propose an
efficient local search for this problem. We focus on the local search such that
the neighborhood is defined by (p,q)-swap, i.e., removing exactly p symbols and
then assigning symbols to at most q empty cells. For p in {1,2,3}, our
neighborhood search algorithm finds an improved solution or concludes that no
such solution exists in O(n^{p+1}) time. We also propose a novel swap
operation, Trellis-swap, which is a generalization of (1,q)-swap and
(2,q)-swap. Our Trellis-neighborhood search algorithm takes O(n^{3.5}) time to
do the same thing. Using these neighborhood search algorithms, we design a
prototype iterated local search algorithm and show its effectiveness in
comparison with state-of-the-art optimization solvers such as IBM ILOG CPLEX
and LocalSolver.Comment: 17 pages, 2 figure
- …