60 research outputs found
Algorithms and Adaptivity Gaps for Stochastic k-TSP
Given a metric and a , the classic
\textsf{k-TSP} problem is to find a tour originating at the
of minimum length that visits at least nodes in . In this work,
motivated by applications where the input to an optimization problem is
uncertain, we study two stochastic versions of \textsf{k-TSP}.
In Stoch-Reward -TSP, originally defined by Ene-Nagarajan-Saket [ENS17],
each vertex in the given metric contains a stochastic reward .
The goal is to adaptively find a tour of minimum expected length that collects
at least reward ; here "adaptively" means our next decision may depend on
previous outcomes. Ene et al. give an -approximation adaptive
algorithm for this problem, and left open if there is an -approximation
algorithm. We totally resolve their open question and even give an
-approximation \emph{non-adaptive} algorithm for this problem.
We also introduce and obtain similar results for the Stoch-Cost -TSP
problem. In this problem each vertex has a stochastic cost , and the
goal is to visit and select at least vertices to minimize the expected
\emph{sum} of tour length and cost of selected vertices. This problem
generalizes the Price of Information framework [Singla18] from deterministic
probing costs to metric probing costs.
Our techniques are based on two crucial ideas: "repetitions" and "critical
scaling". We show using Freedman's and Jogdeo-Samuels' inequalities that for
our problems, if we truncate the random variables at an ideal threshold and
repeat, then their expected values form a good surrogate. Unfortunately, this
ideal threshold is adaptive as it depends on how far we are from achieving our
target , so we truncate at various different scales and identify a
"critical" scale.Comment: ITCS 202
Approximation Algorithms for Stochastic k-TSP
This paper studies the stochastic variant of the classical k-TSP problem where rewards at the vertices are independent random variables which are instantiated upon the tour\u27s visit. The objective is to minimize the expected length of a tour that collects reward at least k. The solution is a policy describing the tour which may (adaptive) or may not (non-adaptive) depend on the observed rewards.
Our work presents an adaptive O(log k)-approximation algorithm for Stochastic k-TSP, along with a non-adaptive O(log^2 k)-approximation algorithm which also upper bounds the adaptivity gap by O(log^2 k). We also show that the adaptivity gap of Stochastic k-TSP is at least e, even in the special case of stochastic knapsack cover
(Near) Optimal Adaptivity Gaps for Stochastic Multi-Value Probing
Consider a kidney-exchange application where we want to find a max-matching in a random graph. To find whether an edge e exists, we need to perform an expensive test, in which case the edge e appears independently with a known probability p_e. Given a budget on the total cost of the tests, our goal is to find a testing strategy that maximizes the expected maximum matching size.
The above application is an example of the stochastic probing problem. In general the optimal stochastic probing strategy is difficult to find because it is adaptive - decides on the next edge to probe based on the outcomes of the probed edges. An alternate approach is to show the adaptivity gap is small, i.e., the best non-adaptive strategy always has a value close to the best adaptive strategy. This allows us to focus on designing non-adaptive strategies that are much simpler. Previous works, however, have focused on Bernoulli random variables that can only capture whether an edge appears or not. In this work we introduce a multi-value stochastic probing problem, which can also model situations where the weight of an edge has a probability distribution over multiple values.
Our main technical contribution is to obtain (near) optimal bounds for the (worst-case) adaptivity gaps for multi-value stochastic probing over prefix-closed constraints. For a monotone submodular function, we show the adaptivity gap is at most 2 and provide a matching lower bound. For a weighted rank function of a k-extendible system (a generalization of intersection of k matroids), we show the adaptivity gap is between O(k log k) and k. None of these results were known even in the Bernoulli case where both our upper and lower bounds also apply, thereby resolving an open question of Gupta et al. [Gupta et al., 2017]
The Price of Information in Combinatorial Optimization
Consider a network design application where we wish to lay down a
minimum-cost spanning tree in a given graph; however, we only have stochastic
information about the edge costs. To learn the precise cost of any edge, we
have to conduct a study that incurs a price. Our goal is to find a spanning
tree while minimizing the disutility, which is the sum of the tree cost and the
total price that we spend on the studies. In a different application, each edge
gives a stochastic reward value. Our goal is to find a spanning tree while
maximizing the utility, which is the tree reward minus the prices that we pay.
Situations such as the above two often arise in practice where we wish to
find a good solution to an optimization problem, but we start with only some
partial knowledge about the parameters of the problem. The missing information
can be found only after paying a probing price, which we call the price of
information. What strategy should we adopt to optimize our expected
utility/disutility?
A classical example of the above setting is Weitzman's "Pandora's box"
problem where we are given probability distributions on values of
independent random variables. The goal is to choose a single variable with a
large value, but we can find the actual outcomes only after paying a price. Our
work is a generalization of this model to other combinatorial optimization
problems such as matching, set cover, facility location, and prize-collecting
Steiner tree. We give a technique that reduces such problems to their non-price
counterparts, and use it to design exact/approximation algorithms to optimize
our utility/disutility. Our techniques extend to situations where there are
additional constraints on what parameters can be probed or when we can
simultaneously probe a subset of the parameters.Comment: SODA 201
Locally Adaptive Optimization: Adaptive Seeding for Monotone Submodular Functions
The Adaptive Seeding problem is an algorithmic challenge motivated by
influence maximization in social networks: One seeks to select among certain
accessible nodes in a network, and then select, adaptively, among neighbors of
those nodes as they become accessible in order to maximize a global objective
function. More generally, adaptive seeding is a stochastic optimization
framework where the choices in the first stage affect the realizations in the
second stage, over which we aim to optimize.
Our main result is a -approximation for the adaptive seeding
problem for any monotone submodular function. While adaptive policies are often
approximated via non-adaptive policies, our algorithm is based on a novel
method we call \emph{locally-adaptive} policies. These policies combine a
non-adaptive global structure, with local adaptive optimizations. This method
enables the -approximation for general monotone submodular functions
and circumvents some of the impossibilities associated with non-adaptive
policies.
We also introduce a fundamental problem in submodular optimization that may
be of independent interest: given a ground set of elements where every element
appears with some small probability, find a set of expected size at most
that has the highest expected value over the realization of the elements. We
show a surprising result: there are classes of monotone submodular functions
(including coverage) that can be approximated almost optimally as the
probability vanishes. For general monotone submodular functions we show via a
reduction from \textsc{Planted-Clique} that approximations for this problem are
not likely to be obtainable. This optimization problem is an important tool for
adaptive seeding via non-adaptive policies, and its hardness motivates the
introduction of \emph{locally-adaptive} policies we use in the main result
Attention, Learn to Solve Routing Problems!
The recently presented idea to learn heuristics for combinatorial
optimization problems is promising as it can save costly development. However,
to push this idea towards practical implementation, we need better models and
better ways of training. We contribute in both directions: we propose a model
based on attention layers with benefits over the Pointer Network and we show
how to train this model using REINFORCE with a simple baseline based on a
deterministic greedy rollout, which we find is more efficient than using a
value function. We significantly improve over recent learned heuristics for the
Travelling Salesman Problem (TSP), getting close to optimal results for
problems up to 100 nodes. With the same hyperparameters, we learn strong
heuristics for two variants of the Vehicle Routing Problem (VRP), the
Orienteering Problem (OP) and (a stochastic variant of) the Prize Collecting
TSP (PCTSP), outperforming a wide range of baselines and getting results close
to highly optimized and specialized algorithms.Comment: Accepted at ICLR 2019. 25 pages, 7 figure
Minimum Latency Submodular Cover
We study the Minimum Latency Submodular Cover problem (MLSC), which consists
of a metric with source and monotone submodular functions
. The goal is to find a path
originating at that minimizes the total cover time of all functions. This
generalizes well-studied problems, such as Submodular Ranking [AzarG11] and
Group Steiner Tree [GKR00]. We give a polynomial time O(\log \frac{1}{\eps}
\cdot \log^{2+\delta} |V|)-approximation algorithm for MLSC, where
is the smallest non-zero marginal increase of any
and is any constant.
We also consider the Latency Covering Steiner Tree problem (LCST), which is
the special case of \mlsc where the s are multi-coverage functions. This
is a common generalization of the Latency Group Steiner Tree
[GuptaNR10a,ChakrabartyS11] and Generalized Min-sum Set Cover [AzarGY09,
BansalGK10] problems. We obtain an -approximation algorithm for
LCST.
Finally we study a natural stochastic extension of the Submodular Ranking
problem, and obtain an adaptive algorithm with an O(\log 1/ \eps)
approximation ratio, which is best possible. This result also generalizes some
previously studied stochastic optimization problems, such as Stochastic Set
Cover [GoemansV06] and Shared Filter Evaluation [MunagalaSW07, LiuPRY08].Comment: 23 pages, 1 figur
- …