43,976 research outputs found
Approximation Algorithms for Distributionally Robust Stochastic Optimization with Black-Box Distributions
Two-stage stochastic optimization is a framework for modeling uncertainty,
where we have a probability distribution over possible realizations of the
data, called scenarios, and decisions are taken in two stages: we make
first-stage decisions knowing only the underlying distribution and before a
scenario is realized, and may take additional second-stage recourse actions
after a scenario is realized. The goal is typically to minimize the total
expected cost. A criticism of this model is that the underlying probability
distribution is itself often imprecise! To address this, a versatile approach
that has been proposed is the {\em distributionally robust 2-stage model}:
given a collection of probability distributions, our goal now is to minimize
the maximum expected total cost with respect to a distribution in this
collection.
We provide a framework for designing approximation algorithms in such
settings when the collection is a ball around a central distribution and the
central distribution is accessed {\em only via a sampling black box}.
We first show that one can utilize the {\em sample average approximation}
(SAA) method to reduce the problem to the case where the central distribution
has {\em polynomial-size} support. We then show how to approximately solve a
fractional relaxation of the SAA (i.e., polynomial-scenario
central-distribution) problem. By complementing this via LP-rounding algorithms
that provide {\em local} (i.e., per-scenario) approximation guarantees, we
obtain the {\em first} approximation algorithms for the distributionally robust
versions of a variety of discrete-optimization problems including set cover,
vertex cover, edge cover, facility location, and Steiner tree, with guarantees
that are, except for set cover, within -factors of the guarantees known
for the deterministic version of the problem
Two Timescale Convergent Q-learning for Sleep--Scheduling in Wireless Sensor Networks
In this paper, we consider an intrusion detection application for Wireless
Sensor Networks (WSNs). We study the problem of scheduling the sleep times of
the individual sensors to maximize the network lifetime while keeping the
tracking error to a minimum. We formulate this problem as a
partially-observable Markov decision process (POMDP) with continuous
state-action spaces, in a manner similar to (Fuemmeler and Veeravalli [2008]).
However, unlike their formulation, we consider infinite horizon discounted and
average cost objectives as performance criteria. For each criterion, we propose
a convergent on-policy Q-learning algorithm that operates on two timescales,
while employing function approximation to handle the curse of dimensionality
associated with the underlying POMDP. Our proposed algorithm incorporates a
policy gradient update using a one-simulation simultaneous perturbation
stochastic approximation (SPSA) estimate on the faster timescale, while the
Q-value parameter (arising from a linear function approximation for the
Q-values) is updated in an on-policy temporal difference (TD) algorithm-like
fashion on the slower timescale. The feature selection scheme employed in each
of our algorithms manages the energy and tracking components in a manner that
assists the search for the optimal sleep-scheduling policy. For the sake of
comparison, in both discounted and average settings, we also develop a function
approximation analogue of the Q-learning algorithm. This algorithm, unlike the
two-timescale variant, does not possess theoretical convergence guarantees.
Finally, we also adapt our algorithms to include a stochastic iterative
estimation scheme for the intruder's mobility model. Our simulation results on
a 2-dimensional network setting suggest that our algorithms result in better
tracking accuracy at the cost of only a few additional sensors, in comparison
to a recent prior work
A Multistage Stochastic Programming Approach to the Dynamic and Stochastic VRPTW - Extended version
We consider a dynamic vehicle routing problem with time windows and
stochastic customers (DS-VRPTW), such that customers may request for services
as vehicles have already started their tours. To solve this problem, the goal
is to provide a decision rule for choosing, at each time step, the next action
to perform in light of known requests and probabilistic knowledge on requests
likelihood. We introduce a new decision rule, called Global Stochastic
Assessment (GSA) rule for the DS-VRPTW, and we compare it with existing
decision rules, such as MSA. In particular, we show that GSA fully integrates
nonanticipativity constraints so that it leads to better decisions in our
stochastic context. We describe a new heuristic approach for efficiently
approximating our GSA rule. We introduce a new waiting strategy. Experiments on
dynamic and stochastic benchmarks, which include instances of different degrees
of dynamism, show that not only our approach is competitive with
state-of-the-art methods, but also enables to compute meaningful offline
solutions to fully dynamic problems where absolutely no a priori customer
request is provided.Comment: Extended version of the same-name study submitted for publication in
conference CPAIOR201
Approximation algorithms for stochastic and risk-averse optimization
We present improved approximation algorithms in stochastic optimization. We
prove that the multi-stage stochastic versions of covering integer programs
(such as set cover and vertex cover) admit essentially the same approximation
algorithms as their standard (non-stochastic) counterparts; this improves upon
work of Swamy \& Shmoys which shows an approximability that depends
multiplicatively on the number of stages. We also present approximation
algorithms for facility location and some of its variants in the -stage
recourse model, improving on previous approximation guarantees. We give a
-approximation algorithm in the standard polynomial-scenario model and
an algorithm with an expected per-scenario -approximation guarantee,
which is applicable to the more general black-box distribution model.Comment: Extension of a SODA'07 paper. To appear in SIAM J. Discrete Mat
Optimal treatment allocations in space and time for on-line control of an emerging infectious disease
A key component in controlling the spread of an epidemic is deciding where, whenand to whom to apply an intervention.We develop a framework for using data to informthese decisionsin realtime.We formalize a treatment allocation strategy as a sequence of functions, oneper treatment period, that map up-to-date information on the spread of an infectious diseaseto a subset of locations where treatment should be allocated. An optimal allocation strategyoptimizes some cumulative outcome, e.g. the number of uninfected locations, the geographicfootprint of the disease or the cost of the epidemic. Estimation of an optimal allocation strategyfor an emerging infectious disease is challenging because spatial proximity induces interferencebetween locations, the number of possible allocations is exponential in the number oflocations, and because disease dynamics and intervention effectiveness are unknown at outbreak.We derive a Bayesian on-line estimator of the optimal allocation strategy that combinessimulation–optimization with Thompson sampling.The estimator proposed performs favourablyin simulation experiments. This work is motivated by and illustrated using data on the spread ofwhite nose syndrome, which is a highly fatal infectious disease devastating bat populations inNorth America
- …