22,964 research outputs found
A Constraint Programming Approach for Non-Preemptive Evacuation Scheduling
Large-scale controlled evacuations require emergency services to select
evacuation routes, decide departure times, and mobilize resources to issue
orders, all under strict time constraints. Existing algorithms almost always
allow for preemptive evacuation schedules, which are less desirable in
practice. This paper proposes, for the first time, a constraint-based
scheduling model that optimizes the evacuation flow rate (number of vehicles
sent at regular time intervals) and evacuation phasing of widely populated
areas, while ensuring a nonpreemptive evacuation for each residential zone. Two
optimization objectives are considered: (1) to maximize the number of evacuees
reaching safety and (2) to minimize the overall duration of the evacuation.
Preliminary results on a set of real-world instances show that the approach can
produce, within a few seconds, a non-preemptive evacuation schedule which is
either optimal or at most 6% away of the optimal preemptive solution.Comment: Submitted to the 21st International Conference on Principles and
Practice of Constraint Programming (CP 2015). 15 pages + 1 reference pag
Two Timescale Convergent Q-learning for Sleep--Scheduling in Wireless Sensor Networks
In this paper, we consider an intrusion detection application for Wireless
Sensor Networks (WSNs). We study the problem of scheduling the sleep times of
the individual sensors to maximize the network lifetime while keeping the
tracking error to a minimum. We formulate this problem as a
partially-observable Markov decision process (POMDP) with continuous
state-action spaces, in a manner similar to (Fuemmeler and Veeravalli [2008]).
However, unlike their formulation, we consider infinite horizon discounted and
average cost objectives as performance criteria. For each criterion, we propose
a convergent on-policy Q-learning algorithm that operates on two timescales,
while employing function approximation to handle the curse of dimensionality
associated with the underlying POMDP. Our proposed algorithm incorporates a
policy gradient update using a one-simulation simultaneous perturbation
stochastic approximation (SPSA) estimate on the faster timescale, while the
Q-value parameter (arising from a linear function approximation for the
Q-values) is updated in an on-policy temporal difference (TD) algorithm-like
fashion on the slower timescale. The feature selection scheme employed in each
of our algorithms manages the energy and tracking components in a manner that
assists the search for the optimal sleep-scheduling policy. For the sake of
comparison, in both discounted and average settings, we also develop a function
approximation analogue of the Q-learning algorithm. This algorithm, unlike the
two-timescale variant, does not possess theoretical convergence guarantees.
Finally, we also adapt our algorithms to include a stochastic iterative
estimation scheme for the intruder's mobility model. Our simulation results on
a 2-dimensional network setting suggest that our algorithms result in better
tracking accuracy at the cost of only a few additional sensors, in comparison
to a recent prior work
Petuum: A New Platform for Distributed Machine Learning on Big Data
What is a systematic way to efficiently apply a wide spectrum of advanced ML
programs to industrial scale problems, using Big Models (up to 100s of billions
of parameters) on Big Data (up to terabytes or petabytes)? Modern
parallelization strategies employ fine-grained operations and scheduling beyond
the classic bulk-synchronous processing paradigm popularized by MapReduce, or
even specialized graph-based execution that relies on graph representations of
ML programs. The variety of approaches tends to pull systems and algorithms
design in different directions, and it remains difficult to find a universal
platform applicable to a wide range of ML programs at scale. We propose a
general-purpose framework that systematically addresses data- and
model-parallel challenges in large-scale ML, by observing that many ML programs
are fundamentally optimization-centric and admit error-tolerant,
iterative-convergent algorithmic solutions. This presents unique opportunities
for an integrative system design, such as bounded-error network synchronization
and dynamic scheduling based on ML program structure. We demonstrate the
efficacy of these system designs versus well-known implementations of modern ML
algorithms, allowing ML programs to run in much less time and at considerably
larger model sizes, even on modestly-sized compute clusters.Comment: 15 pages, 10 figures, final version in KDD 2015 under the same titl
Randomized longest-queue-first scheduling for large-scale buffered systems
We develop diffusion approximations for parallel-queueing systems with the
randomized longest-queue-first scheduling algorithm by establishing new
mean-field limit theorems as the number of buffers . We achieve
this by allowing the number of sampled buffers to depend on the number
of buffers , which yields an asymptotic `decoupling' of the queue length
processes.
We show through simulation experiments that the resulting approximation is
accurate even for moderate values of and . To our knowledge, we are
the first to derive diffusion approximations for a queueing system in the
large-buffer mean-field regime. Another noteworthy feature of our scaling idea
is that the randomized longest-queue-first algorithm emulates the
longest-queue-first algorithm, yet is computationally more attractive. The
analysis of the system performance as a function of is facilitated by
the multi-scale nature in our limit theorems: the various processes we study
have different space scalings. This allows us to show the trade-off between
performance and complexity of the randomized longest-queue-first scheduling
algorithm
- …