7,874 research outputs found
Reinforcement Learning Scheduler for Vehicle-to-Vehicle Communications Outside Coverage
Radio resources in vehicle-to-vehicle (V2V) communication can be scheduled
either by a centralized scheduler residing in the network (e.g., a base station
in case of cellular systems) or a distributed scheduler, where the resources
are autonomously selected by the vehicles. The former approach yields a
considerably higher resource utilization in case the network coverage is
uninterrupted. However, in case of intermittent or out-of-coverage, due to not
having input from centralized scheduler, vehicles need to revert to distributed
scheduling. Motivated by recent advances in reinforcement learning (RL), we
investigate whether a centralized learning scheduler can be taught to
efficiently pre-assign the resources to vehicles for out-of-coverage V2V
communication. Specifically, we use the actor-critic RL algorithm to train the
centralized scheduler to provide non-interfering resources to vehicles before
they enter the out-of-coverage area. Our initial results show that a RL-based
scheduler can achieve performance as good as or better than the state-of-art
distributed scheduler, often outperforming it. Furthermore, the learning
process completes within a reasonable time (ranging from a few hundred to a few
thousand epochs), thus making the RL-based scheduler a promising solution for
V2V communications with intermittent network coverage.Comment: Article published in IEEE VNC 201
Learning Scheduling Algorithms for Data Processing Clusters
Efficiently scheduling data processing jobs on distributed compute clusters
requires complex algorithms. Current systems, however, use simple generalized
heuristics and ignore workload characteristics, since developing and tuning a
scheduling policy for each workload is infeasible. In this paper, we show that
modern machine learning techniques can generate highly-efficient policies
automatically. Decima uses reinforcement learning (RL) and neural networks to
learn workload-specific scheduling algorithms without any human instruction
beyond a high-level objective such as minimizing average job completion time.
Off-the-shelf RL techniques, however, cannot handle the complexity and scale of
the scheduling problem. To build Decima, we had to develop new representations
for jobs' dependency graphs, design scalable RL models, and invent RL training
methods for dealing with continuous stochastic job arrivals. Our prototype
integration with Spark on a 25-node cluster shows that Decima improves the
average job completion time over hand-tuned scheduling heuristics by at least
21%, achieving up to 2x improvement during periods of high cluster load
Petuum: A New Platform for Distributed Machine Learning on Big Data
What is a systematic way to efficiently apply a wide spectrum of advanced ML
programs to industrial scale problems, using Big Models (up to 100s of billions
of parameters) on Big Data (up to terabytes or petabytes)? Modern
parallelization strategies employ fine-grained operations and scheduling beyond
the classic bulk-synchronous processing paradigm popularized by MapReduce, or
even specialized graph-based execution that relies on graph representations of
ML programs. The variety of approaches tends to pull systems and algorithms
design in different directions, and it remains difficult to find a universal
platform applicable to a wide range of ML programs at scale. We propose a
general-purpose framework that systematically addresses data- and
model-parallel challenges in large-scale ML, by observing that many ML programs
are fundamentally optimization-centric and admit error-tolerant,
iterative-convergent algorithmic solutions. This presents unique opportunities
for an integrative system design, such as bounded-error network synchronization
and dynamic scheduling based on ML program structure. We demonstrate the
efficacy of these system designs versus well-known implementations of modern ML
algorithms, allowing ML programs to run in much less time and at considerably
larger model sizes, even on modestly-sized compute clusters.Comment: 15 pages, 10 figures, final version in KDD 2015 under the same titl
- …