11,998 research outputs found
Constraint Centric Scheduling Guide
The advent of architectures with software-exposed resources (Spatial Architectures) has created a demand for universally applicable scheduling techniques. This paper describes our generalized spatial scheduling framework, formulated with Integer Linear Programming, and specifically accomplishes two goals. First, using the ?Simple? architecture, it illustrates how to use our open-source tool to create a customized scheduler and covers problem formulation with ILP and GAMS. Second, it summarizes results on the application to three real architectures (TRIPS,DySER,PLUG), demonstrating the technique?s practicality and competitiveness with existing schedulers
EMM: Energy-Aware Mobility Management for Mobile Edge Computing in Ultra Dense Networks
Merging mobile edge computing (MEC) functionality with the dense deployment
of base stations (BSs) provides enormous benefits such as a real proximity, low
latency access to computing resources. However, the envisioned integration
creates many new challenges, among which mobility management (MM) is a critical
one. Simply applying existing radio access oriented MM schemes leads to poor
performance mainly due to the co-provisioning of radio access and computing
services of the MEC-enabled BSs. In this paper, we develop a novel user-centric
energy-aware mobility management (EMM) scheme, in order to optimize the delay
due to both radio access and computation, under the long-term energy
consumption constraint of the user. Based on Lyapunov optimization and
multi-armed bandit theories, EMM works in an online fashion without future
system state information, and effectively handles the imperfect system state
information. Theoretical analysis explicitly takes radio handover and
computation migration cost into consideration and proves a bounded deviation on
both the delay performance and energy consumption compared to the oracle
solution with exact and complete future system information. The proposed
algorithm also effectively handles the scenario in which candidate BSs randomly
switch on/off during the offloading process of a task. Simulations show that
the proposed algorithms can achieve close-to-optimal delay performance while
satisfying the user energy consumption constraint.Comment: 14 pages, 6 figures, an extended version of the paper submitted to
IEEE JSA
Energy-Efficient Flow Scheduling and Routing with Hard Deadlines in Data Center Networks
The power consumption of enormous network devices in data centers has emerged
as a big concern to data center operators. Despite many
traffic-engineering-based solutions, very little attention has been paid on
performance-guaranteed energy saving schemes. In this paper, we propose a novel
energy-saving model for data center networks by scheduling and routing
"deadline-constrained flows" where the transmission of every flow has to be
accomplished before a rigorous deadline, being the most critical requirement in
production data center networks. Based on speed scaling and power-down energy
saving strategies for network devices, we aim to explore the most energy
efficient way of scheduling and routing flows on the network, as well as
determining the transmission speed for every flow. We consider two general
versions of the problem. For the version of only flow scheduling where routes
of flows are pre-given, we show that it can be solved polynomially and we
develop an optimal combinatorial algorithm for it. For the version of joint
flow scheduling and routing, we prove that it is strongly NP-hard and cannot
have a Fully Polynomial-Time Approximation Scheme (FPTAS) unless P=NP. Based on
a relaxation and randomized rounding technique, we provide an efficient
approximation algorithm which can guarantee a provable performance ratio with
respect to a polynomial of the total number of flows.Comment: 11 pages, accepted by ICDCS'1
Petuum: A New Platform for Distributed Machine Learning on Big Data
What is a systematic way to efficiently apply a wide spectrum of advanced ML
programs to industrial scale problems, using Big Models (up to 100s of billions
of parameters) on Big Data (up to terabytes or petabytes)? Modern
parallelization strategies employ fine-grained operations and scheduling beyond
the classic bulk-synchronous processing paradigm popularized by MapReduce, or
even specialized graph-based execution that relies on graph representations of
ML programs. The variety of approaches tends to pull systems and algorithms
design in different directions, and it remains difficult to find a universal
platform applicable to a wide range of ML programs at scale. We propose a
general-purpose framework that systematically addresses data- and
model-parallel challenges in large-scale ML, by observing that many ML programs
are fundamentally optimization-centric and admit error-tolerant,
iterative-convergent algorithmic solutions. This presents unique opportunities
for an integrative system design, such as bounded-error network synchronization
and dynamic scheduling based on ML program structure. We demonstrate the
efficacy of these system designs versus well-known implementations of modern ML
algorithms, allowing ML programs to run in much less time and at considerably
larger model sizes, even on modestly-sized compute clusters.Comment: 15 pages, 10 figures, final version in KDD 2015 under the same titl
- …