22,552 research outputs found
When Backpressure Meets Predictive Scheduling
Motivated by the increasing popularity of learning and predicting human user
behavior in communication and computing systems, in this paper, we investigate
the fundamental benefit of predictive scheduling, i.e., predicting and
pre-serving arrivals, in controlled queueing systems. Based on a lookahead
window prediction model, we first establish a novel equivalence between the
predictive queueing system with a \emph{fully-efficient} scheduling scheme and
an equivalent queueing system without prediction. This connection allows us to
analytically demonstrate that predictive scheduling necessarily improves system
delay performance and can drive it to zero with increasing prediction power. We
then propose the \textsf{Predictive Backpressure (PBP)} algorithm for achieving
optimal utility performance in such predictive systems. \textsf{PBP}
efficiently incorporates prediction into stochastic system control and avoids
the great complication due to the exponential state space growth in the
prediction window size. We show that \textsf{PBP} can achieve a utility
performance that is within of the optimal, for any ,
while guaranteeing that the system delay distribution is a
\emph{shifted-to-the-left} version of that under the original Backpressure
algorithm. Hence, the average packet delay under \textsf{PBP} is strictly
better than that under Backpressure, and vanishes with increasing prediction
window size. This implies that the resulting utility-delay tradeoff with
predictive scheduling beats the known optimal tradeoff for systems without prediction
Timely-Throughput Optimal Scheduling with Prediction
Motivated by the increasing importance of providing delay-guaranteed services
in general computing and communication systems, and the recent wide adoption of
learning and prediction in network control, in this work, we consider a general
stochastic single-server multi-user system and investigate the fundamental
benefit of predictive scheduling in improving timely-throughput, being the rate
of packets that are delivered to destinations before their deadlines. By
adopting an error rate-based prediction model, we first derive a Markov
decision process (MDP) solution to optimize the timely-throughput objective
subject to an average resource consumption constraint. Based on a packet-level
decomposition of the MDP, we explicitly characterize the optimal scheduling
policy and rigorously quantify the timely-throughput improvement due to
predictive-service, which scales as
,
where are constants, is the
true-positive rate in prediction, is the false-negative rate, is the
packet deadline and is the prediction window size. We also conduct
extensive simulations to validate our theoretical findings. Our results provide
novel insights into how prediction and system parameters impact performance and
provide useful guidelines for designing predictive low-latency control
algorithms.Comment: 14 pages, 7 figure
Cooperative Multi-Bitrate Video Caching and Transcoding in Multicarrier NOMA-Assisted Heterogeneous Virtualized MEC Networks
Cooperative video caching and transcoding in mobile edge computing (MEC)
networks is a new paradigm for future wireless networks, e.g., 5G and 5G
beyond, to reduce scarce and expensive backhaul resource usage by prefetching
video files within radio access networks (RANs). Integration of this technique
with other advent technologies, such as wireless network virtualization and
multicarrier non-orthogonal multiple access (MC-NOMA), provides more flexible
video delivery opportunities, which leads to enhancements both for the
network's revenue and for the end-users' service experience. In this regard, we
propose a two-phase RAF for a parallel cooperative joint multi-bitrate video
caching and transcoding in heterogeneous virtualized MEC networks. In the cache
placement phase, we propose novel proactive delivery-aware cache placement
strategies (DACPSs) by jointly allocating physical and radio resources based on
network stochastic information to exploit flexible delivery opportunities.
Then, for the delivery phase, we propose a delivery policy based on the user
requests and network channel conditions. The optimization problems
corresponding to both phases aim to maximize the total revenue of network
slices, i.e., virtual networks. Both problems are non-convex and suffer from
high-computational complexities. For each phase, we show how the problem can be
solved efficiently. We also propose a low-complexity RAF in which the
complexity of the delivery algorithm is significantly reduced. A Delivery-aware
cache refreshment strategy (DACRS) in the delivery phase is also proposed to
tackle the dynamically changes of network stochastic information. Extensive
numerical assessments demonstrate a performance improvement of up to 30% for
our proposed DACPSs and DACRS over traditional approaches.Comment: 53 pages, 24 figure
- …