30,372 research outputs found
How to Allocate R&D (and Other) Subsidies: An Experimentally Tested Policy Recommendation
This paper evaluates how R&D subsidies to the business sector are typically awarded. We identify two sources of ine_ciency: the selection based on a ranking of individual projects, rather than complete allocations, and the failure to induce competition among applicants in order to extract and use information about the necessary funding. In order to correct these ine_- ciencies we propose mechanisms that include some form of an auction in which applicants bid for subsidies. Our proposals are tested in a simulation and in controlled lab experiments. The results suggest that adopting our proposals may considerably improve the allocation
Online Learning for Offloading and Autoscaling in Energy Harvesting Mobile Edge Computing
Mobile edge computing (a.k.a. fog computing) has recently emerged to enable
in-situ processing of delay-sensitive applications at the edge of mobile
networks. Providing grid power supply in support of mobile edge computing,
however, is costly and even infeasible (in certain rugged or under-developed
areas), thus mandating on-site renewable energy as a major or even sole power
supply in increasingly many scenarios. Nonetheless, the high intermittency and
unpredictability of renewable energy make it very challenging to deliver a high
quality of service to users in energy harvesting mobile edge computing systems.
In this paper, we address the challenge of incorporating renewables into mobile
edge computing and propose an efficient reinforcement learning-based resource
management algorithm, which learns on-the-fly the optimal policy of dynamic
workload offloading (to the centralized cloud) and edge server provisioning to
minimize the long-term system cost (including both service delay and
operational cost). Our online learning algorithm uses a decomposition of the
(offline) value iteration and (online) reinforcement learning, thus achieving a
significant improvement of learning rate and run-time performance when compared
to standard reinforcement learning algorithms such as Q-learning. We prove the
convergence of the proposed algorithm and analytically show that the learned
policy has a simple monotone structure amenable to practical implementation.
Our simulation results validate the efficacy of our algorithm, which
significantly improves the edge computing performance compared to fixed or
myopic optimization schemes and conventional reinforcement learning algorithms.Comment: arXiv admin note: text overlap with arXiv:1701.01090 by other author
Feedback Allocation For OFDMA Systems With Slow Frequency-domain Scheduling
We study the problem of allocating limited feedback resources across multiple
users in an orthogonal-frequency-division-multiple-access downlink system with
slow frequency-domain scheduling. Many flavors of slow frequency-domain
scheduling (e.g., persistent scheduling, semi-persistent scheduling), that
adapt user-sub-band assignments on a slower time-scale, are being considered in
standards such as 3GPP Long-Term Evolution. In this paper, we develop a
feedback allocation algorithm that operates in conjunction with any arbitrary
slow frequency-domain scheduler with the goal of improving the throughput of
the system. Given a user-sub-band assignment chosen by the scheduler, the
feedback allocation algorithm involves solving a weighted sum-rate maximization
at each (slow) scheduling instant. We first develop an optimal
dynamic-programming-based algorithm to solve the feedback allocation problem
with pseudo-polynomial complexity in the number of users and in the total
feedback bit budget. We then propose two approximation algorithms with
complexity further reduced, for scenarios where the problem exhibits additional
structure.Comment: Accepted to IEEE Transactions on Signal Processin
- …