2,983 research outputs found
Joint Latency-Energy optimization scheme for Offloading in Mobile Edge computing environment based in Deep Reinforcement Learning
With the increasing number of mobile devices (MD), IoT devices, and computation-intensive tasks deployed on these devices, there is a need to increase the efficiency and speed of the deliverable. Due to inadequate resources, it is infeasible to compute all the tasks locally. Similarly, due to time constraints, it is not possible to compute the entire task at a remote site. Edge computing (EC) and cloud computing (CC) play the role of providing the resources to these devices on the fly. But a major drawback is increased delay and energy consumption due to transmission and offloading of computation tasks to these remote systems. There is a need to divide the task for computation at local sites, edge servers, and cloud servers to complete tasks with minimum delay and energy consumption. This paper proposes offloading strategy computation using Multi-Period Deep Deterministic Policy Gradient (MP-DDPG) algorithm based on Reinforcement Learning (RL) to optimize the latency caused and energy consumed. We formulate our problem as a Multi-period Markov Decision Process (MP-MDP). In this paper, we use the two-tier offloading architecture including more than one mobile device (MD), two EC-servers, and one CC-server as computation sites. Further, we also compare our proposed algorithm using one-tier architecture and one edge server with the Deep Deterministic Policy Gradient (DDPG) algorithm with similar architecture
A Taxonomy for Management and Optimization of Multiple Resources in Edge Computing
Edge computing is promoted to meet increasing performance needs of
data-driven services using computational and storage resources close to the end
devices, at the edge of the current network. To achieve higher performance in
this new paradigm one has to consider how to combine the efficiency of resource
usage at all three layers of architecture: end devices, edge devices, and the
cloud. While cloud capacity is elastically extendable, end devices and edge
devices are to various degrees resource-constrained. Hence, an efficient
resource management is essential to make edge computing a reality. In this
work, we first present terminology and architectures to characterize current
works within the field of edge computing. Then, we review a wide range of
recent articles and categorize relevant aspects in terms of 4 perspectives:
resource type, resource management objective, resource location, and resource
use. This taxonomy and the ensuing analysis is used to identify some gaps in
the existing research. Among several research gaps, we found that research is
less prevalent on data, storage, and energy as a resource, and less extensive
towards the estimation, discovery and sharing objectives. As for resource
types, the most well-studied resources are computation and communication
resources. Our analysis shows that resource management at the edge requires a
deeper understanding of how methods applied at different levels and geared
towards different resource types interact. Specifically, the impact of mobility
and collaboration schemes requiring incentives are expected to be different in
edge architectures compared to the classic cloud solutions. Finally, we find
that fewer works are dedicated to the study of non-functional properties or to
quantifying the footprint of resource management techniques, including
edge-specific means of migrating data and services.Comment: Accepted in the Special Issue Mobile Edge Computing of the Wireless
Communications and Mobile Computing journa
- …