22,613 research outputs found
VirtFogSim: A parallel toolbox for dynamic energy-delay performance testing and optimization of 5G Mobile-Fog-Cloud virtualized platforms
It is expected that the pervasive deployment of multi-tier 5G-supported Mobile-Fog-Cloudtechnological computing platforms will constitute an effective means to support the real-time execution of future Internet applications by resource- and energy-limited mobile devices. Increasing interest in this emerging networking-computing technology demands the optimization and performance evaluation of several parts of the underlying infrastructures. However, field trials are challenging due to their operational costs, and in every case, the obtained results could be difficult to repeat and customize. These emergingMobile-Fog-Cloud ecosystems still lack, indeed, customizable software tools for the performance simulation of their computing-networking building blocks. Motivated by these considerations, in this contribution, we present VirtFogSim. It is aMATLAB-supported software toolbox that allows the dynamic joint optimization and tracking of the energy and delay performance of Mobile-Fog-Cloud systems for the execution of applications described by general Directed Application Graphs (DAGs). In a nutshell, the main peculiar features of the proposed VirtFogSim toolbox are that: (i) it allows the joint dynamic energy-aware optimization of the placement of the application tasks and the allocation of the needed computing-networking resources under hard constraints on acceptable overall execution times, (ii) it allows the repeatable and customizable simulation of the resulting energy-delay performance of the overall system; (iii) it allows the dynamic tracking of the performed resource allocation under time-varying operational environments, as those typically featuring mobile applications; (iv) it is equipped with a user-friendly Graphic User Interface (GUI) that supports a number of graphic formats for data rendering, and (v) itsMATLAB code is optimized for running atop multi-core parallel execution platforms. To check both the actual optimization and scalability capabilities of the VirtFogSim toolbox, a number of experimental setups featuring different use cases and operational environments are simulated, and their performances are compared
A Taxonomy for Management and Optimization of Multiple Resources in Edge Computing
Edge computing is promoted to meet increasing performance needs of
data-driven services using computational and storage resources close to the end
devices, at the edge of the current network. To achieve higher performance in
this new paradigm one has to consider how to combine the efficiency of resource
usage at all three layers of architecture: end devices, edge devices, and the
cloud. While cloud capacity is elastically extendable, end devices and edge
devices are to various degrees resource-constrained. Hence, an efficient
resource management is essential to make edge computing a reality. In this
work, we first present terminology and architectures to characterize current
works within the field of edge computing. Then, we review a wide range of
recent articles and categorize relevant aspects in terms of 4 perspectives:
resource type, resource management objective, resource location, and resource
use. This taxonomy and the ensuing analysis is used to identify some gaps in
the existing research. Among several research gaps, we found that research is
less prevalent on data, storage, and energy as a resource, and less extensive
towards the estimation, discovery and sharing objectives. As for resource
types, the most well-studied resources are computation and communication
resources. Our analysis shows that resource management at the edge requires a
deeper understanding of how methods applied at different levels and geared
towards different resource types interact. Specifically, the impact of mobility
and collaboration schemes requiring incentives are expected to be different in
edge architectures compared to the classic cloud solutions. Finally, we find
that fewer works are dedicated to the study of non-functional properties or to
quantifying the footprint of resource management techniques, including
edge-specific means of migrating data and services.Comment: Accepted in the Special Issue Mobile Edge Computing of the Wireless
Communications and Mobile Computing journa
Profitable Task Allocation in Mobile Cloud Computing
We propose a game theoretic framework for task allocation in mobile cloud
computing that corresponds to offloading of compute tasks to a group of nearby
mobile devices. Specifically, in our framework, a distributor node holds a
multidimensional auction for allocating the tasks of a job among nearby mobile
nodes based on their computational capabilities and also the cost of
computation at these nodes, with the goal of reducing the overall job
completion time. Our proposed auction also has the desired incentive
compatibility property that ensures that mobile devices truthfully reveal their
capabilities and costs and that those devices benefit from the task allocation.
To deal with node mobility, we perform multiple auctions over adaptive time
intervals. We develop a heuristic approach to dynamically find the best time
intervals between auctions to minimize unnecessary auctions and the
accompanying overheads. We evaluate our framework and methods using both real
world and synthetic mobility traces. Our evaluation results show that our game
theoretic framework improves the job completion time by a factor of 2-5 in
comparison to the time taken for executing the job locally, while minimizing
the number of auctions and the accompanying overheads. Our approach is also
profitable for the nearby nodes that execute the distributor's tasks with these
nodes receiving a compensation higher than their actual costs
Exploiting Non-Causal CPU-State Information for Energy-Efficient Mobile Cooperative Computing
Scavenging the idling computation resources at the enormous number of mobile
devices can provide a powerful platform for local mobile cloud computing. The
vision can be realized by peer-to-peer cooperative computing between edge
devices, referred to as co-computing. This paper considers a co-computing
system where a user offloads computation of input-data to a helper. The helper
controls the offloading process for the objective of minimizing the user's
energy consumption based on a predicted helper's CPU-idling profile that
specifies the amount of available computation resource for co-computing.
Consider the scenario that the user has one-shot input-data arrival and the
helper buffers offloaded bits. The problem for energy-efficient co-computing is
formulated as two sub-problems: the slave problem corresponding to adaptive
offloading and the master one to data partitioning. Given a fixed offloaded
data size, the adaptive offloading aims at minimizing the energy consumption
for offloading by controlling the offloading rate under the deadline and buffer
constraints. By deriving the necessary and sufficient conditions for the
optimal solution, we characterize the structure of the optimal policies and
propose algorithms for computing the policies. Furthermore, we show that the
problem of optimal data partitioning for offloading and local computing at the
user is convex, admitting a simple solution using the sub-gradient method.
Last, the developed design approach for co-computing is extended to the
scenario of bursty data arrivals at the user accounting for data causality
constraints. Simulation results verify the effectiveness of the proposed
algorithms.Comment: Submitted to possible journa
- …