1,801 research outputs found
Joint Service Caching and Task Offloading for Mobile Edge Computing in Dense Networks
Mobile Edge Computing (MEC) pushes computing functionalities away from the
centralized cloud to the network edge, thereby meeting the latency requirements
of many emerging mobile applications and saving backhaul network bandwidth.
Although many existing works have studied computation offloading policies,
service caching is an equally, if not more important, design topic of MEC, yet
receives much less attention. Service caching refers to caching application
services and their related databases/libraries in the edge server (e.g.
MEC-enabled BS), thereby enabling corresponding computation tasks to be
executed. Because only a small number of application services can be cached in
resource-limited edge server at the same time, which services to cache has to
be judiciously decided to maximize the edge computing performance. In this
paper, we investigate the extremely compelling but much less studied problem of
dynamic service caching in MEC-enabled dense cellular networks. We propose an
efficient online algorithm, called OREO, which jointly optimizes dynamic
service caching and task offloading to address a number of key challenges in
MEC systems, including service heterogeneity, unknown system dynamics, spatial
demand coupling and decentralized coordination. Our algorithm is developed
based on Lyapunov optimization and Gibbs sampling, works online without
requiring future information, and achieves provable close-to-optimal
performance. Simulation results show that our algorithm can effectively reduce
computation latency for end users while keeping energy consumption low
How to Place Your Apps in the Fog -- State of the Art and Open Challenges
Fog computing aims at extending the Cloud towards the IoT so to achieve
improved QoS and to empower latency-sensitive and bandwidth-hungry
applications. The Fog calls for novel models and algorithms to distribute
multi-service applications in such a way that data processing occurs wherever
it is best-placed, based on both functional and non-functional requirements.
This survey reviews the existing methodologies to solve the application
placement problem in the Fog, while pursuing three main objectives. First, it
offers a comprehensive overview on the currently employed algorithms, on the
availability of open-source prototypes, and on the size of test use cases.
Second, it classifies the literature based on the application and Fog
infrastructure characteristics that are captured by available models, with a
focus on the considered constraints and the optimised metrics. Finally, it
identifies some open challenges in application placement in the Fog
Secure Cloud-Edge Deployments, with Trust
Assessing the security level of IoT applications to be deployed to
heterogeneous Cloud-Edge infrastructures operated by different providers is a
non-trivial task. In this article, we present a methodology that permits to
express security requirements for IoT applications, as well as infrastructure
security capabilities, in a simple and declarative manner, and to automatically
obtain an explainable assessment of the security level of the possible
application deployments. The methodology also considers the impact of trust
relations among different stakeholders using or managing Cloud-Edge
infrastructures. A lifelike example is used to showcase the prototyped
implementation of the methodology
Adaptive Dispatching of Tasks in the Cloud
The increasingly wide application of Cloud Computing enables the
consolidation of tens of thousands of applications in shared infrastructures.
Thus, meeting the quality of service requirements of so many diverse
applications in such shared resource environments has become a real challenge,
especially since the characteristics and workload of applications differ widely
and may change over time. This paper presents an experimental system that can
exploit a variety of online quality of service aware adaptive task allocation
schemes, and three such schemes are designed and compared. These are a
measurement driven algorithm that uses reinforcement learning, secondly a
"sensible" allocation algorithm that assigns jobs to sub-systems that are
observed to provide a lower response time, and then an algorithm that splits
the job arrival stream into sub-streams at rates computed from the hosts'
processing capabilities. All of these schemes are compared via measurements
among themselves and with a simple round-robin scheduler, on two experimental
test-beds with homogeneous and heterogeneous hosts having different processing
capacities.Comment: 10 pages, 9 figure
- …