999 research outputs found
Joint Data compression and Computation offloading in Hierarchical Fog-Cloud Systems
Data compression has the potential to significantly improve the computation
offloading performance in hierarchical fog-cloud systems. However, it remains
unknown how to optimally determine the compression ratio jointly with the
computation offloading decisions and the resource allocation. This joint
optimization problem is studied in the current paper where we aim to minimize
the maximum weighted energy and service delay cost (WEDC) of all users. First,
we consider a scenario where data compression is performed only at the mobile
users. We prove that the optimal offloading decisions have a threshold
structure. Moreover, a novel three-step approach employing convexification
techniques is developed to optimize the compression ratios and the resource
allocation. Then, we address the more general design where data compression is
performed at both the mobile users and the fog server. We propose three
efficient algorithms to overcome the strong coupling between the offloading
decisions and resource allocation. We show that the proposed optimal algorithm
for data compression at only the mobile users can reduce the WEDC by a few
hundred percent compared to computation offloading strategies that do not
leverage data compression or use sub-optimal optimization approaches. Besides,
the proposed algorithms for additional data compression at the fog server can
further reduce the WEDC
A Taxonomy for Management and Optimization of Multiple Resources in Edge Computing
Edge computing is promoted to meet increasing performance needs of
data-driven services using computational and storage resources close to the end
devices, at the edge of the current network. To achieve higher performance in
this new paradigm one has to consider how to combine the efficiency of resource
usage at all three layers of architecture: end devices, edge devices, and the
cloud. While cloud capacity is elastically extendable, end devices and edge
devices are to various degrees resource-constrained. Hence, an efficient
resource management is essential to make edge computing a reality. In this
work, we first present terminology and architectures to characterize current
works within the field of edge computing. Then, we review a wide range of
recent articles and categorize relevant aspects in terms of 4 perspectives:
resource type, resource management objective, resource location, and resource
use. This taxonomy and the ensuing analysis is used to identify some gaps in
the existing research. Among several research gaps, we found that research is
less prevalent on data, storage, and energy as a resource, and less extensive
towards the estimation, discovery and sharing objectives. As for resource
types, the most well-studied resources are computation and communication
resources. Our analysis shows that resource management at the edge requires a
deeper understanding of how methods applied at different levels and geared
towards different resource types interact. Specifically, the impact of mobility
and collaboration schemes requiring incentives are expected to be different in
edge architectures compared to the classic cloud solutions. Finally, we find
that fewer works are dedicated to the study of non-functional properties or to
quantifying the footprint of resource management techniques, including
edge-specific means of migrating data and services.Comment: Accepted in the Special Issue Mobile Edge Computing of the Wireless
Communications and Mobile Computing journa
Edge and Central Cloud Computing: A Perfect Pairing for High Energy Efficiency and Low-latency
In this paper, we study the coexistence and synergy between edge and central
cloud computing in a heterogeneous cellular network (HetNet), which contains a
multi-antenna macro base station (MBS), multiple multi-antenna small base
stations (SBSs) and multiple single-antenna user equipment (UEs). The SBSs are
empowered by edge clouds offering limited computing services for UEs, whereas
the MBS provides high-performance central cloud computing services to UEs via a
restricted multiple-input multiple-output (MIMO) backhaul to their associated
SBSs. With processing latency constraints at the central and edge networks, we
aim to minimize the system energy consumption used for task offloading and
computation. The problem is formulated by jointly optimizing the cloud
selection, the UEs' transmit powers, the SBSs' receive beamformers, and the
SBSs' transmit covariance matrices, which is {a mixed-integer and non-convex
optimization problem}. Based on methods such as decomposition approach and
successive pseudoconvex approach, a tractable solution is proposed via an
iterative algorithm. The simulation results show that our proposed solution can
achieve great performance gain over conventional schemes using edge or central
cloud alone. Also, with large-scale antennas at the MBS, the massive MIMO
backhaul can significantly reduce the complexity of the proposed algorithm and
obtain even better performance.Comment: Accepted in IEEE Transactions on Wireless Communication
- …