200 research outputs found

    Emerging Edge Computing Technologies for Distributed Internet of Things (IoT) Systems

    Get PDF
    The ever-increasing growth in the number of connected smart devices and various Internet of Things (IoT) verticals is leading to a crucial challenge of handling massive amount of raw data generated from distributed IoT systems and providing real-time feedback to the end-users. Although existing cloud-computing paradigm has an enormous amount of virtual computing power and storage capacity, it is not suitable for latency-sensitive applications and distributed systems due to the involved latency and its centralized mode of operation. To this end, edge/fog computing has recently emerged as the next generation of computing systems for extending cloud-computing functions to the edges of the network. Despite several benefits of edge computing such as geo-distribution, mobility support and location awareness, various communication and computing related challenges need to be addressed in realizing edge computing technologies for future IoT systems. In this regard, this paper provides a holistic view on the current issues and effective solutions by classifying the emerging technologies in regard to the joint coordination of radio and computing resources, system optimization and intelligent resource management. Furthermore, an optimization framework for edge-IoT systems is proposed to enhance various performance metrics such as throughput, delay, resource utilization and energy consumption. Finally, a Machine Learning (ML) based case study is presented along with some numerical results to illustrate the significance of edge computing.Comment: 16 pages, 4 figures, 2 tables, submitted to IEEE Wireless Communications Magazin

    A Survey on UAV-enabled Edge Computing: Resource Management Perspective

    Full text link
    Edge computing facilitates low-latency services at the network's edge by distributing computation, communication, and storage resources within the geographic proximity of mobile and Internet-of-Things (IoT) devices. The recent advancement in Unmanned Aerial Vehicles (UAVs) technologies has opened new opportunities for edge computing in military operations, disaster response, or remote areas where traditional terrestrial networks are limited or unavailable. In such environments, UAVs can be deployed as aerial edge servers or relays to facilitate edge computing services. This form of computing is also known as UAV-enabled Edge Computing (UEC), which offers several unique benefits such as mobility, line-of-sight, flexibility, computational capability, and cost-efficiency. However, the resources on UAVs, edge servers, and IoT devices are typically very limited in the context of UEC. Efficient resource management is, therefore, a critical research challenge in UEC. In this article, we present a survey on the existing research in UEC from the resource management perspective. We identify a conceptual architecture, different types of collaborations, wireless communication models, research directions, key techniques and performance indicators for resource management in UEC. We also present a taxonomy of resource management in UEC. Finally, we identify and discuss some open research challenges that can stimulate future research directions for resource management in UEC.Comment: 36 pages, Accepted to ACM CSU

    A Max-Min Task Offloading Algorithm for Mobile Edge Computing Using Non-Orthogonal Multiple Access

    Full text link
    To mitigate computational power gap between the network core and edges, mobile edge computing (MEC) is poised to play a fundamental role in future generations of wireless networks. In this letter, we consider a non-orthogonal multiple access (NOMA) transmission model to maximize the worst task to be offloaded among all users to the network edge server. A provably convergent and efficient algorithm is developed to solve the considered non-convex optimization problem for maximizing the minimum number of offloaded bits in a multi-user NOMAMEC system. Compared to the approach of optimized orthogonal multiple access (OMA), for given MEC delay, power and energy limits, the NOMA-based system considerably outperforms its OMA-based counterpart in MEC settings. Numerical results demonstrate that the proposed algorithm for NOMA-based MEC is particularly useful for delay sensitive applications.Comment: 5 pages, 5 figure

    Enabling Technologies for Ultra-Reliable and Low Latency Communications: From PHY and MAC Layer Perspectives

    Full text link
    © 1998-2012 IEEE. Future 5th generation networks are expected to enable three key services-enhanced mobile broadband, massive machine type communications and ultra-reliable and low latency communications (URLLC). As per the 3rd generation partnership project URLLC requirements, it is expected that the reliability of one transmission of a 32 byte packet will be at least 99.999% and the latency will be at most 1 ms. This unprecedented level of reliability and latency will yield various new applications, such as smart grids, industrial automation and intelligent transport systems. In this survey we present potential future URLLC applications, and summarize the corresponding reliability and latency requirements. We provide a comprehensive discussion on physical (PHY) and medium access control (MAC) layer techniques that enable URLLC, addressing both licensed and unlicensed bands. This paper evaluates the relevant PHY and MAC techniques for their ability to improve the reliability and reduce the latency. We identify that enabling long-term evolution to coexist in the unlicensed spectrum is also a potential enabler of URLLC in the unlicensed band, and provide numerical evaluations. Lastly, this paper discusses the potential future research directions and challenges in achieving the URLLC requirements

    Holistic resource management in UAV-assisted wireless networks

    Get PDF
    Unmanned aerial vehicles (UAVs) are considered as a promising solution to assist terrestrial networks in future wireless networks (i.e., beyond fifth-generation (B5G) and sixth-generation (6G)). The convergence of various technologies requires future wireless networks to provide multiple functionalities, including communication, computing, control, and caching (4C), necessary for applications such as connected robotics and autonomous systems. The majority of existing works consider the developments in 4C individually, which limits the cooperation among 4C for potential gains. UAVs have been recently introduced to supplement mobile edge computing (MEC) in terrestrial networks to reduce network latency by providing mobile resources at the network edge in future wireless networks. However, compared to ground base stations (BSs), the limited resources at the network edge call for holistic management of the resources, which requires joint optimization. We provide a comprehensive review of holistic resource management in UAV-assisted wireless networks. Integrated resource management considers the challenges associated with aerial networks (such as three-dimensional (3D) placement of UAVs, trajectory planning, channel modelling, and backhaul connectivity) and terrestrial networks (such as limited bandwidth, power, and interference). We present architectures (source-UAV-destination and UAV-destination architecture) and 4C in UAV-assisted wireless networks. We then provide a detailed discussion on resource management by categorizing the optimization problems into individual or combinations of two (communication and computation) or three (communication, computation and control). Moreover, solution approaches and performance metrics are discussed and analyzed for different objectives and problem types. We formulate a mathematical framework for holistic resource management to minimize the linear combination of network latency and cost for user association while guaranteeing the offloading, computing, and caching constraints. Binary decision variables are used to allocate offloading and computing resources. Since the decision variables are binary and constraints are linear, the formulated problem is a binary linear programming problem. We propose a heuristic algorithm based on the interior point method by exploiting the optimization structure of the problem to get a sub-optimal solution with less complexity. Simulation results show the effectiveness of the proposed work when compared to the optimal results obtained using branch and bound. Finally, we discuss insight into the potential future research areas to address the challenges of holistic resource management in UAV-assisted wireless networks

    A lightweight secure adaptive approach for internet-of-medical-things healthcare applications in edge-cloud-based networks

    Get PDF
    Mobile-cloud-based healthcare applications are increasingly growing in practice. For instance, healthcare, transport, and shopping applications are designed on the basis of the mobile cloud. For executing mobile-cloud applications, offloading and scheduling are fundamental mechanisms. However, mobile healthcare workflow applications with these methods are widely ignored, demanding applications in various aspects for healthcare monitoring, live healthcare service, and biomedical firms. However, these offloading and scheduling schemes do not consider the workflow applications' execution in their models. This paper develops a lightweight secure efficient offloading scheduling (LSEOS) metaheuristic model. LSEOS consists of light weight, and secure offloading and scheduling methods whose execution offloading delay is less than that of existing methods. The objective of LSEOS is to run workflow applications on other nodes and minimize the delay and security risk in the system. The metaheuristic LSEOS consists of the following components: adaptive deadlines, sorting, and scheduling with neighborhood search schemes. Compared to current strategies for delay and security validation in a model, computational results revealed that the LSEOS outperformed all available offloading and scheduling methods for process applications by 10% security ratio and by 29% regarding delays

    Towards Massive Machine Type Communications in Ultra-Dense Cellular IoT Networks: Current Issues and Machine Learning-Assisted Solutions

    Get PDF
    The ever-increasing number of resource-constrained Machine-Type Communication (MTC) devices is leading to the critical challenge of fulfilling diverse communication requirements in dynamic and ultra-dense wireless environments. Among different application scenarios that the upcoming 5G and beyond cellular networks are expected to support, such as eMBB, mMTC and URLLC, mMTC brings the unique technical challenge of supporting a huge number of MTC devices, which is the main focus of this paper. The related challenges include QoS provisioning, handling highly dynamic and sporadic MTC traffic, huge signalling overhead and Radio Access Network (RAN) congestion. In this regard, this paper aims to identify and analyze the involved technical issues, to review recent advances, to highlight potential solutions and to propose new research directions. First, starting with an overview of mMTC features and QoS provisioning issues, we present the key enablers for mMTC in cellular networks. Along with the highlights on the inefficiency of the legacy Random Access (RA) procedure in the mMTC scenario, we then present the key features and channel access mechanisms in the emerging cellular IoT standards, namely, LTE-M and NB-IoT. Subsequently, we present a framework for the performance analysis of transmission scheduling with the QoS support along with the issues involved in short data packet transmission. Next, we provide a detailed overview of the existing and emerging solutions towards addressing RAN congestion problem, and then identify potential advantages, challenges and use cases for the applications of emerging Machine Learning (ML) techniques in ultra-dense cellular networks. Out of several ML techniques, we focus on the application of low-complexity Q-learning approach in the mMTC scenarios. Finally, we discuss some open research challenges and promising future research directions.Comment: 37 pages, 8 figures, 7 tables, submitted for a possible future publication in IEEE Communications Surveys and Tutorial
    corecore