383 research outputs found

    Collaborative Inference in DNN-based Satellite Systems with Dynamic Task Streams

    Full text link
    As a driving force in the advancement of intelligent in-orbit applications, DNN models have been gradually integrated into satellites, producing daily latency-constraint and computation-intensive tasks. However, the substantial computation capability of DNN models, coupled with the instability of the satellite-ground link, pose significant challenges, hindering timely completion of tasks. It becomes necessary to adapt to task stream changes when dealing with tasks requiring latency guarantees, such as dynamic observation tasks on the satellites. To this end, we consider a system model for a collaborative inference system with latency constraints, leveraging the multi-exit and model partition technology. To address this, we propose an algorithm, which is tailored to effectively address the trade-off between task completion and maintaining satisfactory task accuracy by dynamically choosing early-exit and partition points. Simulation evaluations show that our proposed algorithm significantly outperforms baseline algorithms across the task stream with strict latency constraints

    Self organising cloud cells: a resource efficient network densification strategy

    Get PDF
    Network densification is envisioned as the key enabler for 2020 vision that requires cellular systems to grow in capacity by hundreds of times to cope with unprecedented traffic growth trends being witnessed since advent of broadband on the move. However, increased energy consumption and complex mobility management associated with network densifications remain as the two main challenges to be addressed before further network densification can be exploited on a wide scale. In the wake of these challenges, this paper proposes and evaluates a novel dense network deployment strategy for increasing the capacity of future cellular systems without sacrificing energy efficiency and compromising mobility performance. Our deployment architecture consists of smart small cells, called cloud nodes, which provide data coverage to individual users on a demand bases while taking into account the spatial and temporal dynamics of user mobility and traffic. The decision to activate the cloud nodes, such that certain performance objectives at system level are targeted, is carried out by the overlaying macrocell based on a fuzzy-logic framework. We also compare the proposed architecture with conventional macrocell only deployment and pure microcell-based dense deployment in terms of blocking probability, handover probability and energy efficiency and discuss and quantify the trade-offs therein

    ENORM: A Framework For Edge NOde Resource Management

    Get PDF
    Current computing techniques using the cloud as a centralised server will become untenable as billions of devices get connected to the Internet. This raises the need for fog computing, which leverages computing at the edge of the network on nodes, such as routers, base stations and switches, along with the cloud. However, to realise fog computing the challenge of managing edge nodes will need to be addressed. This paper is motivated to address the resource management challenge. We develop the first framework to manage edge nodes, namely the Edge NOde Resource Management (ENORM) framework. Mechanisms for provisioning and auto-scaling edge node resources are proposed. The feasibility of the framework is demonstrated on a PokeMon Go-like online game use-case. The benefits of using ENORM are observed by reduced application latency between 20% - 80% and reduced data transfer and communication frequency between the edge node and the cloud by up to 95\%. These results highlight the potential of fog computing for improving the quality of service and experience.Comment: 14 pages; accepted to IEEE Transactions on Services Computing on 12 September 201

    On distributed mobile edge computing

    Get PDF
    Mobile Cloud Computing (MCC) has been proposed to offload the workloads of mobile applications from mobile devices to the cloud in order to not only reduce energy consumption of mobile devices but also accelerate the execution of mobile applications. Owing to the long End-to-End (E2E) delay between mobile devices and the cloud, offloading the workloads of many interactive mobile applications to the cloud may not be suitable. That is, these mobile applications require a huge amount of computing resources to process their workloads as well as a low E2E delay between mobile devices and computing resources, which cannot be satisfied by the current MCC technology. In order to reduce the E2E delay, a novel cloudlet network architecture is proposed to bring the computing and storage resources from the remote cloud to the mobile edge. In the cloudlet network, each mobile user is associated with a specific Avatar (i.e., a dedicated Virtual Machine (VM) providing computing and storage resources to its mobile user) in the nearby cloudlet via its associated Base Station (BS). Thus, mobile users can offload their workloads to their Avatars with low E2E delay (i.e., one wireless hop). However, mobile users may roam among BSs in the mobile network, and so the E2E delay between mobile users and their Avatars may become worse if the Avatars remain in their original cloudlets. Thus, Avatar handoff is proposed to migrate an Avatar from one cloudlet into another to reduce the E2E delay between the Avatar and its mobile user. The LatEncy aware Avatar handDoff (LEAD) algorithm is designed to determine the location of each mobile user\u27s Avatar in each time slot in order to minimize the average E2E delay among all the mobile users and their Avatars. The performance of LEAD is demonstrated via extensive simulations. The cloudlet network architecture not only facilitates mobile users in offloading their computational tasks but also empowers Internet of Things (IoT). Popular IoT resources are proposed to be cached in nearby brokers, which are considered as application layer middleware nodes hosted by cloudlets in the cloudlet network, to reduce the energy consumption of servers. In addition, an Energy Aware and latency guaranteed dynamic reSourcE caching (EASE) strategy is proposed to enable each broker to cache suitable popular resources such that the energy consumption from the servers is minimized and the average delay of delivering the contents of the resources to the corresponding clients is guaranteed. The performance of EASE is demonstrated via extensive simulations. The future work comprises two parts. First, caching popular IoT resources in nearby brokers may incur unbalanced traffic loads among brokers, thus increasing the average delay of delivering the contents of the resources. Thus, how to balance the traffic loads among brokers to speed up IoT content delivery process requires further investigation. Second, drone assisted mobile access network architecture will be briefly investigated to accelerate communications between mobile users and their Avatars

    A Threshold Based Handover Triggering Scheme in Heterogeneous Wireless Networks

    Get PDF
    The widespread popularity of Wireless Local Area Network (WLAN) is recognized as an effective approach to complementing cellular networks for the high data rate and cost effective connectivity delivered to mobile users. Efficient handover and offloading schemes for integrated WLAN and cellular networks, referred to as Heterogeneous Wireless Networks, have thus attracted lots of attentions from both academia and industry. This paper proposes a novel Multiple-Threshold based Triggering (MTT) scheme for Cellular-to-WLAN handover control. Aiming at minimizing the probability of handover failures and unnecessary handovers, three thresholds are calculated based on a variety of network parameters such as system performance requirements, radius of the WLAN coverage, user mobility and handover delays. The thresholds are then compared against the predicted user residence time and estimated channel holding time inside WLAN to make vertical handover decisions (VHDs). Simulations were carried out to evaluate the effectiveness of MTT and results show that MTT minimizes handover failures and avoids unnecessary handovers in integrated cellular and WLAN networks, thus providing satisfactory Quality of Service (QoS) to users and improving system resource utilization

    Context-Aware Handover Policies in HetNets

    Get PDF
    Next generation cellular systems are expected to entail a wide variety of wireless coverage zones, with cells of different sizes and capacities that can overlap in space and share the transmission resources. In this scenario, which is referred to as Heterogeneous Networks (HetNets), a fundamental challenge is the management of the handover process between macro, femto and pico cells. To limit the number of handovers and the signaling between the cells, it will hence be crucial to manage the user's mobility considering the context parameters, such as cells size, traffic loads, and user velocity. In this paper, we propose a theoretical model to characterize the performance of a mobile user in a HetNet scenario as a function of the user's mobility, the power profile of the neighboring cells, the handover parameters, and the traffic load of the different cells. We propose a Markov-based framework to model the handover process for the mobile user, and derive an optimal context-dependent handover criterion. The mathematical model is validated by means of simulations, comparing the performance of our strategy with conventional handover optimization techniques in different scenarios. Finally, we show the impact of the handover regulation on the users performance and how it is possible to improve the users capacity exploiting context information

    A Bilevel Optimization Approach for Joint Offloading Decision and Resource Allocation in Cooperative Mobile Edge Computing

    Get PDF
    This paper studies a multi-user cooperative mobile edge computing offloading (CoMECO) system in a multi-user interference environment, in which delay-sensitive tasks may be executed on local devices, cooperative devices, or the primary MEC server. In this system, we jointly optimize the offloading decision and computation resource allocation for minimizing the total energy consumption of all mobile users under the delay constraint. If this problem is solved directly, the offloading decision and computation resource allocation are generally generated separately at the same time. Note, however, that they are closely coupled. Therefore, under this condition, their dependency is not well considered, thus leading to poor performance. We transform this problem into a bilevel optimization problem, in which the offloading decision is generated in the upper level, and then the optimal allocation of computation resources is obtained in the lower level based on the given offloading decision. In this way, the dependency between the offloading decision and computation resource allocation can be fully taken into account. Subsequently, a bilevel optimization approach, called BiJOR, is proposed. In BiJOR, candidate modes are first pruned to reduce the number of infeasible offloading decisions. Afterward, the upper level optimization problem is solved by ant colony system (ACS). Furthermore, a sorting strategy is incorporated into ACS to construct feasible offloading decisions with a higher probability and a local search operator is designed in ACS to accelerate the convergence. For the lower level optimization problem, it is solved by the monotonic optimization method. In addition, BiJOR is extended to deal with a complex scenario with the channel selection. Extensive experiments are carried out to investigate the performance of BiJOR on two sets of instances with up to 400 mobile users. The experimental results demonstrate the effectiveness of BiJOR and the superiority of the CoMECO system
    • …
    corecore