5,385 research outputs found

    Computation Offloading Decision in Mobile Cloud Computing: Enhance Battery Life of Mobile Device

    Get PDF
    Functionality on mobile device is ever richer in daily life. Mobile devices have limited resources like battery life, storage and processor, etc. Nowadays, Mobile Cloud Computing (MCC) bridges the gap between the limited capabilities of mobile devices and the increasing user demand of mobile applications by offloading the computational workloads from local devices to the remote cloud. Deciding to offload some computing tasks or not is a way to solve the limitations of battery life and computing capability of mobile devices. Application offloading is energy efficient only under various conditions for determining where/which code should be executed. This paper presents a Computational Offloading Decision Algorithm (CODA) , to save the battery life of mobile devices, taking into account the CPU load, state of charge, network bandwidth and transmission data size. The system can take decision which method should be offloaded or not based on different context of the mobile device to obtain minimum processing cost. Numerical study is carried out to evaluate the performance of system. Experimental result will demonstrate that the proposed algorithm can significantly reduce energy consumption of mobile device as well as execution time of application

    Cloud-aided wireless systems: communications and radar applications

    Get PDF
    This dissertation focuses on cloud-assisted radio technologies for communication, including mobile cloud computing and Cloud Radio Access Network (C-RAN), and for radar systems. This dissertation first concentrates on cloud-aided communications. Mobile cloud computing, which allows mobile users to run computationally heavy applications on battery limited devices, such as cell phones, is considered initially. Mobile cloud computing enables the offloading of computation-intensive applications from a mobile device to a cloud processor via a wireless interface. The interplay between offloading decisions at the application layer and physical-layer parameters, which determine the energy and latency associated with the mobile-cloud communication, motivates the inter-layer optimization of fine-grained task offloading across both layers. This problem is modeled by using application call graphs, and the joint optimization of application-layer and physical-layer parameters is carried out via a message passing algorithm by minimizing the total energy expenditure of the mobile user. The concept of cloud radio is also being considered for the development of two cellular architectures known as Distributed RAN (D-RAN) and C-RAN, whereby the baseband processing of base stations is carried out in a remote Baseband Processing Unit (BBU). These architectures can reduce the capital and operating expenses of dense deployments at the cost of increasing the communication latency. The effect of this latency, which is due to the fronthaul transmission between the Remote Radio Head (RRH) and the BBU, is then studied for implementation of Hybrid Automatic Repeat Request (HARQ) protocols. Specifically, two novel solutions are proposed, which are based on the control-data separation architecture. The trade-offs involving resources such as the number of transmitting and receiving antennas, transmission power and the blocklength of the transmitted codeword, and the performance of the proposed solutions is investigated in analysis and numerical results. The detection of a target in radar systems requires processing of the signal that is received by the sensors. Similar to cloud radio access networks in communications, this processing of the signals can be carried out in a remote Fusion Center (FC) that is connected to all sensors via limited-capacity fronthaul links. The last part of this dissertation is dedicated to exploring the application of cloud radio to radar systems. In particular, the problem of maximizing the detection performance at the FC jointly over the code vector used by the transmitting antenna and over the statistics of the noise introduced by quantization at the sensors for fronthaul transmission is investigated by adopting the information-theoretic criterion of the Bhattacharyya distance and information-theoretic bounds on the quantization rate

    Multi-User Computation Partitioning for Latency Sensitive Mobile Cloud Applications

    Get PDF
    Elastic partitioning of computations between mobile devices and cloud is an important and challenging research topic for mobile cloud computing. Existing works focus on the single-user computation partitioning, which aims to optimize the application completion time for one particular single user. These works assume that the cloud always has enough resources to execute the computations immediately when they are offloaded to the cloud. However, this assumption does not hold for large scale mobile cloud applications. In these applications, due to the competition for cloud resources among a large number of users, the offloaded computations may be executed with certain scheduling delay on the cloud. Single user partitioning that does not take into account the scheduling delay on the cloud may yield significant performance degradation. In this paper, we study, for the first time, Multi-user Computation Partitioning Problem (MCPP), which considers the partitioning of multiple usersā€™ computations together with the scheduling of offloaded computations on the cloud resources. Instead of pursuing the minimum application completion time for every single user, we aim to achieve minimum average completion time for all the users, based on the number of provisioned resources on the cloud. We show that MCPP is different from and more difficult than the classical job scheduling problems. We design an offline heuristic algorithm, namely SearchAdjust, to solve MCPP. We demonstrate through benchmarks that SearchAdjust outperforms both the single user partitioning approaches and classical job scheduling approaches by 10% on average in terms of application delay. Based on SearchAdjust, we also design an online algorithm for MCPP that can be easily deployed in practical systems. We validate the effectiveness of our online algorithm using real world load traces. Index Termsā€”mobile cloud computing; offloading; computation partitioning; job schedulin

    Smart Application Division and Time Allocation Policy for Computational Offloading in Wireless Powered Mobile Edge Computing

    Get PDF
    Limited battery life and poor computational resources of mobile terminals are challenging problems for the present and future computation-intensive mobile applications. Wireless powered mobile edge computing is one of the solutions, in which wireless energy transfer technology and cloud serverā€™s capabilities are brought to the edge of cellular networks. In wireless powered mobile edge computing systems, the mobile terminals charge their batteries through radio frequency signals and offload their applications to the nearby hybrid access point in the same time slot to minimize their energy consumption and ensure uninterrupted connectivity with hybrid access point. However, the smart division of application into subtasks as well as intelligent partitioning of time slot for harvesting energy and offloading data is a complex problem. In this paper, we propose a novel deep-learning-based offloading and time allocation policy (DOTP) for training a deep neural network that divides the computation application into optimal number of subtasks, decides for the subtasks to be offloaded or executed locally (offloading policy), and divides the time slot for data offloading and energy harvesting (time allocation policy). DOTP takes into account the current battery level, energy consumption, and time delay of mobile terminal. A comprehensive cost function is formulated, which uses all the aforementioned metrics to calculate the cost for all number of subtasks. We propose an algorithm that selects the optimal number of subtasks, partial offloading policy, and time allocation policy to generate a huge dataset for training a deep neural network and hence avoid huge computational overhead in partial offloading. Simulation results are compared with the benchmark schemes of total offloading, local execution, and partial offloading. It is evident from the results that the proposed algorithm outperforms the other schemes in terms of battery life, time delay, and energy consumption, with 75% accuracy of the trained deep neural network. The achieved decrease in total energy consumption of mobile terminal through DOTP is 45.74%, 36.69%, and 30.59% as compared to total offloading, partial offloading, and local offloading schemes, respectively

    Multi-user Computation Partitioning for Latency Sensitive Mobile Cloud Applications

    Get PDF
    Elastic partitioning of computations between mobile devices and cloud is an important and challenging research topic for mobile cloud computing. Existing works focus on the single-user computation partitioning, which aims to optimize the application completion time for one particular single user. These works assume that the cloud always has enough resources to execute the computations immediately when they are offloaded to the cloud. However, this assumption does not hold for large scale mobile cloud applications. In these applications, due to the competition for cloud resources among a large number of users, the offloaded computations may be executed with certain scheduling delay on the cloud. Single user partitioning that does not take into account the scheduling delay on the cloud may yield significant performance degradation. In this paper, we study, for the first time, Multi-user Computation Partitioning Problem (MCPP), which considers the partitioning of multiple usersā€™ computations together with the scheduling of offloaded computations on the cloud resources. Instead of pursuing the minimum application completion time for every single user, we aim to achieve minimum average completion time for all the users, based on the number of provisioned resources on the cloud. We show that MCPP is different from and more difficult than the classical job scheduling problems. We design an offline heuristic algorithm, namely SearchAdjust, to solve MCPP. We demonstrate through benchmarks that SearchAdjust outperforms both the single user partitioning approaches and classical job scheduling approaches by 10% on average in terms of application delay. Based on SearchAdjust, we also design an online algorithm for MCPP that can be easily deployed in practical systems. We validate the effectiveness of our online algorithm using real world load traces. Index Termsā€”mobile cloud computing; offloading; computation partitioning; job schedulin
    • ā€¦
    corecore