5 research outputs found

    Energy-efficient deployment of edge dataenters for mobile clouds in sustainable iot

    Full text link
    © 2013 IEEE. Achieving quick responses with limited energy consumption in mobile cloud computing is an active area of research. The energy consumption increases when a user's request (task) runs in the local mobile device instead of executing in the cloud. Whereas, latency become an issue when the task executes in the cloud environment instead of the mobile device. Therefore, a tradeoff between energy consumption and latency is required in building sustainable Internet of Things (IoT), and for that, we have introduced a middle layer named an edge computing layer to avoid latency in IoT. There are several real-time applications, such as smart city and smart health, where mobile users upload their tasks into the cloud or execute locally. We have intended to minimize the energy consumption of a mobile device as well as the energy consumption of the cloud system while meeting a task's deadline, by offloading the task to the edge datacenter or cloud. This paper proposes an adaptive technique to optimize both parameters, i.e., energy consumption and latency by offloading the task and also by selecting the appropriate virtual machine for the execution of the task. In the proposed technique, if the specified edge datacenter is unable to provide resources, then the user's request will be sent to the cloud system. Finally, the proposed technique is evaluated using a real-world scenario to measure its performance and efficiency. The simulation results show that the total energy consumption and execution time decrease after introducing an edge datacenters as a middle layer

    Minimization of Energy and Service Latency Computation Offloading using Neural Network in 5G NOMA System

    Get PDF
    The future Internet of Things (IoT) era is anticipated to support computation-intensive and time-critical applications using edge computing for mobile (MEC), which is regarded as promising technique. However, the transmitting uplink performance will be highly impacted by the hostile wireless channel, the low bandwidth, and the low transmission power of IoT devices. Using edge computing for mobile (MEC) to offload tasks becomes a crucial technology to reduce service latency for computation-intensive applications and reduce the computational workloads of mobile devices. Under the restrictions of computation latency and cloud computing capacity, our goal is to reduce the overall energy consumption of all users, including transmission energy and local computation energy. In this article, the Deep Q Network Algorithm (DQNA) to deal with the data rates with respect to the user base in different time slots of 5G NOMA network. The DQNA is optimized by considering more number of cell structures like 2, 4, 6 and 8. Therefore, the DQNA provides the optimal distribution of power among all 3 users in the 5G network, which gives the increased data rates. The existing various power distribution algorithms like frequent pattern (FP), weighted least squares mean error weighted least squares mean error (WLSME), and Random Power and Maximal Power allocation are used to justify the proposed DQNA technique. The proposed technique which gives 81.6% more the data rates when increased the cell structure to 8. Thus 25% more in comparison to other algorithms like FP, WLSME Random Power and Maximal Power allocation

    Green and secure computation offloading for cache-enabled IoT networks

    Get PDF
    The ever-increasing number of diverse and computation-intensive Internet of things (IoT) applications is bringing phenomenal growth in global Internet traffic. Mobile devices with limited resource capacity (i.e., computation and storage resources) and battery lifetime are experiencing technical challenges to satisfy the task requirements. Mobile edge computing (MEC) integrated with IoT applications offloads computation-intensive tasks to the MEC servers at the network edge. This technique shows remarkable potential in reducing energy consumption and delay. Furthermore, caching popular task input data at the edge servers reduces duplicate content transmission, which eventually saves associated energy and time. However, the offloaded tasks are exposed to multiple users and vulnerable to malicious attacks and eavesdropping. Therefore, the assignment of security services to the offloaded tasks is a major requirement to ensure confidentiality and privacy. In this article, we propose a green and secure MEC technique combining caching, cooperative task offloading, and security service assignment for IoT networks. The study not only investigates the synergy between energy and security issues, but also offloads IoT tasks to the edge servers without violating delay requirements. A resource-constrained optimization model is formulated, which minimizes the overall cost combining energy consumption and probable security-breach cost. We also develop a two-stage heuristic algorithm and find an acceptable solution in polynomial time. Simulation results prove that the proposed technique achieves notable improvement over other existing strategies

    Cooperative scheduling and load balancing techniques in fog and edge computing

    Get PDF
    Fog and Edge Computing are two models that reached maturity in the last decade. Today, they are two solid concepts and plenty of literature tried to develop them. Also corroborated by the development of technologies, like for example 5G, they can now be considered de facto standards when building low and ultra-low latency applications, privacy-oriented solutions, industry 4.0 and smart city infrastructures. The common trait of Fog and Edge computing environments regards their inherent distributed and heterogeneous nature where the multiple (Fog or Edge) nodes are able to interact with each other with the essential purpose of pre-processing data gathered by the uncountable number of sensors to which they are connected to, even by running significant ML models and relying upon specific processors (TPU). However, nodes are often placed in a geographic domain, like a smart city, and the dynamic of the traffic during the day may cause some nodes to be overwhelmed by requests while others instead may become completely idle. To achieve the optimal usage of the system and also to guarantee the best possible QoS across all the users connected to the Fog or Edge nodes, the need to design load balancing and scheduling algorithms arises. In particular, a reasonable solution is to enable nodes to cooperate. This capability represents the main objective of this thesis, which is the design of fully distributed algorithms and solutions whose purpose is the one of balancing the load across all the nodes, also by following, if possible, QoS requirements in terms of latency or imposing constraints in terms of power consumption when the nodes are powered by green energy sources. Unfortunately, when a central orchestrator is missing, a crucial element which makes the design of such algorithms difficult is that nodes need to know the state of the others in order to make the best possible scheduling decision. However, it is not possible to retrieve the state without introducing further latency during the service of the request. Furthermore, the retrieved information about the state is always old, and as a consequence, the decision is always relying on imprecise data. In this thesis, the problem is circumvented in two main ways. The first one considers randomised algorithms which avoid probing all of the neighbour nodes in favour of at maximum two nodes picked at random. This is proven to bring an exponential improvement in performance with respect to the probe of a single node. The second approach, instead, considers Reinforcement Learning as a technique for inferring the state of the other nodes thanks to the reward received by the agents when requests are forwarded. Moreover, the thesis will also focus on the energy aspect of the Edge devices. In particular, will be analysed a scenario of Green Edge Computing, where devices are powered only by Photovoltaic Panels and a scenario of mobile offloading targeting ML image inference applications. Lastly, a final glance will be given at a series of infrastructural studies, which will give the foundations for implementing the proposed algorithms on real devices, in particular, Single Board Computers (SBCs). There will be presented a structural scheme of a testbed of Raspberry Pi boards, and a fully-fledged framework called ``P2PFaaS'' which allows the implementation of load balancing and scheduling algorithms based on the Function-as-a-Service (FaaS) paradigm

    Energy-Efficient Deployment of Edge Dataenters for Mobile Clouds in Sustainable IoT

    No full text
    corecore