67 research outputs found

    Simulating fog and edge computing scenarios: an overview and research challenges

    Get PDF
    The fourth industrial revolution heralds a paradigm shift in how people, processes, things, data and networks communicate and connect with each other. Conventional computing infrastructures are struggling to satisfy dramatic growth in demand from a deluge of connected heterogeneous endpoints located at the edge of networks while, at the same time, meeting quality of service levels. The complexity of computing at the edge makes it increasingly difficult for infrastructure providers to plan for and provision resources to meet this demand. While simulation frameworks are used extensively in the modelling of cloud computing environments in order to test and validate technical solutions, they are at a nascent stage of development and adoption for fog and edge computing. This paper provides an overview of challenges posed by fog and edge computing in relation to simulation

    Evaluating system architectures for driving range estimation and charge planning for electric vehicles

    Get PDF
    Due to sparse charging infrastructure and short driving ranges, drivers of battery electric vehicles (BEVs) can experience range anxiety, which is the fear of stranding with an empty battery. To help eliminate range anxiety and make BEVs more attractive for customers, accurate range estimation methods need to be developed. In recent years, many publications have suggested machine learning algorithms as a fitting method to achieve accurate range estimations. However, these algorithms use a large amount of data and have high computational requirements. A traditional placement of the software within a vehicle\u27s electronic control unit could lead to high latencies and thus detrimental to user experience. But since modern vehicles are connected to a backend, where software modules can be implemented, high latencies can be prevented with intelligent distribution of the algorithm parts. On the other hand, communication between vehicle and backend can be slow or expensive. In this article, an intelligent deployment of a range estimation software based on ML is analyzed. We model hardware and software to enable performance evaluation in early stages of the development process. Based on simulations, different system architectures and module placements are then analyzed in terms of latency, network usage, energy usage, and cost. We show that a distributed system with cloud‐based module placement reduces the end‐to‐end latency significantly, when compared with a traditional vehicle‐based placement. Furthermore, we show that network usage is significantly reduced. This intelligent system enables the application of complex, but accurate range estimation with low latencies, resulting in an improved user experience, which enhances the practicality and acceptance of BEVs

    Sphere: Simulator of edge infrastructures for the optimization of performance and resources energy consumption

    Get PDF
    Edge computing constitutes a key paradigm to address the new requirements of areas such as smart cars, industry 4.0, and health care, where massive amounts of heterogeneous data from continuous geographically-distributed sources have to be processed and computed near real-time. To this end, new distributed infrastructures consisting on small computing clusters close to data sources, also known as Cloudlets have emerged. In order to evaluate the performance of these solutions we present Sphere, a simulation tool that enables researchers to establish various scenarios, including: (a) topology and orchestration model of the infrastructure; (b) incoming workload patterns; (c) resource-managing models; and (d) scheduling policies. Moreover, Sphere allows researchers to apply efficiency and performance policies both at infrastructure and cluster levels. The simulator presents the following benefits: (a) Evaluation of various orchestration models; (b) Analysis of resource-efficiency and performance strategies at Edge-infrastructure and cluster (Cloudlet/Cloud) level; (c) Execution of diverse workload generation patterns; (d) Evaluation of strategies for the infrastructure communication, as well as the impact on tasks completion time (makespan); and (e) Simulation of each cluster (Cloudlet/Cloud) independently, including resource-managing, scheduling and resource-efficiency models. Finally, we performed a deep evaluation based on realistic Edge-Computing use cases. The results of the experiments confirm that it is a performant and reliable tool for the analysis of orchestration, graph-resolving, energy-efficiency, resource-managing and scheduling strategies in Edge-computing environments.Ministerio de Ciencia, Innovación y Universidades RTI2018-098062-A-I0

    Performance and efficiency optimization of multi-layer IoT edge architecture

    Get PDF
    Abstract. Internet of Things (IoT) has become a backbone technology that connects together various devices with diverse capabilities. It is a technology, which enables ubiquitously available digital services for end-users. IoT applications for mission-critical scenarios need strict performance indicators such as of latency, scalability, security and privacy. To fulfil these requirements, IoT also requires support from relevant enabling technologies, such as cloud, edge, virtualization and fifth generation mobile communication (5G) technologies. For Latency-critical applications and services, long routes between the traditional cloud server and end-devices (sensors /actuators) is not a feasible approach for computing at these data centres, although these traditional clouds provide very high computational and storage for current IoT system. MEC model can be used to overcome this challenge, which brings the CC computational capacity within or next on the access network base stations. However, the capacity to perform the most critical processes at the local network layer is often necessary to cope with the access network issues. Therefore, this thesis compares the two existing IoT models such as traditional cloud-IoT model, a MEC-based edge-cloud-IoT model, with proposed local edge-cloud-IoT model with respect to their performance and efficiency, using iFogSim simulator. The results consolidate our research team’s previous findings that utilizing the three-tier edge-IoT architecture, capable of optimally utilizing the computational capacity of each of the three tiers, is an effective measure to reduce energy consumption, improve end-to-end latency and minimize operational costs in latency-critical It applications
    corecore