109,143 research outputs found

    Foggy clouds and cloudy fogs: a real need for coordinated management of fog-to-cloud computing systems

    Get PDF
    The recent advances in cloud services technology are fueling a plethora of information technology innovation, including networking, storage, and computing. Today, various flavors have evolved of IoT, cloud computing, and so-called fog computing, a concept referring to capabilities of edge devices and users' clients to compute, store, and exchange data among each other and with the cloud. Although the rapid pace of this evolution was not easily foreseeable, today each piece of it facilitates and enables the deployment of what we commonly refer to as a smart scenario, including smart cities, smart transportation, and smart homes. As most current cloud, fog, and network services run simultaneously in each scenario, we observe that we are at the dawn of what may be the next big step in the cloud computing and networking evolution, whereby services might be executed at the network edge, both in parallel and in a coordinated fashion, as well as supported by the unstoppable technology evolution. As edge devices become richer in functionality and smarter, embedding capacities such as storage or processing, as well as new functionalities, such as decision making, data collection, forwarding, and sharing, a real need is emerging for coordinated management of fog-to-cloud (F2C) computing systems. This article introduces a layered F2C architecture, its benefits and strengths, as well as the arising open and research challenges, making the case for the real need for their coordinated management. Our architecture, the illustrative use case presented, and a comparative performance analysis, albeit conceptual, all clearly show the way forward toward a new IoT scenario with a set of existing and unforeseen services provided on highly distributed and dynamic compute, storage, and networking resources, bringing together heterogeneous and commodity edge devices, emerging fogs, as well as conventional clouds.Peer ReviewedPostprint (author's final draft

    Single-Board-Computer Clusters for Cloudlet Computing in Internet of Things

    Get PDF
    The number of connected sensors and devices is expected to increase to billions in the near future. However, centralised cloud-computing data centres present various challenges to meet the requirements inherent to Internet of Things (IoT) workloads, such as low latency, high throughput and bandwidth constraints. Edge computing is becoming the standard computing paradigm for latency-sensitive real-time IoT workloads, since it addresses the aforementioned limitations related to centralised cloud-computing models. Such a paradigm relies on bringing computation close to the source of data, which presents serious operational challenges for large-scale cloud-computing providers. In this work, we present an architecture composed of low-cost Single-Board-Computer clusters near to data sources, and centralised cloud-computing data centres. The proposed cost-efficient model may be employed as an alternative to fog computing to meet real-time IoT workload requirements while keeping scalability. We include an extensive empirical analysis to assess the suitability of single-board-computer clusters as cost-effective edge-computing micro data centres. Additionally, we compare the proposed architecture with traditional cloudlet and cloud architectures, and evaluate them through extensive simulation. We finally show that acquisition costs can be drastically reduced while keeping performance levels in data-intensive IoT use cases.Ministerio de Economía y Competitividad TIN2017-82113-C2-1-RMinisterio de Economía y Competitividad RTI2018-098062-A-I00European Union’s Horizon 2020 No. 754489Science Foundation Ireland grant 13/RC/209

    Implementing and evaluating an ICON orchestrator

    Get PDF
    The cloud computing paradigm has risen, during the last 20 years, to the task of bringing powerful computational services to the masses. Centralizing the computer hardware to a few large data centers has brought large monetary savings, but at the cost of a greater geographical distance between the server and the client. As a new generation of thin clients have emerged, e.g. smartphones and IoT-devices, the larger latencies induced by these greater distances, can limit the applications that could benefit from using the vast resources available in cloud computing. Not long after the explosive growth of cloud computing, a new paradigm, edge computing has risen. Edge computing aims at bringing the resources generally found in cloud computing closer to the edge where many of the end-users, clients and data producers reside. In this thesis, I will present the edge computing concept as well as the technologies enabling it. Furthermore I will show a few edge computing concepts and architectures, including multi- access edge computing (MEC), Fog computing and intelligent containers (ICON). Finally, I will also present a new edge-orchestrator, the ICON Python Orchestrator (IPO), that enables intelligent containers to migrate closer to the users. The ICON Python orchestrator tests the feasibility of the ICON concept and provides per- formance measurements that can be compared to other contemporary edge computing im- plementations. In this thesis, I will present the IPO architecture design including challenges encountered during the implementation phase and solutions to specific problems. I will also show the testing and validation setup. By using the artificial testing and validation network, client migration speeds were measured using three different cases - redirection, cache hot ICON migration and cache cold ICON migration. While there is room for improvements, the migration speeds measured are on par with other edge computing implementations

    Optimal association of mobile users to multi-access edge computing resources

    Get PDF
    Multi-access edge computing (MEC) plays a key role in fifth-generation (5G) networks in bringing cloud functionalities at the edge of the radio access network, in close proximity to mobile users. In this paper we focus on mobile-edge computation offloading, a way to transfer heavy demanding, and latency-critical applications from mobile handsets to close-located MEC servers, in order to reduce latency and/or energy consumption. Our goal is to provide an optimal strategy to associate mobile users to access points (AP) and MEC hosts, while contextually optimizing the allocation of radio and computational resources to each user, with the objective of minimizing the overall user transmit power under latency constraints incorporating both communication and computation times. The overall problem is a mixed-binary problem. To overcome its inherent computational complexity, we propose two alternative strategies: i) a method based on successive convex approximation (SCA) techniques, proven to converge to local optimal solutions; ii) an approach hinging on matching theory, based on formulating the assignment problem as a matching game

    Hyperprofile-based Computation Offloading for Mobile Edge Networks

    Full text link
    In recent studies, researchers have developed various computation offloading frameworks for bringing cloud services closer to the user via edge networks. Specifically, an edge device needs to offload computationally intensive tasks because of energy and processing constraints. These constraints present the challenge of identifying which edge nodes should receive tasks to reduce overall resource consumption. We propose a unique solution to this problem which incorporates elements from Knowledge-Defined Networking (KDN) to make intelligent predictions about offloading costs based on historical data. Each server instance can be represented in a multidimensional feature space where each dimension corresponds to a predicted metric. We compute features for a "hyperprofile" and position nodes based on the predicted costs of offloading a particular task. We then perform a k-Nearest Neighbor (kNN) query within the hyperprofile to select nodes for offloading computation. This paper formalizes our hyperprofile-based solution and explores the viability of using machine learning (ML) techniques to predict metrics useful for computation offloading. We also investigate the effects of using different distance metrics for the queries. Our results show various network metrics can be modeled accurately with regression, and there are circumstances where kNN queries using Euclidean distance as opposed to rectilinear distance is more favorable.Comment: 5 pages, NSF REU Site publicatio

    EDGE-CoT: next generation cloud computing and its impact on business

    Get PDF
    Purpose – The main objective of this paper is to analyze the potential impact of future cloud computing trends on business, from the perspective of specialists in the area. Design/ methodology/ approach - Qualitative approach that includes literature review and nine semi-structured interviews with proclaimed influencers and global thought leaders in cloud computing, highlighting Jeff Barr, Vice President of Amazon Web Services. Findings -5G networks will enable the emergence of the Edge-CoT architecture, that will consequently drive the increased application of Artificial Intelligence/ Machine Learning (AI/ML) and Robotics. The combination of Edge-CoT, Robotics and AI/ML triggers the development of Smart Cities and Industry 4.0. Simultaneously, Cloud alone will benefit of increased connectivity and will be the preferred business architecture comparing to EdgeCoT. New industries and businesses will result from the Edge-CoT, and the existing companies will benefit mainly from an improved customer experience. Major business challenges triggered by Edge-CoT include workforce re-skilling, promotion of the agile approach and a cultural shift towards risk-taking. Research limitations/implications - The research study was limited to the analysis of a selected set of cloud computing trends. Moreover, the data collection process was limited to 9 cloud experts, hindering a possible generalization. Originality/value – This study uses a qualitative approach to listen to market experts and cross with the theoretical findings to date, consequently bringing theory and practice closer together.Objetivo - O objetivo deste estudo consiste em analisar o potencial impacto das tendências futuras de cloud computing na gestão das empresas, a partir da visão de especialistas da área. Metodologia- Abordagem qualitativa que engloba revisão de literatura e nove entrevistas semiestruturadas com proclamados influencers e lideres globais em cloud computing, destacando-se Jeff Barr, o Vice-presidente da Amazon Web Services. Resultado - As redes 5G possibilitarão o surgimento da arquitetura Edge-CoT, que consequentemente impulsionará o aumento da aplicação de Inteligência Artificial (AI) e robótica. A combinação de Edge-CoT, Robótica e AI desencadeia o desenvolvimento de Smart Cities e Industry 4.0. Simultaneamente, a Cloud sozinha beneficiará do aumento da conectividade e será a arquitetura preferida comparativamente a Edge-CoT. Novos setores e negócios resultarão do Edge-CoT, e as empresas existentes beneficiarão principalmente de uma melhor experiência do cliente. Os principais desafios organizacionais desencadeados pelo Edge-CoT incluem a requalificação da força de trabalho, a adoção da abordagem agile e uma mudança cultural que estimule experimentos tecnológicos. Restrição da pesquisa - O processo de recolha de dados foi limitado a 9 especialistas em cloud computing, dificultando assim uma possível generalização. Originalidade/ Valor - Este estudo utiliza uma abordagem qualitativa para ouvir os especialistas do mercado e cruzar com os resultados teóricos até o momento, aproximando assim a teoria da prática

    Function-as-a-Service for the Cloud-to-Thing Continuum: A Systematic Mapping Study

    Get PDF
    Until recently, Internet of Things applications were mainly seen as a means to gather sensor data for further processing in the Cloud. Nowadays, with the advent of Edge and Fog Computing, digital services are dragged closer to the physical world, with data processing and storage tasks distributed across the whole Cloud-to-Thing continuum. Function-as-a-Service (FaaS) is gaining momentum as one of the promising programming models for such digital services. This work investigates the current research landscape of applying FaaS over the Cloud-to-Thing continuum. In particular, we investigate the support offered by existing FaaS platforms for the deployment, placement, orchestration, and execution of functions across the whole continuum using the Systematic Mapping Study methodology. We selected 33 primary studies and analyzed their data, bringing a broad view on the current research landscape in the area.acceptedVersio
    corecore