12 research outputs found

    Edge Computing in Low-Earth Orbit -- What Could Possibly Go Wrong?

    Full text link
    Large low-Earth orbit (LEO) satellite networks are being built to provide low-latency broadband Internet access to a global subscriber base. In addition to network transmissions, researchers have proposed embedding compute resources in satellites to support LEO edge computing. To make software systems ready for the LEO edge, they need to be adapted for its unique execution environment, e.g., to support handovers in face of satellite mobility. So far, research around LEO edge software systems has focused on the predictable behavior of satellite networks, such as orbital movements. Additionally, we must also consider failure patterns, e.g., effects of radiation on compute hardware in space. In this paper, we present a taxonomy of failures that may occur in LEO edge computing and how they could affect software systems. From there, we derive considerations for LEO edge software systems and lay out avenues for future work.Comment: 1st Workshop on Low Earth Orbit Networking and Communication (LEO-NET '23

    Creating Intelligent Computational Edge through Semantic Mediation

    Get PDF
    This research proposes semantic mediation based on reasoning and the first order logic for mediating the best possible configuration of Computational Edge, relevant for software applications which may benefit for running computations with proximity to their data sources. The mediation considers the context in which these applications exist and exploits the semantic of that context for decision making on where computational elements should reside and which data they should use. The application of semantic mediation could address the initiative to accommodate algorithms from predictive and learning technologies, push AI towards computational edges and potentially contribute towards creating a computing continuum

    Motivations and Challenges in Unmanaged Edge Computing

    Get PDF

    DRAGON: Decentralized fault tolerance in edge federations

    Get PDF
    Edge Federation is a new computing paradigm that seamlessly interconnects the resources of multiple edge service providers. A key challenge in such systems is the deployment of latency-critical and AI based resource-intensive applications in constrained devices. To address this challenge, we propose a novel memory-efficient deep learning based model, namely generative optimization networks (GON). Unlike GANs, GONs use a single network to both discriminate input and generate samples, significantly reducing their memory footprint. Leveraging the low memory footprint of GONs, we propose a decentralized fault-tolerance method called DRAGON that runs simulations (as per a digital modeling twin) to quickly predict and optimize the performance of the edge federation. Extensive experiments with real-world edge computing benchmarks on multiple Raspberry-Pi based federated edge configurations show that DRAGON can outperform the baseline methods in fault-detection and Quality of Service (QoS) metrics. Specifically, the proposed method gives higher F1 scores for fault-detection than the best deep learning (DL) method, while consuming lower memory than the heuristic methods. This allows for improvement in energy consumption, response time and service level agreement violations by up to 74, 63 and 82 percent, respectively

    A Pattern-Language for Self-Healing Internet-of-Things Systems

    Get PDF
    Internet-of-Things systems are assemblies of highly-distributed and heterogeneous parts that, in orchestration, work to provide valuable services to end-users in many scenarios. These systems depend on the correct operation of sensors, actuators, and third-party services, and the failure of a single one can hinder the proper functioning of the whole system, making error detection and recovery of paramount importance, but often overlooked. By drawing inspiration from other research areas, such as cloud, embedded, and mission-critical systems, we present a set of patterns for self-healing IoT systems. We discuss how their implementation can improve system reliability by providing error detection, error recovery, and health mechanisms maintenance. (c) 2020 ACM

    Edge AI for Internet of Energy: Challenges and Perspectives

    Full text link
    The digital landscape of the Internet of Energy (IoE) is on the brink of a revolutionary transformation with the integration of edge Artificial Intelligence (AI). This comprehensive review elucidates the promise and potential that edge AI holds for reshaping the IoE ecosystem. Commencing with a meticulously curated research methodology, the article delves into the myriad of edge AI techniques specifically tailored for IoE. The myriad benefits, spanning from reduced latency and real-time analytics to the pivotal aspects of information security, scalability, and cost-efficiency, underscore the indispensability of edge AI in modern IoE frameworks. As the narrative progresses, readers are acquainted with pragmatic applications and techniques, highlighting on-device computation, secure private inference methods, and the avant-garde paradigms of AI training on the edge. A critical analysis follows, offering a deep dive into the present challenges including security concerns, computational hurdles, and standardization issues. However, as the horizon of technology ever expands, the review culminates in a forward-looking perspective, envisaging the future symbiosis of 5G networks, federated edge AI, deep reinforcement learning, and more, painting a vibrant panorama of what the future beholds. For anyone vested in the domains of IoE and AI, this review offers both a foundation and a visionary lens, bridging the present realities with future possibilities

    ACE: toward application-centric, edge-cloud, collaborative intelligence

    Get PDF
    Constructing a unified platform for the scalable, reliable, robust, and efficient development and deployment of ECCI applications

    Design, development and evaluation of the ruggedized edge computing node (RECON)

    Get PDF
    The increased quality and quantity of sensors provide an ever-increasing capability to collect large quantities of high-quality data in the field. Research devoted to translating that data is progressing rapidly; however, translating field data into usable information can require high performance computing capabilities. While high performance computing (HPC) resources are available in centralized facilities, bandwidth, latency, security and other limitations inherent to edge location in field sensor applications may prevent HPC resources from being used in a timely fashion necessary for potential United States Army Corps of Engineers (USACE) field applications. To address these limitations, the design requirements for RECON are established and derived from a review of edge computing, in order to develop and evaluate a novel high-power, field-deployable HPC platform capable of operating in austere environments at the edge
    corecore