3,568 research outputs found

    Neuromorphic AI Empowered Root Cause Analysis of Faults in Emerging Networks

    Full text link
    Mobile cellular network operators spend nearly a quarter of their revenue on network maintenance and management. A significant portion of that budget is spent on resolving faults diagnosed in the system that disrupt or degrade cellular services. Historically, the operations to detect, diagnose and resolve issues were carried out by human experts. However, with diversifying cell types, increased complexity and growing cell density, this methodology is becoming less viable, both technically and financially. To cope with this problem, in recent years, research on self-healing solutions has gained significant momentum. One of the most desirable features of the self-healing paradigm is automated fault diagnosis. While several fault detection and diagnosis machine learning models have been proposed recently, these schemes have one common tenancy of relying on human expert contribution for fault diagnosis and prediction in one way or another. In this paper, we propose an AI-based fault diagnosis solution that offers a key step towards a completely automated self-healing system without requiring human expert input. The proposed solution leverages Random Forests classifier, Convolutional Neural Network and neuromorphic based deep learning model which uses RSRP map images of faults generated. We compare the performance of the proposed solution against state-of-the-art solution in literature that mostly use Naive Bayes models, while considering seven different fault types. Results show that neuromorphic computing model achieves high classification accuracy as compared to the other models even with relatively small training dat

    Continuous maintenance and the future – Foundations and technological challenges

    Get PDF
    High value and long life products require continuous maintenance throughout their life cycle to achieve required performance with optimum through-life cost. This paper presents foundations and technologies required to offer the maintenance service. Component and system level degradation science, assessment and modelling along with life cycle ‘big data’ analytics are the two most important knowledge and skill base required for the continuous maintenance. Advanced computing and visualisation technologies will improve efficiency of the maintenance and reduce through-life cost of the product. Future of continuous maintenance within the Industry 4.0 context also identifies the role of IoT, standards and cyber security

    Design-time Models for Resiliency

    Get PDF
    Resiliency in process-aware information systems is based on the availability of recovery flows and alternative data for coping with missing data. In this paper, we discuss an approach to process and information modeling to support the specification of recovery flows and alternative data. In particular, we focus on processes using sensor data from different sources. The proposed model can be adopted to specify resiliency levels of information systems, based on event-based and temporal constraints

    Riding out of the storm: How to deal with the complexity of grid and cloud management

    Get PDF
    Over the last decade, Grid computing paved the way for a new level of large scale distributed systems. This infrastructure made it possible to securely and reliably take advantage of widely separated computational resources that are part of several different organizations. Resources can be incorporated to the Grid, building a theoretical virtual supercomputer. In time, cloud computing emerged as a new type of large scale distributed system, inheriting and expanding the expertise and knowledge that have been obtained so far. Some of the main characteristics of Grids naturally evolved into clouds, others were modified and adapted and others were simply discarded or postponed. Regardless of these technical specifics, both Grids and clouds together can be considered as one of the most important advances in large scale distributed computing of the past ten years; however, this step in distributed computing has came along with a completely new level of complexity. Grid and cloud management mechanisms play a key role, and correct analysis and understanding of the system behavior are needed. Large scale distributed systems must be able to self-manage, incorporating autonomic features capable of controlling and optimizing all resources and services. Traditional distributed computing management mechanisms analyze each resource separately and adjust specific parameters of each one of them. When trying to adapt the same procedures to Grid and cloud computing, the vast complexity of these systems can make this task extremely complicated. But large scale distributed systems complexity could only be a matter of perspective. It could be possible to understand the Grid or cloud behavior as a single entity, instead of a set of resources. This abstraction could provide a different understanding of the system, describing large scale behavior and global events that probably would not be detected analyzing each resource separately. In this work we define a theoretical framework that combines both ideas, multiple resources and single entity, to develop large scale distributed systems management techniques aimed at system performance optimization, increased dependability and Quality of Service (QoS). The resulting synergy could be the key 350 J. Montes et al. to address the most important difficulties of Grid and cloud management

    Autonomic Analytics

    Get PDF

    A model-based approach for automatic recovery from memory leaks in enterprise applications

    Get PDF
    Large-scale distributed computing systems such as data centers are hosted on heterogeneous and networked servers that execute in a dynamic and uncertain operating environment, caused by factors such as time-varying user workload and various failures. Therefore, achieving stringent quality-of-service goals is a challenging task, requiring a comprehensive approach to performance control, fault diagnosis, and failure recovery. This work presents a model-based approach for fault management, which integrates limited lookahead control (LLC), diagnosis, and fault-tolerance concepts that: (1) enables systems to adapt to environment variations, (2) maintains the availability and reliability of the system, (3) facilitates system recovery from failures. We focused on memory leak errors in this thesis. A characterization function is designed to detect memory leaks. Then, a LLC is applied to enable the computing system to adapt efficiently to variations in the workload, and to enable the system recover from memory leaks and maintain functionality

    NASA Capability Roadmaps Executive Summary

    Get PDF
    This document is the result of eight months of hard work and dedication from NASA, industry, other government agencies, and academic experts from across the nation. It provides a summary of the capabilities necessary to execute the Vision for Space Exploration and the key architecture decisions that drive the direction for those capabilities. This report is being provided to the Exploration Systems Architecture Study (ESAS) team for consideration in development of an architecture approach and investment strategy to support NASA future mission, programs and budget requests. In addition, it will be an excellent reference for NASA's strategic planning. A more detailed set of roadmaps at the technology and sub-capability levels are available on CD. These detailed products include key driving assumptions, capability maturation assessments, and technology and capability development roadmaps

    06031 Abstracts Collection -- Organic Computing -- Controlled Emergence

    Get PDF
    Organic Computing has emerged recently as a challenging vision for future information processing systems, based on the insight that we will soon be surrounded by large collections of autonomous systems equipped with sensors and actuators to be aware of their environment, to communicate freely, and to organize themselves in order to perform the actions and services required. Organic Computing Systems will adapt dynamically to the current conditions of its environment, they will be self-organizing, self-configuring, self-healing, self-protecting, self-explaining, and context-aware. From 15.01.06 to 20.01.06, the Dagstuhl Seminar 06031 ``Organic Computing -- Controlled Emergence\u27\u27 was held in the International Conference and Research Center (IBFI), Schloss Dagstuhl. The seminar was characterized by the very constructive search for common ground between engineering and natural sciences, between informatics on the one hand and biology, neuroscience, and chemistry on the other. The common denominator was the objective to build practically usable self-organizing and emergent systems or their components. An indicator for the practical orientation of the seminar was the large number of OC application systems, envisioned or already under implementation, such as the Internet, robotics, wireless sensor networks, traffic control, computer vision, organic systems on chip, an adaptive and self-organizing room with intelligent sensors or reconfigurable guiding systems for smart office buildings. The application orientation was also apparent by the large number of methods and tools presented during the seminar, which might be used as building blocks for OC systems, such as an evolutionary design methodology, OC architectures, especially several implementations of observer/controller structures, measures and measurement tools for emergence and complexity, assertion-based methods to control self-organization, wrappings, a software methodology to build reflective systems, and components for OC middleware. Organic Computing is clearly oriented towards applications but is augmented at the same time by more theoretical bio-inspired and nature-inspired work, such as chemical computing, theory of complex systems and non-linear dynamics, control mechanisms in insect swarms, homeostatic mechanisms in the brain, a quantitative approach to robustness, abstraction and instantiation as a central metaphor for understanding complex systems. Compared to its beginnings, Organic Computing is coming of age. The OC vision is increasingly padded with meaningful applications and usable tools, but the path towards full OC systems is still complex. There is progress in a more scientific understanding of emergent processes. In the future, we must understand more clearly how to open the configuration space of technical systems for on-line modification. Finally, we must make sure that the human user remains in full control while allowing the systems to optimize
    • …
    corecore