3,568 research outputs found
Neuromorphic AI Empowered Root Cause Analysis of Faults in Emerging Networks
Mobile cellular network operators spend nearly a quarter of their revenue on
network maintenance and management. A significant portion of that budget is
spent on resolving faults diagnosed in the system that disrupt or degrade
cellular services. Historically, the operations to detect, diagnose and resolve
issues were carried out by human experts. However, with diversifying cell
types, increased complexity and growing cell density, this methodology is
becoming less viable, both technically and financially. To cope with this
problem, in recent years, research on self-healing solutions has gained
significant momentum. One of the most desirable features of the self-healing
paradigm is automated fault diagnosis. While several fault detection and
diagnosis machine learning models have been proposed recently, these schemes
have one common tenancy of relying on human expert contribution for fault
diagnosis and prediction in one way or another. In this paper, we propose an
AI-based fault diagnosis solution that offers a key step towards a completely
automated self-healing system without requiring human expert input. The
proposed solution leverages Random Forests classifier, Convolutional Neural
Network and neuromorphic based deep learning model which uses RSRP map images
of faults generated. We compare the performance of the proposed solution
against state-of-the-art solution in literature that mostly use Naive Bayes
models, while considering seven different fault types. Results show that
neuromorphic computing model achieves high classification accuracy as compared
to the other models even with relatively small training dat
Continuous maintenance and the future – Foundations and technological challenges
High value and long life products require continuous maintenance throughout their life cycle to achieve required performance with optimum through-life cost. This paper presents foundations and technologies required to offer the maintenance service. Component and system level degradation science, assessment and modelling along with life cycle ‘big data’ analytics are the two most important knowledge and skill base required for the continuous maintenance. Advanced computing and visualisation technologies will improve efficiency of the maintenance and reduce through-life cost of the product. Future of continuous maintenance within the Industry 4.0 context also identifies the role of IoT, standards and cyber security
Design-time Models for Resiliency
Resiliency in process-aware information systems is based on the availability of recovery flows and alternative data for coping with missing data. In this paper, we discuss an approach to process and information modeling to support the specification of recovery flows and alternative data. In particular, we focus on processes using sensor data from different sources. The proposed model can be adopted to specify resiliency levels of information systems, based on event-based and temporal constraints
Riding out of the storm: How to deal with the complexity of grid and cloud management
Over the last decade, Grid computing paved the way for a new level of large scale distributed systems. This infrastructure made it possible to securely and reliably take advantage of widely separated computational resources that are part of several different organizations. Resources can be incorporated to the Grid, building a theoretical virtual supercomputer. In time, cloud computing emerged as a new type of large scale distributed system, inheriting and expanding the expertise and knowledge that have been obtained so far. Some of the main characteristics of Grids naturally evolved into clouds, others were modified and adapted and others were simply discarded or postponed. Regardless of these technical specifics, both Grids and clouds together can be considered as one of the most important advances in large scale distributed computing of the past ten years; however, this step in distributed computing has came along with a completely new level of complexity. Grid and cloud management mechanisms play a key role, and correct analysis and understanding of the system behavior are needed. Large scale distributed systems must be able to self-manage, incorporating autonomic features capable of controlling and optimizing all resources and services. Traditional distributed computing management mechanisms analyze each resource separately and adjust specific parameters of each one of them. When trying to adapt the same procedures to Grid and cloud computing, the vast complexity of these systems can make this task extremely complicated. But large scale distributed systems complexity could only be a matter of perspective. It could be possible to understand the Grid or cloud behavior as a single entity, instead of a set of resources. This abstraction could provide a different understanding of the system, describing large scale behavior and global events that probably would not be detected analyzing each resource separately. In this work we define a theoretical framework that combines both ideas, multiple resources and single entity, to develop large scale distributed systems management techniques aimed at system performance optimization, increased dependability and Quality of Service (QoS). The resulting synergy could be the key 350 J. Montes et al. to address the most important difficulties of Grid and cloud management
A model-based approach for automatic recovery from memory leaks in enterprise applications
Large-scale distributed computing systems such as data centers are hosted on heterogeneous and networked servers that execute in a dynamic and uncertain operating environment, caused by factors such as time-varying user workload and various failures. Therefore, achieving stringent quality-of-service goals is a challenging task, requiring a comprehensive approach to performance control, fault diagnosis, and failure recovery. This work presents a model-based approach for fault management, which integrates limited lookahead control (LLC), diagnosis, and fault-tolerance concepts that: (1) enables systems to adapt to environment variations, (2) maintains the availability and reliability of the system, (3) facilitates system recovery from failures. We focused on memory leak errors in this thesis. A characterization function is designed to detect memory leaks. Then, a LLC is applied to enable the computing system to adapt efficiently to variations in the workload, and to enable the system recover from memory leaks and maintain functionality
NASA Capability Roadmaps Executive Summary
This document is the result of eight months of hard work and dedication from NASA, industry, other government agencies, and academic experts from across the nation. It provides a summary of the capabilities necessary to execute the Vision for Space Exploration and the key architecture decisions that drive the direction for those capabilities. This report is being provided to the Exploration Systems Architecture Study (ESAS) team for consideration in development of an architecture approach and investment strategy to support NASA future mission, programs and budget requests. In addition, it will be an excellent reference for NASA's strategic planning. A more detailed set of roadmaps at the technology and sub-capability levels are available on CD. These detailed products include key driving assumptions, capability maturation assessments, and technology and capability development roadmaps
06031 Abstracts Collection -- Organic Computing -- Controlled Emergence
Organic Computing has emerged recently as a challenging vision for
future information processing systems, based on the insight that we
will soon be surrounded by large collections of autonomous systems
equipped with sensors and actuators to be aware of their environment,
to communicate freely, and to organize themselves in order to perform
the actions and services required. Organic Computing Systems will
adapt dynamically to the current conditions of its environment, they
will be self-organizing, self-configuring, self-healing,
self-protecting, self-explaining, and context-aware.
From 15.01.06 to 20.01.06, the Dagstuhl Seminar 06031 ``Organic
Computing -- Controlled Emergence\u27\u27 was held in the International
Conference and Research Center (IBFI), Schloss Dagstuhl.
The seminar was characterized by the very constructive search for
common ground between engineering and natural sciences, between
informatics on the one hand and biology, neuroscience, and chemistry
on the other. The common denominator was the objective to build
practically usable self-organizing and emergent systems or their
components.
An indicator for the practical orientation of the seminar was the
large number of OC application systems, envisioned or already under
implementation, such as the Internet, robotics, wireless sensor
networks, traffic control, computer vision, organic systems on chip,
an adaptive and self-organizing room with intelligent sensors or
reconfigurable guiding systems for smart office buildings. The
application orientation was also apparent by the large number of
methods and tools presented during the seminar, which might be used as
building blocks for OC systems, such as an evolutionary design
methodology, OC architectures, especially several implementations of
observer/controller structures, measures and measurement tools for
emergence and complexity, assertion-based methods to control
self-organization, wrappings, a software methodology to build
reflective systems, and components for OC middleware.
Organic Computing is clearly oriented towards applications but is
augmented at the same time by more theoretical bio-inspired and
nature-inspired work, such as chemical computing, theory of complex
systems and non-linear dynamics, control mechanisms in insect swarms,
homeostatic mechanisms in the brain, a quantitative approach to
robustness, abstraction and instantiation as a central metaphor for
understanding complex systems.
Compared to its beginnings, Organic Computing is coming of age. The OC
vision is increasingly padded with meaningful applications and usable
tools, but the path towards full OC systems is still complex. There is
progress in a more scientific understanding of emergent processes. In
the future, we must understand more clearly how to open the
configuration space of technical systems for on-line
modification. Finally, we must make sure that the human user remains
in full control while allowing the systems to optimize
- …