213,666 research outputs found
Modeling and Analyzing Adaptive User-Centric Systems in Real-Time Maude
Pervasive user-centric applications are systems which are meant to sense the
presence, mood, and intentions of users in order to optimize user comfort and
performance. Building such applications requires not only state-of-the art
techniques from artificial intelligence but also sound software engineering
methods for facilitating modular design, runtime adaptation and verification of
critical system requirements.
In this paper we focus on high-level design and analysis, and use the
algebraic rewriting language Real-Time Maude for specifying applications in a
real-time setting. We propose a generic component-based approach for modeling
pervasive user-centric systems and we show how to analyze and prove crucial
properties of the system architecture through model checking and simulation.
For proving time-dependent properties we use Metric Temporal Logic (MTL) and
present analysis algorithms for model checking two subclasses of MTL formulas:
time-bounded response and time-bounded safety MTL formulas. The underlying idea
is to extend the Real-Time Maude model with suitable clocks, to transform the
MTL formulas into LTL formulas over the extended specification, and then to use
the LTL model checker of Maude. It is shown that these analyses are sound and
complete for maximal time sampling. The approach is illustrated by a simple
adaptive advertising scenario in which an adaptive advertisement display can
react to actions of the users in front of the display.Comment: In Proceedings RTRTS 2010, arXiv:1009.398
Modeling Adaptive Middleware and Its Applications to Military Tactical Datalinks
Open systems solutions and techniques have become the de facto standard for achieving interoperability between disparate, large-scale, legacy software systems. A key technology among open systems solutions and techniques is middleware. Middleware, in general, is used to isolate applications from dependencies introduced by hardware, operating systems, and other low-level aspects of system architectures. While middleware approaches are or will be integrated into operational military systems, many open questions exist about the appropriate areas to applying middleware. Adaptive middleware is middleware that provides an application with a run-time adaptation strategy, based upon system-level interfaces and properties. Adaptive middleware is an example of an active applied research area. Adaptive middleware is being developed and applied to meet the ever-increasing challenges set forth by the next generation of mission-critical distributed real-time and embedded (DRE) systems. The driving force behind many next-generation DRE systems is the establishment of QoS requirements typically associated with workloads that vary dynamically. The Weapon System Open Architecture (WSOA), an adaptive middleware platform developed by Boeing, is modeled as a part of this research to determine the scalability of the architecture. The WSOA adaptive middleware was previously flight-tested with one tactical node, and the test results represent the performance baseline the architecture. The WSOA adaptive middleware is modeled with 1, 2, 4, 8 and 16 tactical nodes. The results of the modeling and simulation is that the WSOA adaptive middleware can achieve the performance baseline achieved during the original flight-test, in the cases of 1, 2, and 4 tactical nodes. In addition, the results of the modeling and simulation also demonstrate that the WSOA adaptive middleware cannot achiev
MIMO First and Second Order Discrete Sliding Mode Controls of Uncertain Linear Systems under Implementation Imprecisions
The performance of a conventional model-based controller significantly
depends on the accuracy of the modeled dynamics. The model of a plant's
dynamics is subjected to errors in estimating the numerical values of the
physical parameters, and variations over operating environment conditions and
time. These errors and variations in the parameters of a model are the major
sources of uncertainty within the controller structure. Digital implementation
of controller software on an actual electronic control unit (ECU) introduces
another layer of uncertainty at the controller inputs/outputs. The
implementation uncertainties are mostly due to data sampling and quantization
via the analog-to-digital conversion (ADC) unit. The failure to address the
model and ADC uncertainties during the early stages of a controller design
cycle results in a costly and time consuming verification and validation (V&V)
process. In this paper, new formulations of the first and second order discrete
sliding mode controllers (DSMC) are presented for a general class of uncertain
linear systems. The knowledge of the ADC imprecisions is incorporated into the
proposed DSMCs via an online ADC uncertainty prediction mechanism to improve
the controller robustness characteristics. Moreover, the DSMCs are equipped
with adaptation laws to remove two different types of modeling uncertainties
(multiplicative and additive) from the parameters of the linear system model.
The proposed adaptive DSMCs are evaluated on a DC motor speed control problem
in real-time using a processor-in-the-loop (PIL) setup with an actual ECU. The
results show that the proposed SISO and MIMO second order DSMCs improve the
conventional SISO first order DSMC tracking performance by 69% and 84%,
respectively. Moreover, the proposed adaptation mechanism is able to remove the
uncertainties in the model by up to 90%.Comment: 10 pages, 11 figures, ASME 2017 Dynamic Systems and Control
Conferenc
Prototype of Fault Adaptive Embedded Software for Large-Scale Real-Time Systems
This paper describes a comprehensive prototype of large-scale fault adaptive
embedded software developed for the proposed Fermilab BTeV high energy physics
experiment. Lightweight self-optimizing agents embedded within Level 1 of the
prototype are responsible for proactive and reactive monitoring and mitigation
based on specified layers of competence. The agents are self-protecting,
detecting cascading failures using a distributed approach. Adaptive,
reconfigurable, and mobile objects for reliablility are designed to be
self-configuring to adapt automatically to dynamically changing environments.
These objects provide a self-healing layer with the ability to discover,
diagnose, and react to discontinuities in real-time processing. A generic
modeling environment was developed to facilitate design and implementation of
hardware resource specifications, application data flow, and failure mitigation
strategies. Level 1 of the planned BTeV trigger system alone will consist of
2500 DSPs, so the number of components and intractable fault scenarios involved
make it impossible to design an `expert system' that applies traditional
centralized mitigative strategies based on rules capturing every possible
system state. Instead, a distributed reactive approach is implemented using the
tools and methodologies developed by the Real-Time Embedded Systems group.Comment: 2nd Workshop on Engineering of Autonomic Systems (EASe), in the 12th
Annual IEEE International Conference and Workshop on the Engineering of
Computer Based Systems (ECBS), Washington, DC, April, 200
Transfer Learning for Improving Model Predictions in Highly Configurable Software
Modern software systems are built to be used in dynamic environments using
configuration capabilities to adapt to changes and external uncertainties. In a
self-adaptation context, we are often interested in reasoning about the
performance of the systems under different configurations. Usually, we learn a
black-box model based on real measurements to predict the performance of the
system given a specific configuration. However, as modern systems become more
complex, there are many configuration parameters that may interact and we end
up learning an exponentially large configuration space. Naturally, this does
not scale when relying on real measurements in the actual changing environment.
We propose a different solution: Instead of taking the measurements from the
real system, we learn the model using samples from other sources, such as
simulators that approximate performance of the real system at low cost. We
define a cost model that transform the traditional view of model learning into
a multi-objective problem that not only takes into account model accuracy but
also measurements effort as well. We evaluate our cost-aware transfer learning
solution using real-world configurable software including (i) a robotic system,
(ii) 3 different stream processing applications, and (iii) a NoSQL database
system. The experimental results demonstrate that our approach can achieve (a)
a high prediction accuracy, as well as (b) a high model reliability.Comment: To be published in the proceedings of the 12th International
Symposium on Software Engineering for Adaptive and Self-Managing Systems
(SEAMS'17
Accelerating scientific codes by performance and accuracy modeling
Scientific software is often driven by multiple parameters that affect both
accuracy and performance. Since finding the optimal configuration of these
parameters is a highly complex task, it extremely common that the software is
used suboptimally. In a typical scenario, accuracy requirements are imposed,
and attained through suboptimal performance. In this paper, we present a
methodology for the automatic selection of parameters for simulation codes, and
a corresponding prototype tool. To be amenable to our methodology, the target
code must expose the parameters affecting accuracy and performance, and there
must be formulas available for error bounds and computational complexity of the
underlying methods. As a case study, we consider the particle-particle
particle-mesh method (PPPM) from the LAMMPS suite for molecular dynamics, and
use our tool to identify configurations of the input parameters that achieve a
given accuracy in the shortest execution time. When compared with the
configurations suggested by expert users, the parameters selected by our tool
yield reductions in the time-to-solution ranging between 10% and 60%. In other
words, for the typical scenario where a fixed number of core-hours are granted
and simulations of a fixed number of timesteps are to be run, usage of our tool
may allow up to twice as many simulations. While we develop our ideas using
LAMMPS as computational framework and use the PPPM method for dispersion as
case study, the methodology is general and valid for a range of software tools
and methods
An Adaptive Design Methodology for Reduction of Product Development Risk
Embedded systems interaction with environment inherently complicates
understanding of requirements and their correct implementation. However,
product uncertainty is highest during early stages of development. Design
verification is an essential step in the development of any system, especially
for Embedded System. This paper introduces a novel adaptive design methodology,
which incorporates step-wise prototyping and verification. With each adaptive
step product-realization level is enhanced while decreasing the level of
product uncertainty, thereby reducing the overall costs. The back-bone of this
frame-work is the development of Domain Specific Operational (DOP) Model and
the associated Verification Instrumentation for Test and Evaluation, developed
based on the DOP model. Together they generate functionally valid test-sequence
for carrying out prototype evaluation. With the help of a case study 'Multimode
Detection Subsystem' the application of this method is sketched. The design
methodologies can be compared by defining and computing a generic performance
criterion like Average design-cycle Risk. For the case study, by computing
Average design-cycle Risk, it is shown that the adaptive method reduces the
product development risk for a small increase in the total design cycle time.Comment: 21 pages, 9 figure
- …