37,904 research outputs found
Techniques for the Fast Simulation of Models of Highly dependable Systems
With the ever-increasing complexity and requirements of highly dependable systems, their evaluation during design and operation is becoming more crucial. Realistic models of such systems are often not amenable to analysis using conventional analytic or numerical methods. Therefore, analysts and designers turn to simulation to evaluate these models. However, accurate estimation of dependability measures of these models requires that the simulation frequently observes system failures, which are rare events in highly dependable systems. This renders ordinary Simulation impractical for evaluating such systems. To overcome this problem, simulation techniques based on importance sampling have been developed, and are very effective in certain settings. When importance sampling works well, simulation run lengths can be reduced by several orders of magnitude when estimating transient as well as steady-state dependability measures. This paper reviews some of the importance-sampling techniques that have been developed in recent years to estimate dependability measures efficiently in Markov and nonMarkov models of highly dependable system
Statistical analysis of chemical computational systems with MULTIVESTA and ALCHEMIST
The chemical-oriented approach is an emerging paradigm for programming the behaviour of densely distributed and context-aware devices (e.g. in ecosystems of displays tailored to crowd steering, or to obtain profile-based coordinated visualization). Typically, the evolution of such systems cannot be easily predicted, thus making of paramount importance the availability of techniques and tools supporting prior-to-deployment analysis. Exact analysis techniques do not scale well when the complexity of systems grows: as a consequence, approximated techniques based on simulation assumed a relevant role. This work presents a new simulation-based distributed tool addressing the statistical analysis of such a kind of systems, which has been obtained by chaining two existing tools: MultiVeStA and Alchemist. The former is a recently proposed lightweight tool which allows to enrich existing discrete event simulators with distributed statistical analysis capabilities, while the latter is an efficient simulator for chemical-oriented computational systems. The tool is validated against a crowd steering scenario, and insights on the performance are provided by discussing how these scale distributing the analysis tasks on a multi-core architecture
Engine performance characteristics and evaluation of variation in the length of intake plenum
In the engine with multipoint fuel injection system using electronically controlled fuel injectors has an intake manifold in which only the air flows and, the fuel is injected into the intake valve. Since the intake manifolds transport mainly air, the supercharging effects of the variable length intake plenum will be different from carbureted engine. Engine tests have been carried out with the aim of constituting a base study to design a new variable length intake manifold plenum. The objective in this research is to study the engine performance characteristics and to evaluate the effects of the variation in the length of intake plenum. The engine test bed used for experimental work consists of a control panel, a hydraulic dynamometer and measurement instruments to measure the parameters of engine performance characteristics. The control panel is being used to perform administrative and management operating system. Besides that, the hydraulic dynamometer was used to measure the power of an engine by using a cell filled with liquid to increase its load. Thus, measurement instrument is provided in this test to measure the as brake torque, brake power, thermal efficiency and specific fuel consumption. The results showed that the variation in the plenum length causes an improvement on the engine performance characteristics especially on the fuel consumption at high load and low engine speeds which are put forward the system using for urban roads. From this experiment, it will show the behavior of engine performance
Computation-Communication Trade-offs and Sensor Selection in Real-time Estimation for Processing Networks
Recent advances in electronics are enabling substantial processing to be
performed at each node (robots, sensors) of a networked system. Local
processing enables data compression and may mitigate measurement noise, but it
is still slower compared to a central computer (it entails a larger
computational delay). However, while nodes can process the data in parallel,
the centralized computational is sequential in nature. On the other hand, if a
node sends raw data to a central computer for processing, it incurs
communication delay. This leads to a fundamental communication-computation
trade-off, where each node has to decide on the optimal amount of preprocessing
in order to maximize the network performance. We consider a network in charge
of estimating the state of a dynamical system and provide three contributions.
First, we provide a rigorous problem formulation for optimal real-time
estimation in processing networks in the presence of delays. Second, we show
that, in the case of a homogeneous network (where all sensors have the same
computation) that monitors a continuous-time scalar linear system, the optimal
amount of local preprocessing maximizing the network estimation performance can
be computed analytically. Third, we consider the realistic case of a
heterogeneous network monitoring a discrete-time multi-variate linear system
and provide algorithms to decide on suitable preprocessing at each node, and to
select a sensor subset when computational constraints make using all sensors
suboptimal. Numerical simulations show that selecting the sensors is crucial.
Moreover, we show that if the nodes apply the preprocessing policy suggested by
our algorithms, they can largely improve the network estimation performance.Comment: 15 pages, 16 figures. Accepted journal versio
Recommended from our members
Potential applications of simulation modelling techniques in healthcare: lessons learned from aerospace and military
The Aerospace and Military areas are to do with complex missions and situations. Modelling and Simulation (M&S) has been applied in many areas of defence ranging from space sciences, satellite engineering to multi-warfare (air warfare, undersea warfare), air & missile defence, acquisition, tactical military trainings & exercises, national security analysis and strategic decision making & planning, etc. The application of simulation modelling techniques in healthcare would improve the provision of healthcare services; however, their application has been much relatively feeble in the healthcare sector as compared to the defence sector. This paper presents results from a systematic literature survey on applications of modelling simulation techniques in the Aerospace & Military. The knowledge gained or lessons learned from the survey were finally used to analyze the potential applications of the simulation modelling techniques to the healthcare sector. Results show that in the defence sector, Distributed Simulation has now become a widely adopted technique. However, System Dynamics (SD) and Discrete Event Simulation (DSE) have also gained relative attention. From this survey it becomes clear that various simulation modelling techniques are useful for specific purposes and have potential applications in the healthcare sector
Sequence-based Anytime Control
We present two related anytime algorithms for control of nonlinear systems
when the processing resources available are time-varying. The basic idea is to
calculate tentative control input sequences for as many time steps into the
future as allowed by the available processing resources at every time step.
This serves to compensate for the time steps when the processor is not
available to perform any control calculations. Using a stochastic Lyapunov
function based approach, we analyze the stability of the resulting closed loop
system for the cases when the processor availability can be modeled as an
independent and identically distributed sequence and via an underlying Markov
chain. Numerical simulations indicate that the increase in performance due to
the proposed algorithms can be significant.Comment: 14 page
Event-based State Estimation: An Emulation-based Approach
An event-based state estimation approach for reducing communication in a
networked control system is proposed. Multiple distributed sensor agents
observe a dynamic process and sporadically transmit their measurements to
estimator agents over a shared bus network. Local event-triggering protocols
ensure that data is transmitted only when necessary to meet a desired
estimation accuracy. The event-based design is shown to emulate the performance
of a centralised state observer design up to guaranteed bounds, but with
reduced communication. The stability results for state estimation are extended
to the distributed control system that results when the local estimates are
used for feedback control. Results from numerical simulations and hardware
experiments illustrate the effectiveness of the proposed approach in reducing
network communication.Comment: 21 pages, 8 figures, this article is based on the technical report
arXiv:1511.05223 and is accepted for publication in IET Control Theory &
Application
Resource-aware IoT Control: Saving Communication through Predictive Triggering
The Internet of Things (IoT) interconnects multiple physical devices in
large-scale networks. When the 'things' coordinate decisions and act
collectively on shared information, feedback is introduced between them.
Multiple feedback loops are thus closed over a shared, general-purpose network.
Traditional feedback control is unsuitable for design of IoT control because it
relies on high-rate periodic communication and is ignorant of the shared
network resource. Therefore, recent event-based estimation methods are applied
herein for resource-aware IoT control allowing agents to decide online whether
communication with other agents is needed, or not. While this can reduce
network traffic significantly, a severe limitation of typical event-based
approaches is the need for instantaneous triggering decisions that leave no
time to reallocate freed resources (e.g., communication slots), which hence
remain unused. To address this problem, novel predictive and self triggering
protocols are proposed herein. From a unified Bayesian decision framework, two
schemes are developed: self triggers that predict, at the current triggering
instant, the next one; and predictive triggers that check at every time step,
whether communication will be needed at a given prediction horizon. The
suitability of these triggers for feedback control is demonstrated in hardware
experiments on a cart-pole, and scalability is discussed with a multi-vehicle
simulation.Comment: 16 pages, 15 figures, accepted article to appear in IEEE Internet of
Things Journal. arXiv admin note: text overlap with arXiv:1609.0753
- …