18,371 research outputs found
Time granularity impact on propagation of disruptions in a system-of-systems simulation of infrastructure and business networks
System-of-systems (SoS) approach is often used for simulating disruptions to
business and infrastructure system networks allowing for integration of several
models into one simulation. However, the integration is frequently challenging
as each system is designed individually with different characteristics, such as
time granularity. Understanding the impact of time granularity on propagation
of disruptions between businesses and infrastructure systems and finding the
appropriate granularity for the SoS simulation remain as major challenges. To
tackle these, we explore how time granularity, recovery time, and disruption
size affect the propagation of disruptions between constituent systems of an
SoS simulation. To address this issue, we developed a High Level Architecture
(HLA) simulation of 3 networks and performed a series of simulation
experiments. Our results revealed that time granularity and especially recovery
time have huge impact on propagation of disruptions. Consequently, we developed
a model for selecting an appropriate time granularity for an SoS simulation
based on expected recovery time. Our simulation experiments show that time
granularity should be less than 1.13 of expected recovery time. We identified
some areas for future research centered around extending the experimental
factors space.Comment: 26 pages, 11 figures, 2 tables, Submitted to International Journal of
Environmental Research and Public Health: Special Issue on Cascading Disaster
Modelling and Preventio
W-NINE: a two-stage emulation platform for mobile and wireless systems
More and more applications and protocols are now running on wireless networks. Testing the implementation of such applications and protocols is a real challenge as the position of the mobile terminals and environmental effects strongly affect the overall performance. Network emulation is often perceived as a good trade-off between experiments on operational wireless networks and discrete-event simulations on Opnet or ns-2. However, ensuring repeatability and realism in network emulation while taking into account mobility in a wireless environment is very difficult. This paper proposes a network emulation platform, called W-NINE, based on off-line computations preceding online pattern-based traffic shaping. The underlying concepts of repeatability, dynamicity, accuracy and realism are defined in the emulation context. Two different simple case studies illustrate the validity of our approach with respect to these concepts
Many-Task Computing and Blue Waters
This report discusses many-task computing (MTC) generically and in the
context of the proposed Blue Waters systems, which is planned to be the largest
NSF-funded supercomputer when it begins production use in 2012. The aim of this
report is to inform the BW project about MTC, including understanding aspects
of MTC applications that can be used to characterize the domain and
understanding the implications of these aspects to middleware and policies.
Many MTC applications do not neatly fit the stereotypes of high-performance
computing (HPC) or high-throughput computing (HTC) applications. Like HTC
applications, by definition MTC applications are structured as graphs of
discrete tasks, with explicit input and output dependencies forming the graph
edges. However, MTC applications have significant features that distinguish
them from typical HTC applications. In particular, different engineering
constraints for hardware and software must be met in order to support these
applications. HTC applications have traditionally run on platforms such as
grids and clusters, through either workflow systems or parallel programming
systems. MTC applications, in contrast, will often demand a short time to
solution, may be communication intensive or data intensive, and may comprise
very short tasks. Therefore, hardware and software for MTC must be engineered
to support the additional communication and I/O and must minimize task dispatch
overheads. The hardware of large-scale HPC systems, with its high degree of
parallelism and support for intensive communication, is well suited for MTC
applications. However, HPC systems often lack a dynamic resource-provisioning
feature, are not ideal for task communication via the file system, and have an
I/O system that is not optimized for MTC-style applications. Hence, additional
software support is likely to be required to gain full benefit from the HPC
hardware
Distributed Hybrid Simulation of the Internet of Things and Smart Territories
This paper deals with the use of hybrid simulation to build and compose
heterogeneous simulation scenarios that can be proficiently exploited to model
and represent the Internet of Things (IoT). Hybrid simulation is a methodology
that combines multiple modalities of modeling/simulation. Complex scenarios are
decomposed into simpler ones, each one being simulated through a specific
simulation strategy. All these simulation building blocks are then synchronized
and coordinated. This simulation methodology is an ideal one to represent IoT
setups, which are usually very demanding, due to the heterogeneity of possible
scenarios arising from the massive deployment of an enormous amount of sensors
and devices. We present a use case concerned with the distributed simulation of
smart territories, a novel view of decentralized geographical spaces that,
thanks to the use of IoT, builds ICT services to manage resources in a way that
is sustainable and not harmful to the environment. Three different simulation
models are combined together, namely, an adaptive agent-based parallel and
distributed simulator, an OMNeT++ based discrete event simulator and a
script-language simulator based on MATLAB. Results from a performance analysis
confirm the viability of using hybrid simulation to model complex IoT
scenarios.Comment: arXiv admin note: substantial text overlap with arXiv:1605.0487
Fine-grained traffic state estimation and visualisation
Tools for visualising the current traffic state are used by local authorities for strategic monitoring of the traffic network and by everyday users for planning their journey. Popular visualisations include those provided by Google Maps and by Inrix. Both employ a traffic lights colour-coding system, where roads on a map are coloured green if traffic is flowing normally and red or black if there is congestion. New sensor technology, especially from wireless sources, is allowing resolution down to lane level. A case study is reported in which a traffic micro-simulation test bed is used to generate high-resolution estimates. An interactive visualisation of the fine-grained traffic state is presented. The visualisation is demonstrated using Google Earth and affords the user a detailed three-dimensional view of the traffic state down to lane level in real time
Integrating heterogeneous distributed COTS discrete-event simulation packages: An emerging standards-based approach
This paper reports on the progress made toward the emergence of standards to support the integration of heterogeneous discrete-event simulations (DESs) created in specialist support tools called commercial-off-the-shelf (COTS) discrete-event simulation packages (CSPs). The general standard for heterogeneous integration in this area has been developed from research in distributed simulation and is the IEEE 1516 standard The High Level Architecture (HLA). However, the specific needs of heterogeneous CSP integration require that the HLA is augmented by additional complementary standards. These are the suite of CSP interoperability (CSPI) standards being developed under the Simulation Interoperability Standards Organization (SISO-http://www.sisostds.org) by the CSPI Product Development Group (CSPI-PDG). The suite consists of several interoperability reference models (IRMs) that outline different integration needs of CSPI, interoperability frameworks (IFs) that define the HLA-based solution to each IRM, appropriate data exchange representations to specify the data exchanged in an IF, and benchmarks termed CSP emulators (CSPEs). This paper contributes to the development of the Type I IF that is intended to represent the HLA-based solution to the problem outlined by the Type I IRM (asynchronous entity passing) by developing the entity transfer specification (ETS) data exchange representation. The use of the ETS in an illustrative case study implemented using a prototype CSPE is shown. This case study also allows us to highlight the importance of event granularity and lookahead in the performance and development of the Type I IF, and to discuss possible methods to automate the capture of appropriate values of lookahead
Hybrid Radio-map for Noise Tolerant Wireless Indoor Localization
In wireless networks, radio-map based locating techniques are commonly used
to cope the complex fading feature of radio signal, in which a radio-map is
built by calibrating received signal strength (RSS) signatures at training
locations in the offline phase. However, in severe hostile environments, such
as in ship cabins where severe shadowing, blocking and multi-path fading
effects are posed by ubiquitous metallic architecture, even radio-map cannot
capture the dynamics of RSS. In this paper, we introduced multiple feature
radio-map location method for severely noisy environments. We proposed to add
low variance signature into radio map. Since the low variance signatures are
generally expensive to obtain, we focus on the scenario when the low variance
signatures are sparse. We studied efficient construction of multi-feature
radio-map in offline phase, and proposed feasible region narrowing down and
particle based algorithm for online tracking. Simulation results show the
remarkably performance improvement in terms of positioning accuracy and
robustness against RSS noises than the traditional radio-map method.Comment: 6 pages, 11th IEEE International Conference on Networking, Sensing
and Control, April 7-9, 2014, Miami, FL, US
A Taxonomy of Data Grids for Distributed Data Sharing, Management and Processing
Data Grids have been adopted as the platform for scientific communities that
need to share, access, transport, process and manage large data collections
distributed worldwide. They combine high-end computing technologies with
high-performance networking and wide-area storage management techniques. In
this paper, we discuss the key concepts behind Data Grids and compare them with
other data sharing and distribution paradigms such as content delivery
networks, peer-to-peer networks and distributed databases. We then provide
comprehensive taxonomies that cover various aspects of architecture, data
transportation, data replication and resource allocation and scheduling.
Finally, we map the proposed taxonomy to various Data Grid systems not only to
validate the taxonomy but also to identify areas for future exploration.
Through this taxonomy, we aim to categorise existing systems to better
understand their goals and their methodology. This would help evaluate their
applicability for solving similar problems. This taxonomy also provides a "gap
analysis" of this area through which researchers can potentially identify new
issues for investigation. Finally, we hope that the proposed taxonomy and
mapping also helps to provide an easy way for new practitioners to understand
this complex area of research.Comment: 46 pages, 16 figures, Technical Repor
A State-of-the-art Integrated Transportation Simulation Platform
Nowadays, universities and companies have a huge need for simulation and
modelling methodologies. In the particular case of traffic and transportation,
making physical modifications to the real traffic networks could be highly
expensive, dependent on political decisions and could be highly disruptive to
the environment. However, while studying a specific domain or problem,
analysing a problem through simulation may not be trivial and may need several
simulation tools, hence raising interoperability issues. To overcome these
problems, we propose an agent-directed transportation simulation platform,
through the cloud, by means of services. We intend to use the IEEE standard HLA
(High Level Architecture) for simulators interoperability and agents for
controlling and coordination. Our motivations are to allow multiresolution
analysis of complex domains, to allow experts to collaborate on the analysis of
a common problem and to allow co-simulation and synergy of different
application domains. This paper will start by presenting some preliminary
background concepts to help better understand the scope of this work. After
that, the results of a literature review is shown. Finally, the general
architecture of a transportation simulation platform is proposed
- …