177 research outputs found
Recommended from our members
Parallel and distributed cyber-physical system simulation
textThe traditions of real-time and embedded system engineering have evolved into a new field of cyber-physical systems (CPSs). The increase in complexity of CPS components and the multi-domain engineering composition of CPSs challenge the current best practices in design and simulation. To address the challenges of CPS simulation, this work introduces a simulator coordination method drawing from strengths of the field of parallel and distributed simulation (PADS), yet offering benefits aimed towards the challenges of coordinating CPS engineering design simulators. The method offers the novel concept of Interpolated Event data types applied to Kahn Process Networks in order to provide simulator coordination. This can enable conservative and optimistic coordination of multiple heterogeneous and homogeneous simulators, but provide important benefits for CPS simulation, such as the opportunity to reduce functional requirements for simulator interfacing compared to existing solutions. The method is analyzed in theoretical properties and instantiated in software tools SimConnect and SimTalk. Finally, an experimental study applies the method and tools to accelerate Spice circuit simulation with tradeoffs in speed versus accuracy, and demonstrates the coordination of three heterogeneous simulators for a CPS simulation with increasing component model refinement and realism.Electrical and Computer Engineerin
Modeling, Design And Evaluation Of Networking Systems And Protocols Through Simulation
Computer modeling and simulation is a practical way to design and test a system without actually having to build it. Simulation has many benefits which apply to many different domains: it reduces costs creating different prototypes for mechanical engineers, increases the safety of chemical engineers exposed to dangerous chemicals, speeds up the time to model physical reactions, and trains soldiers to prepare for battle. The motivation behind this work is to build a common software framework that can be used to create new networking simulators on top of an HLA-based federation for distributed simulation. The goals are to model and simulate networking architectures and protocols by developing a common underlying simulation infrastructure and to reduce the time a developer has to learn the semantics of message passing and time management to free more time for experimentation and data collection and reporting. This is accomplished by evolving the simulation engine through three different applications that model three different types of network protocols. Computer networking is a good candidate for simulation because of the Internet\u27s rapid growth that has spawned off the need for new protocols and algorithms and the desire for a common infrastructure to model these protocols and algorithms. One simulation, the 3DInterconnect simulator, simulates data transmitting through a hardware k-array n-cube network interconnect. Performance results show that k-array n-cube topologies can sustain higher traffic load than the currently used interconnects. The second simulator, Cluster Leader Logic Algorithm Simulator, simulates an ad-hoc wireless routing protocol that uses a data distribution methodology based on the GPS-QHRA routing protocol. CLL algorithm can realize a maximum of 45% power savings and maximum 25% reduced queuing delay compared to GPS-QHRA. The third simulator simulates a grid resource discovery protocol for helping Virtual Organizations to find resource on a grid network to compute or store data on. Results show that worst-case 99.43% of the discovery messages are able to find a resource provider to use for computation. The simulation engine was then built to perform basic HLA operations. Results show successful HLA functions including creating, joining, and resigning from a federation, time management, and event publication and subscription
An Architectural Framework for Performance Analysis: Supporting the Design, Configuration, and Control of DIS /HLA Simulations
Technology advances are providing greater capabilities for most distributed computing environments. However, the advances in capabilities are paralleled by progressively increasing amounts of system complexity. In many instances, this complexity can lead to a lack of understanding regarding bottlenecks in run-time performance of distributed applications. This is especially true in the domain of distributed simulations where a myriad of enabling technologies are used as building blocks to provide large-scale, geographically disperse, dynamic virtual worlds. Persons responsible for the design, configuration, and control of distributed simulations need to understand the impact of decisions made regarding the allocation and use of the logical and physical resources that comprise a distributed simulation environment and how they effect run-time performance. Distributed Interactive Simulation (DIS) and High Level Architecture (HLA) simulation applications historically provide some of the most demanding distributed computing environments in terms of performance, and as such have a justified need for performance information sufficient to support decision-makers trying to improve system behavior.
This research addresses two fundamental questions: (1) Is there an analysis framework suitable for characterizing DIS and HLA simulation performance? and (2) what kind of mechanism can be used to adequately monitor, measure, and collect performance data to support different performance analysis objectives for DIS and HLA simulations? This thesis presents a unified, architectural framework for DIS and HLA simulations, provides details on a performance monitoring system, and shows its effectiveness through a series of use cases that include practical applications of the framework to support real-world U.S. Department of Defense (DoD) programs. The thesis also discusses the robustness of the constructed framework and its applicability to performance analysis of more general distributed computing applications
Application of Web Services to a Simulation Framework
The Joint Semi-Automated Forces (JSAF) simulator is an excellent tool for military training and a great testbed for new SAF behaviors. However, it has the drawback that behaviors must be ported into its own Finite State Machine (FSM) language. Web Services is a growing technology that seamlessly connects service providers to service consumers. This work attempts to merge these two technologies by modeling SAF behaviors as web services. The JSAF simulator is then modeled as a web service consumer. This approach allows new Semi-Automated Forces (SAF) behaviors to be developed independently of the simulator, which provides the developer with greater ļ¬exibility when choosing a programming language, development environment, and development platform. In addition to new SAF behaviors, this approach also supports any external component that can be modeled as a web service. Furthermore, these services are often run over a network, which distributes the computational load across several computers. Finally, hosting copies of a single service on several machines, a concept similar to ļ¬le-sharing mirrors, oļ¬ers an environment for load-balancing. This means if several entities are running the same behavior, a single server does not perform the computation for every entity. Instead, each entity is assigned to a speciļ¬c server, increasing the quality of service seen by the system. A Web Services framework linking JSAF with several services is designed and implemented. Suppression of Enemy Air Defense (SEAD) behaviors written in MATLAB and a behavior recognition system are integrated with JSAF. These behaviors and the recognition tool were developed by other researchers, independent of this work. Results show that oļ¬oading computation to other machines is beneļ¬cial, especially when the simulation system is under heavy load. Preliminary results also indicate that load-balancing performs much better than using a single server
Managing Bandwidth and Traffic via Bundling and Filtration in Large-Scale Distributed Simulations
Research has shown that bandwidth can be a limiting factor in the performance of distributed simulations. The Air Force\u27s Distributed Mission Operations Center (DMOC) periodically hosts one of the largest distributed simulation events in the world. The engineers at the DMOC have dealt with the difficult problem of limited bandwidth by implementing application level filters that process all DIS PDUs between the various networks connected to the exercise. This thesis examines their implemented filter and proposes: adaptive range-based filtering and bundling together of PDUs. The goals are to reduce the number of PDUs passed by the adaptive filter and to reduce network overhead and the total amount of data transferred by maximizing packet size up to the MTU. The proposed changes were implemented and logged data from previous events were used on a test network in order to measure the improvement from the base filter to the improved filter. The results showed that the adaptive range based filter was effective, though minimally so, and that the PDU bundling resulted in a reduction of 17% to 20% of the total traffic transmitted across the network
Extensible Modeling and Simulation Framework (XMSF) Opportunities for Web-Based Modeling and Simulation
Technical Opportunities Workshop Whitepaper, 14 June 2002Purpose: As the Department of Defense (DoD) is engaged in both warfighting and institutional
transformation for the new millennium, DoD Modeling & Simulation (M&S) also needs to identify
and adopt transformational technologies which provide direct tactical relevance to warfighters.
Because the only software systems that composably scale to worldwide scope utilize the World
Wide Web, it is evident that an extensible Web-based framework shows great promise to scale up
the capabilities of M&S systems to meet the needs of training, analysis, acquisition, and the
operational warfighter. By embracing commercial web technologies as a shared-communications
platform and a ubiquitous-delivery framework, DoD M&S can fully leverage mainstream practices
for enterprise-wide software development
Middleware services for distributed virtual environments
PhD ThesisDistributed Virtual Environments (DVEs) are virtual environments which allow
dispersed users to interact with each other and the virtual world through the
underlying network.
Scalability is a major challenge in building a successful DVE, which is directly
affected by the volume of message exchange. Different techniques have been
deployed to reduce the volume of message exchange in order to support large
numbers of simultaneous participants in a DVE. Interest management is a
popular technique for filtering unnecessary message exchange between users.
The rationale behind interest management is to resolve the "interests" of users
and decide whether messages should be exchanged between them. There are
three basic interest management approaches: region-based, aura-based and
hybrid approaches. However, if the time taken for an interest management
approach to determine interests is greater than the duration of the interaction, it
is not possible to guarantee interactions will occur correctly or at all. This is
termed the Missed Interaction Problem, which all existing interest management
approaches are susceptible to.
This thesis provides a new aura-based interest management approach, termed
Predictive Interest management (PIM), to alleviate the missed interaction
problem. PIM uses an enlarged aura to detect potential aura-intersections and
iii
initiate message exchange. It utilises variable message exchange frequencies,
proportional to the intersection degree of the objects' expanded auras, to restrict
bandwidth usage. This thesis provides an experimental system, the PIM system,
which couples predictive interest management with the de-centralised server
communication model. It utilises the Common Object Request Broker
Architecture (CORBA) middleware standard to provide an interoperable
middleware for DVEs. Experimental results are provided to demonstrate that
PIM provides a scalable interest management approach which alleviates the
missed interaction problem
Middleware services for distributed virtual environments
PhD ThesisDistributed Virtual Environments (DVEs) are virtual environments which allow
dispersed users to interact with each other and the virtual world through the
underlying network.
Scalability is a major challenge in building a successful DVE, which is directly
affected by the volume of message exchange. Different techniques have been
deployed to reduce the volume of message exchange in order to support large
numbers of simultaneous participants in a DVE. Interest management is a
popular technique for filtering unnecessary message exchange between users.
The rationale behind interest management is to resolve the "interests" of users
and decide whether messages should be exchanged between them. There are
three basic interest management approaches: region-based, aura-based and
hybrid approaches. However, if the time taken for an interest management
approach to determine interests is greater than the duration of the interaction, it
is not possible to guarantee interactions will occur correctly or at all. This is
termed the Missed Interaction Problem, which all existing interest management
approaches are susceptible to.
This thesis provides a new aura-based interest management approach, termed
Predictive Interest management (PIM), to alleviate the missed interaction
problem. PIM uses an enlarged aura to detect potential aura-intersections and
iii
initiate message exchange. It utilises variable message exchange frequencies,
proportional to the intersection degree of the objects' expanded auras, to restrict
bandwidth usage. This thesis provides an experimental system, the PIM system,
which couples predictive interest management with the de-centralised server
communication model. It utilises the Common Object Request Broker
Architecture (CORBA) middleware standard to provide an interoperable
middleware for DVEs. Experimental results are provided to demonstrate that
PIM provides a scalable interest management approach which alleviates the
missed interaction problem
Integrated Space Asset Management Database and Modeling
Effective Space Asset Management is one key to addressing the ever-growing issue of space congestion. It is imperative that agencies around the world have access to data regarding the numerous active assets and pieces of space junk currently tracked in orbit around the Earth. At the center of this issues is the effective management of data of many types related to orbiting objects. As the population of tracked objects grows, so too should the data management structure used to catalog technical specifications, orbital information, and metadata related to those populations. Marshall Space Flight Center's Space Asset Management Database (SAM-D) was implemented in order to effectively catalog a broad set of data related to known objects in space by ingesting information from a variety of database and processing that data into useful technical information. Using the universal NORAD number as a unique identifier, the SAM-D processes two-line element data into orbital characteristics and cross-references this technical data with metadata related to functional status, country of ownership, and application category. The SAM-D began as an Excel spreadsheet and was later upgraded to an Access database. While SAM-D performs its task very well, it is limited by its current platform and is not available outside of the local user base. Further, while modeling and simulation can be powerful tools to exploit the information contained in SAM-D, the current system does not allow proper integration options for combining the data with both legacy and new M&S tools. This paper provides a summary of SAM-D development efforts to date and outlines a proposed data management infrastructure that extends SAM-D to support the larger data sets to be generated. A service-oriented architecture model using an information sharing platform named SIMON will allow it to easily expand to incorporate new capabilities, including advanced analytics, M&S tools, fusion techniques and user interface for visualizations. In addition, tight control of information sharing policy will increase confidence in the system, which would encourage industry partners to provide commercial data. Combined with the integration of new and legacy M&S tools, a SIMON-based architecture will provide a robust environment that can be extended and expanded indefinitely
Recommended from our members
A grid computing framework for commercial simulation packages
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.An increased need for collaborative research among different organizations, together with continuing advances in communication technology and computer hardware, has facilitated the development of distributed systems that can provide users non-trivial access to geographically dispersed computing resources (processors, storage, applications, data, instruments, etc.) that are administered in multiple computer domains. The term grid computing or grids is popularly used to refer to such distributed systems. A broader definition of grid computing includes the use of computing resources within an organization for running organization-specific applications. This research is in the context of using grid computing within an enterprise to maximize the use of available hardware and software resources for processing enterprise applications. Large scale scientific simulations have traditionally been the primary benefactor of grid computing. The application of this technology to simulation in industry has, however, been negligible. This research investigates how grid technology can be effectively exploited by simulation practitioners using Windows-based commercially available simulation packages to model simulations in industry. These packages are commonly referred to as Commercial Off-The-Shelf (COTS) Simulation Packages (CSPs). The study identifies several higher level grid services that could be potentially used to support the practise of simulation in industry. It proposes a grid computing framework to investigate these services in the context of CSP-based simulations. This framework is called the CSP-Grid Computing (CSP-GC) Framework. Each identified higher level grid service in this framework is referred to as a CSP-specific service. A total of six case studies are presented to experimentally evaluate how grid computing technologies can be used together with unmodified simulation packages to support some of the CSP-specific services. The contribution of this thesis is the CSP-GC framework that identifies how simulation practise in industry may benefit from the use of grid technology. A further contribution is the recognition of specific grid computing software (grid middleware) that can possibly be used together with existing CSPs to provide grid support. With its focus on end-users and end-user tools, it is intended that this research will encourage wider adoption of grid computing in the workplace and that simulation users will derive benefit from using this technology
- ā¦