221 research outputs found
A distributed computing environment (DCE) based object request broker
Includes bibliographical references.Object oriented technology has moved beyond being a tool for design and programming and is now being used to implement enterprise wide computer systems. Also, there has been a move from centralised mainframe systems to distributed computing due to the advent of more powerful workstations and faster, more reliable networks. The integration of object oriented technology and distributed computing is becoming a generally accepted method for implementing networked computer solutions. The purpose of the research presented in this thesis is to investigate how the evolving object oriented technologies can build upon the current distributed computing technology by using there underlying infrastructure and then to implement a CORBA compliant distributed Object Request Broker. This involves the design and implementation of a compiler which maps CORBA objects to DCE remote procedure calls. Our objective is to investigate the operation of a distributed object implementation and in particular the performance which can be achieved by a DCE-based Object Request Broker which is CORBA compliant
Adaptive object management for distributed systems
This thesis describes an architecture supporting the management of pluggable software components and evaluates it against the requirement for an enterprise integration platform for the manufacturing and petrochemical industries. In a distributed environment, we need mechanisms to manage objects and their interactions. At the least, we must be able to create objects in different processes on different nodes; we must be able to link them together so that they can pass messages to each other across the network; and we must deliver their messages in a timely and reliable manner. Object based environments which support these services already exist, for example ANSAware(ANSA, 1989), DEC's Objectbroker(ACA,1992), Iona's Orbix(Orbix,1994)Yet such environments provide limited support for composing applications from pluggable components. Pluggability is the ability to install and configure a component into an environment dynamically when the component is used, without specifying static dependencies between components when they are produced. Pluggability is supported to a degree by dynamic binding. Components may be programmed to import references to other components and to explore their interfaces at runtime, without using static type dependencies. Yet thus overloads the component with the responsibility to explore bindings. What is still generally missing is an efficient general-purpose binding model for managing bindings between independently produced components. In addition, existing environments provide no clear strategy for dealing with fine grained objects. The overhead of runtime binding and remote messaging will severely reduce performance where there are a lot of objects with complex patterns of interaction. We need an adaptive approach to managing configurations of pluggable components according to the needs and constraints of the environment. Management is made difficult by embedding bindings in component implementations and by relying on strong typing as the only means of verifying and validating bindings. To solve these problems we have built a set of configuration tools on top of an existing distributed support environment. Specification tools facilitate the construction of independent pluggable components. Visual composition tools facilitate the configuration of components into applications and the verification of composite behaviours. A configuration model is constructed which maintains the environmental state. Adaptive management is made possible by changing the management policy according to this state. Such policy changes affect the location of objects, their bindings, and the choice of messaging system
Deep Space Network information system architecture study
The purpose of this article is to describe an architecture for the Deep Space Network (DSN) information system in the years 2000-2010 and to provide guidelines for its evolution during the 1990s. The study scope is defined to be from the front-end areas at the antennas to the end users (spacecraft teams, principal investigators, archival storage systems, and non-NASA partners). The architectural vision provides guidance for major DSN implementation efforts during the next decade. A strong motivation for the study is an expected dramatic improvement in information-systems technologies, such as the following: computer processing, automation technology (including knowledge-based systems), networking and data transport, software and hardware engineering, and human-interface technology. The proposed Ground Information System has the following major features: unified architecture from the front-end area to the end user; open-systems standards to achieve interoperability; DSN production of level 0 data; delivery of level 0 data from the Deep Space Communications Complex, if desired; dedicated telemetry processors for each receiver; security against unauthorized access and errors; and highly automated monitor and control
Open solutions to distributed control in ground tracking stations
The advent of high speed local area networks has made it possible to interconnect small, powerful computers to function together as a single large computer. Today, distributed computer systems are the new paradigm for large scale computing systems. However, the communications provided by the local area network is only one part of the solution. The services and protocols used by the application programs to communicate across the network are as indispensable as the local area network. And the selection of services and protocols that do not match the system requirements will limit the capabilities, performance, and expansion of the system. Proprietary solutions are available but are usually limited to a select set of equipment. However, there are two solutions based on 'open' standards. The question that must be answered is 'which one is the best one for my job?' This paper examines a model for tracking stations and their requirements for interprocessor communications in the next century. The model and requirements are matched with the model and services provided by the five different software architectures and supporting protocol solutions. Several key services are examined in detail to determine which services and protocols most closely match the requirements for the tracking station environment. The study reveals that the protocols are tailored to the problem domains for which they were originally designed. Further, the study reveals that the process control model is the closest match to the tracking station model
Recommended from our members
Distributed computing testbed for a remote experimental environment
Collaboration is increasing as physics research becomes concentrated on a few large, expensive facilities, particularly in magnetic fusion energy research, with national and international participation. These facilities are designed for steady state operation and interactive, real-time experimentation. We are developing tools to provide for the establishment of geographically distant centers for interactive operations; such centers would allow scientists to participate in experiments from their home institutions. A testbed is being developed for a Remote Experimental Environment (REE), a ``Collaboratory.`` The testbed will be used to evaluate the ability of a remotely located group of scientists to conduct research on the DIII-D Tokamak at General Atomics. The REE will serve as a testing environment for advanced control and collaboration concepts applicable to future experiments. Process-to-process communications over high speed wide area networks provide real-time synchronization and exchange of data among multiple computer networks, while the ability to conduct research is enhanced by adding audio/video communication capabilities. The Open Software Foundation`s Distributed Computing Environment is being used to test concepts in distributed control, security, naming, remote procedure calls and distributed file access using the Distributed File Services. We are exploring the technology and sociology of remotely participating in the operation of a large scale experimental facility
Recommended from our members
Information integrity and privacy for computerized medical patient records
Sandia National Laboratories and Oceania, Inc. entered into a Cooperative Research and Development Agreement (CRADA) in November 1993 to provide ``Information Integrity and Privacy for Computerized Medical Patient Records`` (CRADA No. SC93/01183). The main objective of the project was to develop information protection methods that are appropriate for databases of patient records in health information systems. This document describes the findings and alternative solutions that resulted from this CRADA
NASA/ASEE Summer Faculty Fellowship Program
This document is a collection of technical reports on research conducted by the participants in the 1996 NASA/ASEE Summer Faculty Fellowship Program at the Kennedy Space Center (KSC). This was the twelfth year that a NASA/ASEE program has been conducted at KSC. The 1996 program was administered by the University of Central Florida in cooperation with KSC. The program was operated under the auspices of the American Society for Engineering Education (ASEE) with sponsorship and funding from the Office of Educational Affairs, NASA Headquarters, Washington, DC and KSC. The KSC Program was one of nine such Aeronautics and Space Research Program funded by NASA in 1996. The NASA/ASEE Program is intended to be a two-year program to allow in-depth research by the University faculty member. The editors of this document were responsible for selecting appropriately qualified faculty to address some of the many problems of current interest to NASA/KSC
Delegating remote operation execution in a mobile computing environment
Remote operation execution is nowadays the most popular paradigm
used to build distributed systems and applications. This success
originates in the simplicity exhibited by programming along the
client--server paradigm. Unfortunately, connectivity and bandwidth
restrictions defy the unchanged porting of this well known
mechanisms to the mobile computing field.
In this paper we present an approach that allows to develop
applications which are tailored for the specific requirements of
mobile computing, while retaining the simple and well understood
remote execution paradigm. The approach provides the additional
benefit that established services could easily be used from mobile
platforms. The cornerstone of our approach is integrated linguistic
support for dynamically delegating the execution and control of
remote procedure calls (RPC) to a delegate located on the fixed part
of the network. Besides presenting the language constructs, we
discuss the extensions to the RPC--based development process and the
necessary run--time support
- …