3,554 research outputs found
Experiences in Multi-domain Management Service Development
The developers of management systems and the management services that operate over them will be faced with increasing complexity as services are developed for the open service market. This paper presents experiences in the development of management services that span several administrative domains and which are therefore representative of the complexities of the open service market. The work described involved the development of TMN based management systems that provided management services in support of multimedia teleservices operating over broadband networks
Recommended from our members
Computing infrastructure issues in distributed communications systems : a survey of operating system transport system architectures
The performance of distributed applications (such as file transfer, remote login, tele-conferencing, full-motion video, and scientific visualization) is influenced by several factors that interact in complex ways. In particular, application performance is significantly affected both by communication infrastructure factors and computing infrastructure factors. Several communication infrastructure factors include channel speed, bit-error rate, and congestion at intermediate switching nodes. Computing infrastructure factors include (among other things) both protocol processing activities (such as connection management, flow control, error detection, and retransmission) and general operating system factors (such as memory latency, CPU speed, interrupt and context switching overhead, process architecture, and message buffering). Due to a several orders of magnitude increase in network channel speed and an increase in application diversity, performance bottlenecks are shifting from the network factors to the transport system factors.This paper defines an abstraction called an "Operating System Transport System Architecture" (OSTSA) that is used to classify the major components and services in the computing infrastructure. End-to-end network protocols such as TCP, TP4, VMTP, XTP, and Delta-t typically run on general-purpose computers, where they utilize various operating system resources such as processors, virtual memory, and network controllers. The OSTSA provides services that integrate these resources to support distributed applications running on local and wide area networks.A taxonomy is presented to evaluate OSTSAs in terms of their support for protocol processing activities. We use this taxonomy to compare and contrast five general-purpose commercial and experimental operating systems including System V UNIX, BSD UNIX, the x-kernel, Choices, and Xinu
A Generic Network and System Management Framework
Networks and distributed systems have formed the basis of an ongoing communications revolution
that has led to the genesis of a wide variety of services. The constantly increasing size and
complexity of these systems does not come without problems. In some organisations, the
deployment of Information Technology has reached a state where the benefits from downsizing and
rightsizing by adding new services are undermined by the effort required to keep the system
running.
Management of networks and distributed systems in general has a straightforward goal: to provide
a productive environment in which work can be performed effectively. The work required for
management should be a small fraction of the total effort. Most IT systems are still managed in an
ad hoc style without any carefully elaborated plan. In such an environment the success of
management decisions depends totally on the qualification and knowledge of the administrator.
The thesis provides an analysis of the state of the art in the area of Network and System
Management and identifies the key requirements that must be addressed for the provisioning of
Integrated Management Services. These include the integration of the different management related
aspects (i.e. integration of heterogeneous Network, System and Service Management).
The thesis then proposes a new framework, INSMware, for the provision of Management Services.
It provides a fundamental basis for the realisation of a new approach to Network and System
Management. It is argued that Management Systems can be derived from a set of pre-fabricated
and reusable Building Blocks that break up the required functionality into a number of separate
entities rather than being developed from scratch. It proposes a high-level logical model in order to
accommodate the range of requirements and environments applicable to Integrated Network and
System Management that can be used as a reference model.
A development methodology is introduced that reflects principles of the proposed approach, and
provides guidelines to structure the analysis, design and implementation phases of a management
system. The INSMware approach can further be combined with the componentware paradigm for
the implementation of the management system. Based on these principles, a prototype for the
management of SNMP systems has been implemented using industry standard middleware
technologies. It is argued that development of a management system based on Componentware
principles can offer a number of benefits. INSMware Components may be re-used and system
solutions will become more modular and thereby easier to construct and maintain
High performance computing simulator for the performance assessment of trajectory based operations
High performance computing (HPC), both at hardware and software level, has demonstrated significant improve-
ments in processing large datasets in a timely manner. However, HPC in the field of air traffic management (ATM) can be much more than only a time reducing tool. It could also be used to build an ATM simulator in which distributed scenarios where decentralized elements (airspace users) interact through a centralized manager in order to generate a trajectory-optimized conflict-free scenario. In this work, we introduce an early prototype of an ATM simulator, focusing on air traffic flow management at strategic, pre-tactical and tactical levels, which allows the calculation of safety and efficiency indicators for optimized trajectories, both at individual and network level. The software architecture of the simulator, relying on a HPC cluster of computers, has been preliminary tested with a set of flights whose trajectory vertical profiles have been optimized according to two different concepts of operations: conventional cruise operations (i.e. flying at constant altitudes and according to the flight levels scheme rules) and continuous climb cruise operations (i.e., optimizing the trajectories with no vertical constraints). The novel ATM simulator has been tested to show preliminary benchmarking results between these two concepts of operations. The simulator here presented can contribute as a testbed to evaluate the potential benefits of future Trajectory Based Operations and to understand the complex relationships among the different ATM key performance areasPeer ReviewedPostprint (published version
Multilevel Contracts for Trusted Components
This article contributes to the design and the verification of trusted
components and services. The contracts are declined at several levels to cover
then different facets, such as component consistency, compatibility or
correctness. The article introduces multilevel contracts and a
design+verification process for handling and analysing these contracts in
component models. The approach is implemented with the COSTO platform that
supports the Kmelia component model. A case study illustrates the overall
approach.Comment: In Proceedings WCSI 2010, arXiv:1010.233
The Impact of the Internet on Telecommunication Architectures
The ever-growing popularity of the Internet is dramatically changing the landscape of the communications market place. The two separate worlds of the Internet and Telecommunications are converging. The respective advantages of the two environments are being integrated to fulfill the promise of the information super-highways. In this paper, we examine the impact of the Internet on the main telecommunication architectures, namely the IN, the TMN and TINA. There are two new tendencies for implementing telephony services in combination with the Internet: running part of the control sys tem over the Internet, or conveying both the user data and the control information over the Internet. We examine these two trends, and elaborate on possible ways of salvaging the best parts of the work achieved by the TINA-Consortium in the Internet context
Towards a Taxonomy of Inter-network Architectures
Over the past decade, research on network architecture design has intensified. However, contributions to the field have mainly been idiosyncratic and architectural descriptions remain idiomatic. This state of affairs has led to the emergence of a large body of network architecture proposals with no clear indication of their compatibility points, their cross similarities, and their differences. Thus, a taxonomy of network architectures that provides a framework for better understanding, organizing, and thinking about the complex architecture design space would be a timely contribution. This paper presents a first step in that direction by attempting a classification based on the architecture\u27s information model. The taxonomy is applied to a special network architecture highlighting its descriptive and classification powers
Comparing a Traditional and a Multi-Agent Load-Balancing System
This article presents a comparison between agent and non-agent based approaches to building network-load-balancing systems. In particular, two large software systems are compared, one traditional and the other agent-based, both performing the same load balancing functions. Due to the two different architectures, several differences emerge. The differences are analyzed theoretically and practically in terms of design, scalability and fault-tolerance. The advantages and disadvantages of both approaches are presented by combining an analysis of the system and gathering the experience of designers, developers and users. Traditionally, designers specify rigid software structure, while for multi-agent systems the emphasis is on specifying the different tasks and roles, as well as the interconnections between the agents that cooperate autonomously and simultaneously. The major advantages of the multi-agent approach are the introduced abstract design layers and, as a consequence, the more comprehendible top-level design, the increased redundancy, and the improved fault tolerance. The major improvement in performance due to the agent architecture is observed in the case of one or more failed computers. Although the agent-oriented design might not be a silver bullet for building large distributed systems, our analysis and application confirm that it does have a number of advantages over non-agent approaches
- …