23,198 research outputs found

    Service level agreement framework for differentiated survivability in GMPLS-based IP-over-optical networks

    Get PDF
    In the next generation optical internet, GMPLS based IP-over-optical networks, ISPs will be required to support a wide variety of applications each having their own requirements. These requirements are contracted by means of the SLA. This paper describes a recovery framework that may be included in the SLA contract between ISP and customers in order to provide the required level of survivability. A key concern with such a recovery framework is how to present the different survivability alternatives including recovery techniques, failure scenario and layered integration into a transparent manner for customers. In this paper, two issues are investigated. First, the performance of the recovery framework when applying a proposed mapping procedure as an admission control mechanism in the edge router considering a smart-edge simple-core GMPLS-based IP/WDM network is considered. The second issue pertains to the performance of a pre-allocated restoration and its ability to provide protected connections under different failure scenarios

    Computing in the RAIN: a reliable array of independent nodes

    Get PDF
    The RAIN project is a research collaboration between Caltech and NASA-JPL on distributed computing and data-storage systems for future spaceborne missions. The goal of the project is to identify and develop key building blocks for reliable distributed systems built with inexpensive off-the-shelf components. The RAIN platform consists of a heterogeneous cluster of computing and/or storage nodes connected via multiple interfaces to networks configured in fault-tolerant topologies. The RAIN software components run in conjunction with operating system services and standard network protocols. Through software-implemented fault tolerance, the system tolerates multiple node, link, and switch failures, with no single point of failure. The RAIN-technology has been transferred to Rainfinity, a start-up company focusing on creating clustered solutions for improving the performance and availability of Internet data centers. In this paper, we describe the following contributions: 1) fault-tolerant interconnect topologies and communication protocols providing consistent error reporting of link failures, 2) fault management techniques based on group membership, and 3) data storage schemes based on computationally efficient error-control codes. We present several proof-of-concept applications: a highly-available video server, a highly-available Web server, and a distributed checkpointing system. Also, we describe a commercial product, Rainwall, built with the RAIN technology

    Spare capacity modelling and its applications in survivable iP-over-optical networks

    Get PDF
    As the interest in IP-over-optical networks are becoming the preferred core network architecture, survivability has emerged as a major concern for network service providers; a result of the potentially huge traffic volumes that will be supported by optical infrastructure. Therefore, implementing recovery strategies is critical. In addition to the traditional recovery schemes based around protection and restoration mechanisms, pre-allocated restoration represents a potential candidate to effect and maintain network resilience under failure conditions. Preallocated restoration technique is particularly interesting because it provides a trade-off in terms of recovery performance and resources between protection and restoration schemes. In this paper, the pre-allocated restoration performance is investigated under single and dual-link failures considering a distributed GMPLSbased IP/WDM mesh network. Two load-based spare capacity optimisation methods are proposed in this paper; Local Spare Capacity Optimisation (LSCO) and Global Spare Capacity Optimisation (GSCO)

    Reliable Messaging to Millions of Users with MigratoryData

    Full text link
    Web-based notification services are used by a large range of businesses to selectively distribute live updates to customers, following the publish/subscribe (pub/sub) model. Typical deployments can involve millions of subscribers expecting ordering and delivery guarantees together with low latencies. Notification services must be vertically and horizontally scalable, and adopt replication to provide a reliable service. We report our experience building and operating MigratoryData, a highly-scalable notification service. We discuss the typical requirements of MigratoryData customers, and describe the architecture and design of the service, focusing on scalability and fault tolerance. Our evaluation demonstrates the ability of MigratoryData to handle millions of concurrent connections and support a reliable notification service despite server failures and network disconnections

    Quality of Service challenges for Voice over Internet Protocol (VoIP) within the wireless environment

    Get PDF

    Testing Embedded Memories in Telecommunication Systems

    Get PDF
    Extensive system testing is mandatory nowadays to achieve high product quality. Telecommunication systems are particularly sensitive to such a requirement; to maintain market competitiveness, manufacturers need to combine reduced costs, shorter life cycles, advanced technologies, and high quality. Moreover, strict reliability constraints usually impose very low fault latencies and a high degree of fault detection for both permanent and transient faults. This article analyzes major problems related to testing complex telecommunication systems, with particular emphasis on their memory modules, often so critical from the reliability point of view. In particular, advanced BIST-based solutions are analyzed, and two significant industrial case studies presente

    Quality assessment technique for ubiquitous software and middleware

    Get PDF
    The new paradigm of computing or information systems is ubiquitous computing systems. The technology-oriented issues of ubiquitous computing systems have made researchers pay much attention to the feasibility study of the technologies rather than building quality assurance indices or guidelines. In this context, measuring quality is the key to developing high-quality ubiquitous computing products. For this reason, various quality models have been defined, adopted and enhanced over the years, for example, the need for one recognised standard quality model (ISO/IEC 9126) is the result of a consensus for a software quality model on three levels: characteristics, sub-characteristics, and metrics. However, it is very much unlikely that this scheme will be directly applicable to ubiquitous computing environments which are considerably different to conventional software, trailing a big concern which is being given to reformulate existing methods, and especially to elaborate new assessment techniques for ubiquitous computing environments. This paper selects appropriate quality characteristics for the ubiquitous computing environment, which can be used as the quality target for both ubiquitous computing product evaluation processes ad development processes. Further, each of the quality characteristics has been expanded with evaluation questions and metrics, in some cases with measures. In addition, this quality model has been applied to the industrial setting of the ubiquitous computing environment. These have revealed that while the approach was sound, there are some parts to be more developed in the future
    • 

    corecore