593 research outputs found

    Scheduling for Proportional Differentiated Services on the Internet

    Get PDF
    Proportional Differentiated Services can be provisioned in terms of bandwidth, delay, or packet loss. Several studies contributed schedulers and packet droppers that achieved proportional bandwidth, delay, or loss differentiation. However, all these schemes differentiated in terms of only one of the three metrics. A simple, unified, scalable, and robust scheme to simultaneously control all three metrics was felt important. By controlling just delay and packet loss, proportional differentiation can be achieved in terms of all three metrics. A robust adaptive scheduler for proportional delay differentiation services is presented. Proportional services are further policed by a class based packet dropper. The combination of the adaptive scheduler and the packet dropper treats different traffic classes proportionally in terms of all three metrics. Simulation experiments show that regardless of the network traffic characteristics, our scheme can effectively differentiate services in terms of bandwidth, delay, and loss simultaneously

    An emission and discard priority scheme for optical burst switched networks

    Get PDF
    Word processed copy.Includes bibliographical references (leaves 76-79).Optical burst switching (OBS) is a promising technology designed to meet the growing demands for internet handwidth and better Quality of Service (QoS). This technology provides all optical and high speed switching to overcome the bottleneck of electronic routers in the core network. In this thesis, I describe several critical issues that affect OBS networks. I highlight the need to resolve contention efficiently and cost-effectively to improve QoS in OBS networks

    Cost functions in optical burst-switched networks

    Get PDF
    Optical Burst Switching (OBS) is a new paradigm for an all-optical Internet. It combines the best features of Optical Circuit Switching (OCS) and Optical Packet Switching (OPS) while avoidmg the mam problems associated with those networks .Namely, it offers good granularity, but its hardware requirements are lower than those of OPS. In a backbone network, low loss ratio is of particular importance. Also, to meet varying user requirements, it should support multiple classes of service. In Optical Burst-Switched networks both these goals are closely related to the way bursts are arranged in channels. Unlike the case of circuit switching, scheduling decisions affect the loss probability of future burst This thesis proposes the idea of a cost function. The cost function is used to judge the quality of a burst arrangement and estimate the probability that this burst will interfere with future bursts. Two applications of the cost functio n are proposed. A scheduling algorithm uses the value of the cost function to optimize the alignment of the new burst with other bursts in a channel, thus minimising the loss ratio. A cost-based burst droppmg algorithm, that can be used as a part of a Quality of Service scheme, drops only those bursts, for which the cost function value indicates that are most likely to cause a contention. Simulation results, performed using a custom-made OBS extension to the ns-2 simulator, show that the cost-based algorithms improve network performanc

    Performance of the transmission control protocol (TCP) over wireless with quality of service.

    Get PDF
    Thesis (M.Sc.Eng.)-University of Natal, Durban, 2001.The Transmission Control Protocol (TCP) is the most widely used transport protocol in the Internet. TCP is a reliable transport protocol that is tuned to perform well in wired networks where packet losses are mainly due to congestion. Wireless channels are characterized by losses due to transmission errors and handoffs. TCP interprets these losses as congestion and invokes congestion control mechanisms resulting in degradation of performance. TCP is usually layered over the Internet protocol (lP) at the network layer. JP is not reliable and does not provide for any Quality of Service (QoS). The Internet Engineering Task Force (IETF) has provided two techniques for providing QoS in the Internet. These include Integrated Services (lntServ) and Differentiated Services (DiffServ). IntServ provides flow based quality of service and thus it is not scalable on connections with large flows. DiffServ has grown in popularity since it is scalable. A packet in a DiffServ domain is classified into a class of service according to its contract profile and treated differently by its class. To provide end-to-end QoS there is a strong interaction between the transport protocol and the network protocol. In this dissertation we consider the performance of the TCP over a wireless channel. We study whether the current TCP protocols can deliver the desired quality of service faced with the challenges they have on wireless channel. The dissertation discusses the methods of providing for QoS in the Internet. We derive an analytical model for TCP protocol. It is extended to cater for the wireless channel and then further differentiated services. The model is shown to be accurate when compared to simulation. We then conclude by deducing to what degree you can provide the desired QoS with TCP on a wireless channel

    Traffic Profiles and Performance Modelling of Heterogeneous Networks

    Get PDF
    This thesis considers the analysis and study of short and long-term traffic patterns of heterogeneous networks. A large number of traffic profiles from different locations and network environments have been determined. The result of the analysis of these patterns has led to a new parameter, namely the 'application signature'. It was found that these signatures manifest themselves in various granularities over time, and are usually unique to an application, permanent virtual circuit (PVC), user or service. The differentiation of the application signatures into different categories creates a foundation for short and long-term management of networks. The thesis therefore looks from the micro and macro perspective on traffic management, covering both aspects. The long-term traffic patterns have been used to develop a novel methodology for network planning and design. As the size and complexity of interconnected systems grow steadily, usually covering different time zones, geographical and political areas, a new methodology has been developed as part of this thesis. A part of the methodology is a new overbooking mechanism, which stands in contrast to existing overbooking methods created by companies like Bell Labs. The new overbooking provides companies with cheaper network design and higher average throughput. In addition, new requirements like risk factors have been incorporated into the methodology, which lay historically outside the design process. A large network service provider has implemented the overbooking mechanism into their network planning process, enabling practical evaluation. The other aspect of the thesis looks at short-term traffic patterns, to analyse how congestion can be controlled. Reoccurring short-term traffic patterns, the application signatures, have been used for this research to develop the "packet train model" further. Through this research a new congestion control mechanism was created to investigate how the application signatures and the "extended packet train model" could be used. To validate the results, a software simulation has been written that executes the proprietary congestion mechanism and the new mechanism for comparison. Application signatures for the TCP/IP protocols have been applied in the simulation and the results are displayed and discussed in the thesis. The findings show the effects that frame relay congestion control mechanisms have on TCP/IP, where the re-sending of segments, buffer allocation, delay and throughput are compared. The results prove that application signatures can be used effectively to enhance existing congestion control mechanisms.AT&T (UK) Ltd, Englan

    Advances in Internet Quality of Service

    Get PDF
    We describe recent advances in theories and architecture that support performance guarantees needed for quality of service networks. We start with deterministic computations and give applications to integrated services, differentiated services, and playback delays. We review the methods used for obtaining a scalable integrated services support, based on the concept of a stateless core. New probabilistic results that can be used for a statistical dimensioning of differentiated services are explained; some are based on classical queuing theory, while others capitalize on the deterministic results. Then we discuss performance guarantees in a best effort context; we review: methods to provide some quality of service in a pure best effort environment; methods to provide some quality of service differentiation without access control, and methods that allow an application to control the performance it receives, in the absence of network support

    Theories and Models for Internet Quality of Service

    Get PDF
    We survey recent advances in theories and models for Internet Quality of Service (QoS). We start with the theory of network calculus, which lays the foundation for support of deterministic performance guarantees in networks, and illustrate its applications to integrated services, differentiated services, and streaming media playback delays. We also present mechanisms and architecture for scalable support of guaranteed services in the Internet, based on the concept of a stateless core. Methods for scalable control operations are also briefly discussed. We then turn our attention to statistical performance guarantees, and describe several new probabilistic results that can be used for a statistical dimensioning of differentiated services. Lastly, we review recent proposals and results in supporting performance guarantees in a best effort context. These include models for elastic throughput guarantees based on TCP performance modeling, techniques for some quality of service differentiation without access control, and methods that allow an application to control the performance it receives, in the absence of network support

    A user-centered approach to network quality of service and charging

    Get PDF
    The number of network users is expected to triple between 1998 and 2002 (Cullinane, 1998). While a vision of the future Internet offers the potential to break traditional barriers in communications and commerce, the current level of service does not satisfy the requirements of many users (Network Reliability Steering Committee, 1998, Cullinane, 1998). This thesis is concerned with users' perceptions of Quality of Service (QoS), and their attitudes to charging mechanisms applied to wide-area networks. Whilst the majority of research in this area has been conducted from a technical point of view, studies addressing issues of QoS and charging from a users' perspective are limited. The aim of this research was to investigate the latter issue to provide a more complete and integrated perspective on QoS and charging in the user-network system. The thesis first addresses previous work that looks at QoS and charging, establishing a justification for the new research. This part of the thesis concludes that, whilst part of our understanding of QoS requirements can be explained by technical and economic paradigms, additional research is required to examine the perceptions and concomitant behaviour of users. The methodology employed is outlined in relation to obtaining this objective. The second part of the thesis details work undertaken. This work has made the following main contributions: *Developed a set of conceptual models that describe users' perceptions of network QoS. *Shown that these models can be used to predict users' behaviour in different contexts by capturing subjective evaluations of QoS. * Shown how a combination of established and new methods can be successfully applied in capturing and assessing users' perceptions of QoS. *Shown how the new data relates to technical and econometric research. *Provided concrete examples of how the new research can inform network systems design. The work documented in this thesis has implications for user-centred, technical and econometric research. This thesis therefore contributes, not only to the field of HCI to which it is most closely related, but provides guidelines that can be used by econometricians and network designers. The research from all three of these perspectives is concerned with the efficient function of network resource allocation systems. The work documented in this thesis has suggested how it is possible to integrate these perspectives to provide valued levels of QoS to users

    A cross-layer middleware architecture for time and safety critical applications in MANETs

    Get PDF
    Mobile Ad hoc Networks (MANETs) can be deployed instantaneously and adaptively, making them highly suitable to military, medical and disaster-response scenarios. Using real-time applications for provision of instantaneous and dependable communications, media streaming, and device control in these scenarios is a growing research field. Realising timing requirements in packet delivery is essential to safety-critical real-time applications that are both delay- and loss-sensitive. Safety of these applications is compromised by packet loss, both on the network and by the applications themselves that will drop packets exceeding delay bounds. However, the provision of this required Quality of Service (QoS) must overcome issues relating to the lack of reliable existing infrastructure, conservation of safety-certified functionality. It must also overcome issues relating to the layer-2 dynamics with causal factors including hidden transmitters and fading channels. This thesis proposes that bounded maximum delay and safety-critical application support can be achieved by using cross-layer middleware. Such an approach benefits from the use of established protocols without requiring modifications to safety-certified ones. This research proposes ROAM: a novel, adaptive and scalable cross-layer Real-time Optimising Ad hoc Middleware framework for the provision and maintenance of performance guarantees in self-configuring MANETs. The ROAM framework is designed to be scalable to new optimisers and MANET protocols and requires no modifications of protocol functionality. Four original contributions are proposed: (1) ROAM, a middleware entity abstracts information from the protocol stack using application programming interfaces (APIs) and that implements optimisers to monitor and autonomously tune conditions at protocol layers in response to dynamic network conditions. The cross-layer approach is MANET protocol generic, using minimal imposition on the protocol stack, without protocol modification requirements. (2) A horizontal handoff optimiser that responds to time-varying link quality to ensure optimal and most robust channel usage. (3) A distributed contention reduction optimiser that reduces channel contention and related delay, in response to detection of the presence of a hidden transmitter. (4) A feasibility evaluation of the ROAM architecture to bound maximum delay and jitter in a comprehensive range of ns2-MIRACLE simulation scenarios that demonstrate independence from the key causes of network dynamics: application setting and MANET configuration; including mobility or topology. Experimental results show that ROAM can constrain end-to-end delay, jitter and packet loss, to support real-time applications with critical timing requirements
    • 

    corecore