5,532 research outputs found
Time Driven Priority Router Implementation and First Experiments
This paper reports on the implementation of Time-Driven Priority (TDP) scheduling on a FreeBSD platform. This work is part of a TDP prototyping and demonstration project aimed at showing the implications of TDP deployment in packet-switched networks, especially benefits for real-time applications. This paper focuses on practical aspects related to the implementation of the technology on a Personal Computer (PC)-based router and presents the experimental results obtained on a testbed network. The basic building blocks of a TDP router are described and implementation choices are discussed. The relevant results achieved and here presented can be categorized into two types: qualitative results, including the successful integration of all needed blocks and the insight obtained on the complexity related to the implementation of a TDP router, and quantitative ones, including measures of achievable network utilization and of jitter experienced on a fully-loaded TDP network. The outcome demonstrates the effectiveness of the presented implementation while confirming TDP points of strengt
Traffic Engineering in G-MPLS networks with QoS guarantees
In this paper a new Traffic Engineering (TE) scheme to efficiently route sub-wavelength requests with different QoS requirements is proposed for G-MPLS networks. In most previous studies on TE based on dynamic traffic grooming, the objectives were to minimize the rejection probability by respecting the constraints of the optical node architecture, but without considering service differentiation. In practice, some high-priority (HP) connections can instead be characterized by specific constraints on the maximum tolerable end-to-end delay and packet-loss ratio. The proposed solution consists of a distributed two-stage scheme: each time a new request arrives, an on-line dynamic grooming scheme finds a route which fulfills the QoS requirements. If a HP request is blocked at the ingress router, a preemption algorithm is executed locally in order to create room for this traffic. The proposed preemption mechanism minimizes the network disruption, both in term of number of rerouted low-priority connections and new set-up lightpaths, and the signaling complexity. Extensive simulation experiments are performed to demonstrate the efficiency of our scheme
Container network functions: bringing NFV to the network edge
In order to cope with the increasing network utilization driven by new mobile clients, and to satisfy demand for new network services and performance guarantees, telecommunication service providers are exploiting virtualization over their network by implementing network services in virtual machines, decoupled from legacy hardware accelerated appliances. This effort, known as NFV, reduces OPEX and provides new business opportunities. At the same time, next generation mobile, enterprise, and IoT networks are introducing the concept of computing capabilities being pushed at the network edge, in close proximity of the users. However, the heavy footprint of today's NFV platforms prevents them from operating at the network edge. In this article, we identify the opportunities of virtualization at the network edge and present Glasgow Network Functions (GNF), a container-based NFV platform that runs and orchestrates lightweight container VNFs, saving core network utilization and providing lower latency. Finally, we demonstrate three useful examples of the platform: IoT DDoS remediation, on-demand troubleshooting for telco networks, and supporting roaming of network functions
TCP throughput guarantee in the DiffServ Assured Forwarding service: what about the results?
Since the proposition of Quality of Service architectures by the IETF, the
interaction between TCP and the QoS services has been intensively studied. This
paper proposes to look forward to the results obtained in terms of TCP
throughput guarantee in the DiffServ Assured Forwarding (DiffServ/AF) service
and to present an overview of the different proposals to solve the problem. It
has been demonstrated that the standardized IETF DiffServ conditioners such as
the token bucket color marker and the time sliding window color maker were not
good TCP traffic descriptors. Starting with this point, several propositions
have been made and most of them presents new marking schemes in order to
replace or improve the traditional token bucket color marker. The main problem
is that TCP congestion control is not designed to work with the AF service.
Indeed, both mechanisms are antagonists. TCP has the property to share in a
fair manner the bottleneck bandwidth between flows while DiffServ network
provides a level of service controllable and predictable. In this paper, we
build a classification of all the propositions made during these last years and
compare them. As a result, we will see that these conditioning schemes can be
separated in three sets of action level and that the conditioning at the
network edge level is the most accepted one. We conclude that the problem is
still unsolved and that TCP, conditioned or not conditioned, remains
inappropriate to the DiffServ/AF service
Quality of Service over Specific Link Layers: state of the art report
The Integrated Services concept is proposed as an enhancement to the current Internet architecture, to provide a better Quality of Service (QoS) than that provided by the traditional Best-Effort service. The features of the Integrated Services are explained in this report. To support Integrated Services, certain requirements are posed on the underlying link layer. These requirements are studied by the Integrated Services over Specific Link Layers (ISSLL) IETF working group. The status of this ongoing research is reported in this document. To be more specific, the solutions to provide Integrated Services over ATM, IEEE 802 LAN technologies and low-bitrate links are evaluated in detail. The ISSLL working group has not yet studied the requirements, that are posed on the underlying link layer, when this link layer is wireless. Therefore, this state of the art report is extended with an identification of the requirements that are posed on the underlying wireless link, to provide differentiated Quality of Service
Recommended from our members
Towards scalable end-to-end QoS provision for VoIP applications
The growth of the Internet and the development of its new applications have increased the demand for providing a certain level of resource assurance and service support. The concept of ensuring quality of service (QoS) has been introduced in order to provide the support and assurance for these services. Different QoS mechanisms, such as integrated services (IntServ) and differentiated services (DiffServ), have been developed and introduced to provide different levels of QoS provision. However, IntServ can suffer from scalability issues that make it infeasible for large-scale network implementations. On the other hand, the aggregated-based per-flow technique of DiffServ does not provide such an end-to-end QoS guarantee. Recently, the IETF have proposed a new QoS architecture that implements IntServ over DiffServ in order to provide an end-to-end QoS for scalable networks. Hence, it became possible to provide and support a certain level of QoS for some delay sensitive and bandwidth-demanding applications such as voice over Internet Protocol (VoIP). With regard to VoIP applications, delay, jitter and packet loss are crucial issues that have to be taken into consideration for any VoIP system design and such parameters need a distinct level of QoS support
- …