696 research outputs found
Quality of service assurance for the next generation Internet
The provisioning for multimedia applications has been of increasing interest among researchers and Internet Service Providers. Through the migration from resource-based to service-driven networks, it has become evident that the Internet model should be enhanced to provide support for a variety of differentiated services that match applications and customer requirements, and not stay limited under the flat best-effort service that is currently provided.
In this paper, we describe and critically appraise the major achievements of the efforts to introduce Quality of Service (QoS) assurance and provisioning within the Internet model. We then propose a research path for the creation of a network services management architecture,
through which we can move towards a QoS-enabled network environment, offering support for a variety of different services, based on traffic characteristics and user expectations
Efficient resource management for end-to-end QoS guarantees in DiffServ networks
The Differentiated Services (DiffServ) architecture has been proposed as a scalable solution for delivering end-to-end Quality of Service (QoS) guarantees over the Internet. While the scalability of the data plane emerges from the definition of only a small number of different service classes, the issue of a scalable control plane is still an open research problem. The initial proposal was to use a centralized agent, called Bandwidth Broker (BB), to manage the resources within each DiffServ domain and make local admission control decisions. In this paper, we propose an alternative distributed approach, where the local admission decisions are made independently at the edge routers of each domain. We will show, through simulation results, that this distributed approach can manage the network resources very efficiently, leading to lower bandwidth blocking rates when compared to traditional shortest path admission control. Moreover, its simplicity and distributed implementation make it a very scalable solution for resource management in DiffServ networks.published_or_final_versio
Design and Implementation of a Measurement-Based Policy-Driven Resource Management Framework For Converged Networks
This paper presents the design and implementation of a measurement-based QoS
and resource management framework, CNQF (Converged Networks QoS Management
Framework). CNQF is designed to provide unified, scalable QoS control and
resource management through the use of a policy-based network management
paradigm. It achieves this via distributed functional entities that are
deployed to co-ordinate the resources of the transport network through
centralized policy-driven decisions supported by measurement-based control
architecture. We present the CNQF architecture, implementation of the prototype
and validation of various inbuilt QoS control mechanisms using real traffic
flows on a Linux-based experimental test bed.Comment: in Ictact Journal On Communication Technology: Special Issue On Next
Generation Wireless Networks And Applications, June 2011, Volume 2, Issue 2,
Issn: 2229-6948(Online
A Survey on the Contributions of Software-Defined Networking to Traffic Engineering
Since the appearance of OpenFlow back in 2008, software-defined networking (SDN) has gained momentum. Although there are some discrepancies between the standards developing organizations working with SDN about what SDN is and how it is defined, they all outline traffic engineering (TE) as a key application. One of the most common objectives of TE is the congestion minimization, where techniques such as traffic splitting among multiple paths or advanced reservation systems are used. In such a scenario, this manuscript surveys the role of a comprehensive list of SDN protocols in TE solutions, in order to assess how these protocols can benefit TE. The SDN protocols have been categorized using the SDN architecture proposed by the open networking foundation, which differentiates among data-controller plane interfaces, application-controller plane interfaces, and management interfaces, in order to state how the interface type in which they operate influences TE. In addition, the impact of the SDN protocols on TE has been evaluated by comparing them with the path computation element (PCE)-based architecture. The PCE-based architecture has been selected to measure the impact of SDN on TE because it is the most novel TE architecture until the date, and because it already defines a set of metrics to measure the performance of TE solutions. We conclude that using the three types of interfaces simultaneously will result in more powerful and enhanced TE solutions, since they benefit TE in complementary ways.European Commission through the Horizon 2020 Research and Innovation Programme (GN4) under Grant 691567
Spanish Ministry of Economy and Competitiveness under the Secure Deployment of Services Over SDN and NFV-based Networks Project S&NSEC under Grant TEC2013-47960-C4-3-
Recommended from our members
Multimedia delivery in the future internet
The term “Networked Media” implies that all kinds of media including text, image, 3D graphics, audio
and video are produced, distributed, shared, managed and consumed on-line through various networks,
like the Internet, Fiber, WiFi, WiMAX, GPRS, 3G and so on, in a convergent manner [1]. This white
paper is the contribution of the Media Delivery Platform (MDP) cluster and aims to cover the Networked
challenges of the Networked Media in the transition to the Future of the Internet.
Internet has evolved and changed the way we work and live. End users of the Internet have been confronted
with a bewildering range of media, services and applications and of technological innovations concerning
media formats, wireless networks, terminal types and capabilities. And there is little evidence that the pace
of this innovation is slowing. Today, over one billion of users access the Internet on regular basis, more
than 100 million users have downloaded at least one (multi)media file and over 47 millions of them do so
regularly, searching in more than 160 Exabytes1 of content. In the near future these numbers are expected
to exponentially rise. It is expected that the Internet content will be increased by at least a factor of 6, rising
to more than 990 Exabytes before 2012, fuelled mainly by the users themselves. Moreover, it is envisaged
that in a near- to mid-term future, the Internet will provide the means to share and distribute (new)
multimedia content and services with superior quality and striking flexibility, in a trusted and personalized
way, improving citizens’ quality of life, working conditions, edutainment and safety.
In this evolving environment, new transport protocols, new multimedia encoding schemes, cross-layer inthe
network adaptation, machine-to-machine communication (including RFIDs), rich 3D content as well as
community networks and the use of peer-to-peer (P2P) overlays are expected to generate new models of
interaction and cooperation, and be able to support enhanced perceived quality-of-experience (PQoE) and
innovative applications “on the move”, like virtual collaboration environments, personalised services/
media, virtual sport groups, on-line gaming, edutainment. In this context, the interaction with content
combined with interactive/multimedia search capabilities across distributed repositories, opportunistic P2P
networks and the dynamic adaptation to the characteristics of diverse mobile terminals are expected to
contribute towards such a vision.
Based on work that has taken place in a number of EC co-funded projects, in Framework Program 6 (FP6)
and Framework Program 7 (FP7), a group of experts and technology visionaries have voluntarily
contributed in this white paper aiming to describe the status, the state-of-the art, the challenges and the way
ahead in the area of Content Aware media delivery platforms
An Efficient Algorithm for Congestion Control in Highly Loaded DiffServ/MPLS Networks
Optimal QoS path provisioning of coexisted and aggregated traffic in networks is still demanding problem. All traffic flows in a domain are distributed among LSPs (Label Switching Path) related to N service classes, but the congestion problem of concurrent flows can appear. As we know the IGP (Interior Getaway Protocol) uses simple on-line routing algorithms (e.g. OSPFS, IS-IS) based on shortest path methodology. In QoS end-to-end provisioning where some links may be reserved for certain traffic classes (for particular set of users) it becomes insufficient technique. On other hand, constraint-based explicit routing (CR) based on IGP metric ensures traffic engineering (TE) capabilities. But in overloaded and poorly connected MPLS/DiffServ networks the CR becomes insufficient technique. As we need firm correlation with bandwidth management and traffic engineering (TE) the initial (pro-active) routing can be pre-computed in the context of all priority traffic flows (former contracted SLAs) traversing the network simultaneously. It mean that LSP can be pre-computed much earlier, possibly during SLA (Service Level Agreement) negotiation process. In the paper a new load simulation technique for load balancing control purpose is proposed. The algorithm proposed in the paper may find a longer but lightly loaded path, better than the heavily loaded shortest path. It could be a very good solution for congestion avoidance and for better load-balancing purpose where links are running close to capacity. Also, such technique could be useful in inter-domain end-to-end provisioning, where bandwidth reservation has to be negotiated with neighbor ASes (Autonomous System). To be acceptable for real applications such complicated routing algorithm can be significantly improved. Algorithm was tested on the network of M core routers on the path (between edge routers) and results are given for N=3 service classes. Further improvements through heuristic approach are made and results are discussed
- …