3,186 research outputs found

    The troubled journey of QoS: From ATM to content networking, edge-computing and distributed internet governance

    Get PDF
    Network Quality of Service (QoS) and the associated user Quality of Experience (QoE) have always been the networking “holy grail” and have been sought after through various different approaches and networking technologies over the last decades. Despite substantial amounts of effort invested in the area, there has been very little actual deployment of mechanisms to guarantee QoS in the Internet. As a result, the Internet is largely operating on a “best effort” basis in terms of QoS. Here, we attempt a historical overview in order to better understand how we got to the point where we are today and consider the evolution of QoS/QoE in the future. As we move towards more demanding networking environments where enormous amounts of data is produced at the edge of the network (e.g., from IoT devices), computation will also need to migrate to the edge in order to guarantee QoS. In turn, we argue that distributed computing at the edge of the network will inevitably require infrastructure decentralisation. That said, trust to the infrastructure provider is more difficult to guarantee and new components need to be incorporated into the Internet landscape in order to be able to support emerging applications, but also achieve acceptable service quality. We start from the first steps of ATM and related IP-based technologies, we consider recent proposals for content-oriented and Information-Centric Networking, mobile edge and fog computing, and finally we see how distributed Internet governance through Distributed Ledger Technology and blockchains can influence QoS in future networks

    Overlay networks for smart grids

    Get PDF

    User-Centric Quality of Service Provisioning in IP Networks

    Get PDF
    The Internet has become the preferred transport medium for almost every type of communication, continuing to grow, both in terms of the number of users and delivered services. Efforts have been made to ensure that time sensitive applications receive sufficient resources and subsequently receive an acceptable Quality of Service (QoS). However, typical Internet users no longer use a single service at a given point in time, as they are instead engaged in a multimedia-rich experience, comprising of many different concurrent services. Given the scalability problems raised by the diversity of the users and traffic, in conjunction with their increasing expectations, the task of QoS provisioning can no longer be approached from the perspective of providing priority to specific traffic types over coexisting services; either through explicit resource reservation, or traffic classification using static policies, as is the case with the current approach to QoS provisioning, Differentiated Services (Diffserv). This current use of static resource allocation and traffic shaping methods reveals a distinct lack of synergy between current QoS practices and user activities, thus highlighting a need for a QoS solution reflecting the user services. The aim of this thesis is to investigate and propose a novel QoS architecture, which considers the activities of the user and manages resources from a user-centric perspective. The research begins with a comprehensive examination of existing QoS technologies and mechanisms, arguing that current QoS practises are too static in their configuration and typically give priority to specific individual services rather than considering the user experience. The analysis also reveals the potential threat that unresponsive application traffic presents to coexisting Internet services and QoS efforts, and introduces the requirement for a balance between application QoS and fairness. This thesis proposes a novel architecture, the Congestion Aware Packet Scheduler (CAPS), which manages and controls traffic at the point of service aggregation, in order to optimise the overall QoS of the user experience. The CAPS architecture, in contrast to traditional QoS alternatives, places no predetermined precedence on a specific traffic; instead, it adapts QoS policies to each individual’s Internet traffic profile and dynamically controls the ratio of user services to maintain an optimised QoS experience. The rationale behind this approach was to enable a QoS optimised experience to each Internet user and not just those using preferred services. Furthermore, unresponsive bandwidth intensive applications, such as Peer-to-Peer, are managed fairly while minimising their impact on coexisting services. The CAPS architecture has been validated through extensive simulations with the topologies used replicating the complexity and scale of real-network ISP infrastructures. The results show that for a number of different user-traffic profiles, the proposed approach achieves an improved aggregate QoS for each user when compared with Best effort Internet, Traditional Diffserv and Weighted-RED configurations. Furthermore, the results demonstrate that the proposed architecture not only provides an optimised QoS to the user, irrespective of their traffic profile, but through the avoidance of static resource allocation, can adapt with the Internet user as their use of services change.France Teleco

    Decoupling Information and Connectivity via Information-Centric Transport

    Get PDF
    The power of Information-Centric Networking (ICN) architectures lies in their abstraction for communication --- the request for named data. This abstraction promises that applications can choose to operate only in the information plane, agnostic to the mechanisms implemented in the connectivity plane. However, despite this powerful promise, the information and connectivity planes are presently coupled in today\u27s incarnations of leading ICNs by a core architectural component, the forwarding strategy. Presently, this component is not sustainable: it implements both the information and connectivity mechanisms without specifying who should choose a forwarding strategy --- an application developer or the network operator. In practice, application developers can specify a strategy only if they understand connectivity details, while network operators can assign strategies only if they understand application expectations. In this paper, we define the role of forwarding strategies, and we introduce Information-Centric Transport (ICT) as an abstraction for cleanly decoupling the information plane from the connectivity plane. We discuss how ICTs allow applications to operate in the information plane, concerned only with namespaces and trust identities, leaving network node operators free to deploy whatever strategy mechanisms make sense for the connectivity that they manage. To illustrate the ICT concept, we demonstrate ICT-Sync and ICT-Notify. We show how these ICTs 1) enable applications to operate regardless of connectivity details, 2) are designed to satisfy a predefined set of application requirements and are free from application-specifics, and 3) can be deployed by network operators where needed, without requiring any change to the application logic
    • …
    corecore