1,323 research outputs found

    IETF standardization in the field of the Internet of Things (IoT): a survey

    Get PDF
    Smart embedded objects will become an important part of what is called the Internet of Things. However, the integration of embedded devices into the Internet introduces several challenges, since many of the existing Internet technologies and protocols were not designed for this class of devices. In the past few years, there have been many efforts to enable the extension of Internet technologies to constrained devices. Initially, this resulted in proprietary protocols and architectures. Later, the integration of constrained devices into the Internet was embraced by IETF, moving towards standardized IP-based protocols. In this paper, we will briefly review the history of integrating constrained devices into the Internet, followed by an extensive overview of IETF standardization work in the 6LoWPAN, ROLL and CoRE working groups. This is complemented with a broad overview of related research results that illustrate how this work can be extended or used to tackle other problems and with a discussion on open issues and challenges. As such the aim of this paper is twofold: apart from giving readers solid insights in IETF standardization work on the Internet of Things, it also aims to encourage readers to further explore the world of Internet-connected objects, pointing to future research opportunities

    Less-than-Best-Effort capacity sharing over high BDP networks with LEDBAT

    Get PDF
    There has been a renewed interest at the Internet Engineering Task Force (IETF) in using Less-than-Best Effort (LBE) methods for background applications. IETF recently published a RFC for Low Extra Delay Background Transport (LEDBAT), a congestion control algorithm for LBE transmissions. This paper provides an analysis of LEDBAT performance over congested large bandwidth X delay product (LBDP) networks, and assesses the validity of having a fixed target queuing time. In particular, we lead a study of the impact of this target queuing delay when LEDBAT is used over 4G satellite networks. The rationale is to explore the possibility to grab the unused 4G satellite links' capacity to carry non-commercial traffic. We show that this is achievable with LEDBAT. However, depending on the fluctuation of the load, performance improvements could be obtained by properly setting the target value. We generalize this evaluation over different congested LBDP networks and confirm that the target value might need to be adjusted to networks' and traffic's characteristics. Further work will study whether and how this parameter should be dynamically adapted, and LEDBAT's congestion control improved

    User-Centric Quality of Service Provisioning in IP Networks

    Get PDF
    The Internet has become the preferred transport medium for almost every type of communication, continuing to grow, both in terms of the number of users and delivered services. Efforts have been made to ensure that time sensitive applications receive sufficient resources and subsequently receive an acceptable Quality of Service (QoS). However, typical Internet users no longer use a single service at a given point in time, as they are instead engaged in a multimedia-rich experience, comprising of many different concurrent services. Given the scalability problems raised by the diversity of the users and traffic, in conjunction with their increasing expectations, the task of QoS provisioning can no longer be approached from the perspective of providing priority to specific traffic types over coexisting services; either through explicit resource reservation, or traffic classification using static policies, as is the case with the current approach to QoS provisioning, Differentiated Services (Diffserv). This current use of static resource allocation and traffic shaping methods reveals a distinct lack of synergy between current QoS practices and user activities, thus highlighting a need for a QoS solution reflecting the user services. The aim of this thesis is to investigate and propose a novel QoS architecture, which considers the activities of the user and manages resources from a user-centric perspective. The research begins with a comprehensive examination of existing QoS technologies and mechanisms, arguing that current QoS practises are too static in their configuration and typically give priority to specific individual services rather than considering the user experience. The analysis also reveals the potential threat that unresponsive application traffic presents to coexisting Internet services and QoS efforts, and introduces the requirement for a balance between application QoS and fairness. This thesis proposes a novel architecture, the Congestion Aware Packet Scheduler (CAPS), which manages and controls traffic at the point of service aggregation, in order to optimise the overall QoS of the user experience. The CAPS architecture, in contrast to traditional QoS alternatives, places no predetermined precedence on a specific traffic; instead, it adapts QoS policies to each individual’s Internet traffic profile and dynamically controls the ratio of user services to maintain an optimised QoS experience. The rationale behind this approach was to enable a QoS optimised experience to each Internet user and not just those using preferred services. Furthermore, unresponsive bandwidth intensive applications, such as Peer-to-Peer, are managed fairly while minimising their impact on coexisting services. The CAPS architecture has been validated through extensive simulations with the topologies used replicating the complexity and scale of real-network ISP infrastructures. The results show that for a number of different user-traffic profiles, the proposed approach achieves an improved aggregate QoS for each user when compared with Best effort Internet, Traditional Diffserv and Weighted-RED configurations. Furthermore, the results demonstrate that the proposed architecture not only provides an optimised QoS to the user, irrespective of their traffic profile, but through the avoidance of static resource allocation, can adapt with the Internet user as their use of services change.France Teleco

    Mesmerizer: A Effective Tool for a Complete Peer-to-Peer Software Development Life-cycle

    Get PDF
    In this paper we present what are, in our experience, the best practices in Peer-To-Peer(P2P) application development and how we combined them in a middleware platform called Mesmerizer. We explain how simulation is an integral part of the development process and not just an assessment tool. We then present our component-based event-driven framework for P2P application development, which can be used to execute multiple instances of the same application in a strictly controlled manner over an emulated network layer for simulation/testing, or a single application in a concurrent environment for deployment purpose. We highlight modeling aspects that are of critical importance for designing and testing P2P applications, e.g. the emulation of Network Address Translation and bandwidth dynamics. We show how our simulator scales when emulating low-level bandwidth characteristics of thousands of concurrent peers while preserving a good degree of accuracy compared to a packet-level simulator

    Application acceleration for wireless and mobile data networks

    Get PDF
    This work studies application acceleration for wireless and mobile data networks. The problem of accelerating application can be addressed along multiple dimensions. The first dimension is advanced network protocol design, i.e., optimizing underlying network protocols, particulary transport layer protocol and link layer protocol. Despite advanced network protocol design, in this work we observe that certain application behaviors can fundamentally limit the performance achievable when operating over wireless and mobile data networks. The performance difference is caused by the complex application behaviors of these non-FTP applications. Explicitly dealing with application behaviors can improve application performance for new environments. Along this overcoming application behavior dimension, we accelerate applications by studying specific types of applications including Client-server, Peer-to-peer and Location-based applications. In exploring along this dimension, we identify a set of application behaviors that significantly affect application performance. To accommodate these application behaviors, we firstly extract general design principles that can apply to any applications whenever possible. These design principles can also be integrated into new application designs. We also consider specific applications by applying these design principles and build prototypes to demonstrate the effectiveness of the solutions. In the context of application acceleration, even though all the challenges belong to the two aforementioned dimensions of advanced network protocol design and overcoming application behavior are addressed, application performance can still be limited by the underlying network capability, particularly physical bandwidth. In this work, we study the possibility of speeding up data delivery by eliminating traffic redundancy present in application traffics. Specifically, we first study the traffic redundancy along multiple dimensions using traces obtained from multiple real wireless network deployments. Based on the insights obtained from the analysis, we propose Wireless Memory (WM), a two-ended AP-client solution to effectively exploit traffic redundancy in wireless and mobile environments. Application acceleration can be achieved along two other dimensions: network provision ing and quality of service (QoS). Network provisioning allocates network resources such as physical bandwidth or wireless spectrum, while QoS provides different priority to different applications, users, or data flows. These two dimensions have their respective limitations in the context of application acceleration. In this work, we focus on the two dimensions of overcoming application behavior and Eliminating traffic redundancy to improve application performance. The contribution of this work is as follows. First, we study the problem of application acceleration for wireless and mobile data networks, and we characterize the dimensions along which to address the problem. Second, we identify that application behaviors can significantly affect application performance, and we propose a set of design principles to deal with the behaviors. We also build prototypes to conduct system research. Third, we consider traffic redundancy elimination and propose a wireless memory approach.Ph.D.Committee Chair: Sivakumar, Raghupathy; Committee Member: Ammar, Mostafa; Committee Member: Fekri, Faramarz; Committee Member: Ji, Chuanyi; Committee Member: Ramachandran, Umakishor

    Performance Evaluation of MPLS in a Virtualized Service Provider Core (with/without Class of Service)

    Get PDF
    The last decade has witnessed a major change in the types of traffic scaling the Internet. With the development of real-time applications several challenges were faced within traditional IP networks. Some of these challenges are delay, increased costs faced by the service provider and customer, limited scalability, separate infrastructure costs and high administrative overheads to manage large networks etc. To combat these challenges, researchers have steered towards finding alternate solutions. Over the recent years, we have seen an introduction of a number of virtualized platforms and solutions being offered in the networking industry. Virtual load balancers, virtual firewalls, virtual routers, virtual intrusion detection and preventions systems are just a few examples within the Network Function Virtualization world! Service Providers are trying to find solutions where they could reduce operational expenses while at the same time meet the growing bandwidth demands of their customers. The main aim of this thesis is to evaluate the performance of voice, data and video traffic in a virtualized service provider core. Observations are made on how these traffic types perform on congested vs uncongested links and how Quality of Service treats traffic in a virtualized Service Provider Core using Round Trip Time as a performance metric. This thesis also tries to find if resiliency features such as Fast Reroute provide an additional advantage in failover scenarios within virtualized service provider cores. Juniper Networks vSRX are used to replicate virtual routers in a virtualized service provider core. Twenty-Four tests are carried out to gain a better understanding of how real-time applications and resiliency methods perform in virtualized networks. It is observed that a trade-off exists when introducing QoS on congested primary and secondary label switched paths. What can be observed thru the graphs is having Quality of Service enabled drops more packets however gives us the advantage of lower Round Trip Time for in-profile traffic. On the hand, having Quality of Service disabled, permits more traffic but leads to bandwidth contention between the three traffic classes leading to higher Round-Trip Times. The true benefit of QoS is seen in traffic congestion scenarios. The test bed built in this thesis, shows us that Fast Reroute does not add a significant benefit to aid in the reduction of packet loss during failover scenarios between primary and secondary paths. However, in certain scenarios fast reroute does seem to reduce packet loss specifically for data traffic
    • 

    corecore