141 research outputs found
Analysis of Multiple Flows using Different High Speed TCP protocols on a General Network
We develop analytical tools for performance analysis of multiple TCP flows
(which could be using TCP CUBIC, TCP Compound, TCP New Reno) passing through a
multi-hop network. We first compute average window size for a single TCP
connection (using CUBIC or Compound TCP) under random losses. We then consider
two techniques to compute steady state throughput for different TCP flows in a
multi-hop network. In the first technique, we approximate the queues as M/G/1
queues. In the second technique, we use an optimization program whose solution
approximates the steady state throughput of the different flows. Our results
match well with ns2 simulations.Comment: Submitted to Performance Evaluatio
Model based analysis of some high speed network issues
The study of complex problems in science and engineering today typically involves large scale data, huge number of large-scale scientific breakthroughs critically depends on large multi-disciplinary and geographically-dispersed research teams, where the high speed network becomes the integral part. To serve the ongoing bandwidth requirement and scalability of these networks, there has been a continuous evolution of different TCPs for high speed networks. Testing these protocols on a real network would be expensive, time consuming and more over not easily available to the researchers worldwide. Network simulation is well accepted and widely used method for performance evaluation, it is well known that packet-based simulators like NS2 and Opnet are not adequate in high speed also in large scale networks because of its inherent bottlenecks in terms of message overhead and execution time. In that case model based approach with the help of a set of coupled differential equations is preferred for simulations. This dissertation is focused on the key challenges on research and development of TCPs on high-speed network. To address these issues/challenges this thesis has three objectives: design an analytical simulation methodology; model behaviors of high speed networks and other components including TCP flows and queue using the analytical simulation method; analyze them and explore impacts and interrelationship among them. To decrease the simulation time and speed up the process of testing and development of high speed TCP, we present a scalable simulation methodology for high speed network. We present the fluid model equations for various high-speed TCP variants. With the help of these fluid model equations, the behavior of high-speed TCP variants under various scenarios and its effect on queue size variations are presented. High speed network is not feasible unless we understand effect of bottleneck buffer size on performance of these high-speed TCP variants. A fluid model is introduced to accommodate the new observations of synchronization and de-synchronization phenomena of packet losses at bottleneck link and a microscopic analysis is presented on different buffer sizes at drop-tail queuing scheme. The proposed model based methods promotes principal understanding of the future heterogeneous networks and accelerates protocol developments
Equilibrium of Heterogeneous Congestion Control: Optimality and Stability
When heterogeneous congestion control protocols
that react to different pricing signals share the same network,
the current theory based on utility maximization fails to predict
the network behavior. The pricing signals can be different types
of signals such as packet loss, queueing delay, etc, or different
values of the same type of signal such as different ECN marking
values based on the same actual link congestion level. Unlike in a
homogeneous network, the bandwidth allocation now depends on
router parameters and flow arrival patterns. It can be non-unique,
suboptimal and unstable. In Tang et al. (“Equilibrium of heterogeneous
congestion control: Existence and uniqueness,” IEEE/ACM
Trans. Netw., vol. 15, no. 4, pp. 824–837, Aug. 2007), existence and
uniqueness of equilibrium of heterogeneous protocols are investigated.
This paper extends the study with two objectives: analyzing
the optimality and stability of such networks and designing control
schemes to improve those properties. First, we demonstrate the
intricate behavior of a heterogeneous network through simulations
and present a framework to help understand its equilibrium
properties. Second, we propose a simple source-based algorithm
to decouple bandwidth allocation from router parameters and
flow arrival patterns by only updating a linear parameter in the
sources’ algorithms on a slow timescale. It steers a network to
the unique optimal equilibrium. The scheme can be deployed
incrementally as the existing protocol needs no change and only
new protocols need to adopt the slow timescale adaptation
Agile-SD: A Linux-based TCP Congestion Control Algorithm for Supporting High-speed and Short-distance Networks
Recently, high-speed and short-distance networks are widely deployed and
their necessity is rapidly increasing everyday. This type of networks is used
in several network applications; such as Local Area Networks (LAN) and Data
Center Networks (DCN). In LANs and DCNs, high-speed and short-distance networks
are commonly deployed to connect between computing and storage elements in
order to provide rapid services. Indeed, the overall performance of such
networks is significantly influenced by the Congestion Control Algorithm (CCA)
which suffers from the problem of bandwidth under-utilization, especially if
the applied buffer regime is very small. In this paper, a novel loss-based CCA
tailored for high-speed and Short-Distance (SD) networks, namely Agile-SD, has
been proposed. The main contribution of the proposed CCA is to implement the
mechanism of agility factor. Further, intensive simulation experiments have
been carried out to evaluate the performance of Agile-SD compared to Compound
and Cubic which are the default CCAs of the most commonly used operating
systems. The results of the simulation experiments show that the proposed CCA
outperforms the compared CCAs in terms of average throughput, loss ratio and
fairness, especially when a small buffer is applied. Moreover, Agile-SD shows
lower sensitivity to the buffer size change and packet error rate variation
which increases its efficiency.Comment: 12 Page
A study on fairness and latency issues over high speed networks and data center networks
Newly emerging computer networks, such as high speed networks and data center networks, have characteristics of high bandwidth and high burstiness which make it difficult to address issues such as fairness, queuing latency and link utilization. In this study, we first conduct extensive experimental evaluation of the performance of 10Gbps high speed networks. We found inter-protocol unfairness and larger queuing latency are two outstanding issues in high speed networks and data center networks. There have been several proposals to address fairness and latency issues at switch level via queuing schemes. These queuing schemes have been fairly successful in addressing either fairness issue or large latency but not both at the same time. We propose a new queuing scheme called Approximated-Fair and Controlled-Delay (AFCD) queuing scheme that meets following goals for high speed networks: approximated fairness, controlled low queuing delay, high link utilization and simple implementation. The design of AFCD utilizes a novel synergistic approach by forming an alliance between approximated fair queuing and controlled delay queuing. AFCD maintains very small amount of state information in sending rate estimation of flows and makes drop decision based on a target delay of individual flow. We then present FaLL, a Fair and Low Latency queuing scheme that meets stringent performance requirements of data center networks: fair share of bandwidth, low queuing latency, high throughput, and ease of deployment. FaLL uses an efficiency module, a fairness module and a target delay based dropping scheme to meet these goals. Through rigorous experiments on real testbed, we show that FaLL outperforms various peer solutions in variety of network conditions over data center networks
Study on the Performance of TCP over 10Gbps High Speed Networks
Internet traffic is expected to grow phenomenally over the next five to ten years. To cope with such large traffic volumes, high-speed networks are expected to scale to capacities of terabits-per-second and beyond. Increasing the role of optics for packet forwarding and transmission inside the high-speed networks seems to be the most promising way to accomplish this capacity scaling. Unfortunately, unlike electronic memory, it remains a formidable challenge to build even a few dozen packets of integrated all-optical buffers. On the other hand, many high-speed networks depend on the TCP/IP protocol for reliability which is typically implemented in software and is sensitive to buffer size. For example, TCP requires a buffer size of bandwidth delay product in switches/routers to maintain nearly 100\% link utilization. Otherwise, the performance will be much downgraded. But such large buffer will challenge hardware design and power consumption, and will generate queuing delay and jitter which again cause problems. Therefore, improve TCP performance over tiny buffered high-speed networks is a top priority. This dissertation studies the TCP performance in 10Gbps high-speed networks. First, a 10Gbps reconfigurable optical networking testbed is developed as a research environment. Second, a 10Gbps traffic sniffing tool is developed for measuring and analyzing TCP performance. New expressions for evaluating TCP loss synchronization are presented by carefully examining the congestion events of TCP. Based on observation, two basic reasons that cause performance problems are studied. We find that minimize TCP loss synchronization and reduce flow burstiness impact are critical keys to improve TCP performance in tiny buffered networks. Finally, we present a new TCP protocol called Multi-Channel TCP and a new congestion control algorithm called Desynchronized Multi-Channel TCP (DMCTCP). Our algorithm implementation takes advantage of a potential parallelism from the Multi-Path TCP in Linux. Over an emulated 10Gbps network ruled by routers with only a few dozen packets of buffers, our experimental results confirm that bottleneck link utilization can be much better improved by DMCTCP than by many other TCP variants. Our study is a new step towards the deployment of optical packet switching/routing networks
Transport Layer solution for bulk data transfers over Heterogeneous Long Fat Networks in Next Generation Networks
Aquesta tesi per compendi centra les seves contribucions en l'aprenentatge i innovació de les Xarxes de Nova Generació. És per això que es proposen diferents contribucions en diferents àmbits (Smart Cities, Smart Grids, Smart Campus, Smart Learning, Mitjana, eHealth, Indústria 4.0 entre d'altres) mitjançant l'aplicació i combinació de diferents disciplines (Internet of Things, Building Information Modeling, Cloud Storage, Ciberseguretat, Big Data, Internet de el Futur, Transformació Digital).
Concretament, es detalla el monitoratge sostenible del confort a l'Smart Campus, la que potser es la meva aportació més representativa dins de la conceptualització de Xarxes de Nova Generació. Dins d'aquest innovador concepte de monitorització s'integren diferents disciplines, per poder oferir informació sobre el nivell de confort de les persones. Aquesta investigació demostra el llarg recorregut que hi ha en la transformació digital dels sectors tradicionals i les NGNs.
Durant aquest llarg aprenentatge sobre les NGN a través de les diferents investigacions, es va poder observar una problemàtica que afectava de manera transversal als diferents camps d'aplicació de les NGNs i que aquesta podia tenir una afectació en aquests sectors. Aquesta problemàtica consisteix en el baix rendiment durant l'intercanvi de grans volums de dades sobre xarxes amb gran capacitat d'ample de banda i remotament separades geogràficament, conegudes com a xarxes elefant. Concretament, això afecta al cas d'ús d'intercanvi massiu de dades entre regions Cloud (Cloud Data Sharing use case). És per això que es va estudiar aquest cas d'ús i les diferents alternatives a nivell de protocols de transport,. S'estudien les diferents problemàtiques que pateixen els protocols i s'observa per què aquests no són capaços d'arribar a rendiments òptims.
Deguda a aquesta situació, s'hipotetiza que la introducció de mecanismes que analitzen les mètriques de la xarxa i que exploten eficientment la capacitat de la mateixa milloren el rendiment dels protocols de transport sobre xarxes elefant heterogènies durant l'enviament massiu de dades.
Primerament, es dissenya l’Adaptative and Aggressive Transport Protocol (AATP), un protocol de transport adaptatiu i eficient amb l'objectiu de millorar el rendiment sobre aquest tipus de xarxes elefant. El protocol AATP s'implementa i es prova en un simulador de xarxes i un testbed sota diferents situacions i condicions per la seva validació.
Implementat i provat amb èxit el protocol AATP, es decideix millorar el propi protocol, Enhanced-AATP, sobre xarxes elefant heterogènies. Per això, es dissenya un mecanisme basat en el Jitter Ràtio que permet fer aquesta diferenciació. A més, per tal de millorar el comportament del protocol, s’adapta el seu sistema de fairness per al repartiment just dels recursos amb altres fluxos Enhanced-AATP. Aquesta evolució s'implementa en el simulador de xarxes i es realitzen una sèrie de proves.
A l'acabar aquesta tesi, es conclou que les Xarxes de Nova Generació tenen molt recorregut i moltes coses a millorar causa de la transformació digital de la societat i de l'aparició de nova tecnologia disruptiva. A més, es confirma que la introducció de mecanismes específics en la concepció i operació dels protocols de transport millora el rendiment d'aquests sobre xarxes elefant heterogènies.Esta tesis por compendio centra sus contribuciones en el aprendizaje e innovación de las Redes de Nueva Generación. Es por ello que se proponen distintas contribuciones en diferentes ámbitos (Smart Cities, Smart Grids, Smart Campus, Smart Learning, Media, eHealth, Industria 4.0 entre otros) mediante la aplicación y combinación de diferentes disciplinas (Internet of Things, Building Information Modeling, Cloud Storage, Ciberseguridad, Big Data, Internet del Futuro, Transformación Digital).
Concretamente, se detalla la monitorización sostenible del confort en el Smart Campus, la que se podría considerar mi aportación más representativa dentro de la conceptualización de Redes de Nueva Generación. Dentro de este innovador concepto de monitorización se integran diferentes disciplinas, para poder ofrecer información sobre el nivel de confort de las personas. Esta investigación demuestra el recorrido que existe en la transformación digital de los sectores tradicionales y las NGNs.
Durante este largo aprendizaje sobre las NGN a través de las diferentes investigaciones, se pudo observar una problemática que afectaba de manera transversal a los diferentes campos de aplicación de las NGNs y que ésta podía tener una afectación en estos sectores. Esta problemática consiste en el bajo rendimiento durante el intercambio de grandes volúmenes de datos sobre redes con gran capacidad de ancho de banda y remotamente separadas geográficamente, conocidas como redes elefante, o Long Fat Networks (LFNs). Concretamente, esto afecta al caso de uso de intercambio de datos entre regiones Cloud (Cloud Data Data use case). Es por ello que se estudió este caso de uso y las diferentes alternativas a nivel de protocolos de transporte. Se estudian las diferentes problemáticas que sufren los protocolos y se observa por qué no son capaces de alcanzar rendimientos óptimos.
Debida a esta situación, se hipotetiza que la introducción de mecanismos que analizan las métricas de la red y que explotan eficientemente la capacidad de la misma mejoran el rendimiento de los protocolos de transporte sobre redes elefante heterogéneas durante el envío masivo de datos.
Primeramente, se diseña el Adaptative and Aggressive Transport Protocol (AATP), un protocolo de transporte adaptativo y eficiente con el objetivo maximizar el rendimiento sobre este tipo de redes elefante. El protocolo AATP se implementa y se prueba en un simulador de redes y un testbed bajo diferentes situaciones y condiciones para su validación.
Implementado y probado con éxito el protocolo AATP, se decide mejorar el propio protocolo, Enhanced-AATP, sobre redes elefante heterogéneas. Además, con tal de mejorar el comportamiento del protocolo, se mejora su sistema de fairness para el reparto justo de los recursos con otros flujos Enhanced-AATP. Esta evolución se implementa en el simulador de redes y se realizan una serie de pruebas.
Al finalizar esta tesis, se concluye que las Redes de Nueva Generación tienen mucho recorrido y muchas cosas a mejorar debido a la transformación digital de la sociedad y a la aparición de nueva tecnología disruptiva. Se confirma que la introducción de mecanismos específicos en la concepción y operación de los protocolos de transporte mejora el rendimiento de estos sobre redes elefante heterogéneas.This compendium thesis focuses its contributions on the learning and innovation of the New Generation Networks. That is why different contributions are proposed in different areas (Smart Cities, Smart Grids, Smart Campus, Smart Learning, Media, eHealth, Industry 4.0, among others) through the application and combination of different disciplines (Internet of Things, Building Information Modeling, Cloud Storage, Cybersecurity, Big Data, Future Internet, Digital Transformation).
Specifically, the sustainable comfort monitoring in the Smart Campus is detailed, which can be considered my most representative contribution within the conceptualization of New Generation Networks. Within this innovative monitoring concept, different disciplines are integrated in order to offer information on people's comfort levels. . This research demonstrates the long journey that exists in the digital transformation of traditional sectors and New Generation Networks.
During this long learning about the NGNs through the different investigations, it was possible to observe a problematic that affected the different application fields of the NGNs in a transversal way and that, depending on the service and its requirements, it could have a critical impact on any of these sectors. This issue consists of a low performance operation during the exchange of large volumes of data on networks with high bandwidth capacity and remotely geographically separated, also known as Elephant networks, or Long Fat Networks (LFNs). Specifically, this critically affects the Cloud Data Sharing use case. That is why this use case and the different alternatives at the transport protocol level were studied. For this reason, the performance and operation problems suffered by layer 4 protocols are studied and it is observed why these traditional protocols are not capable of achieving optimal performance.
Due to this situation, it is hypothesized that the introduction of mechanisms that analyze network metrics and efficiently exploit network’s capacity meliorates the performance of Transport Layer protocols over Heterogeneous Long Fat Networks during bulk data transfers.
First, the Adaptive and Aggressive Transport Protocol (AATP) is designed. An adaptive and efficient transport protocol with the aim of maximizing its performance over this type of elephant network.. The AATP protocol is implemented and tested in a network simulator and a testbed under different situations and conditions for its validation.
Once the AATP protocol was designed, implemented and tested successfully, it was decided to improve the protocol itself, Enhanced-AATP, to improve its performance over heterogeneous elephant networks. In addition, in order to upgrade the behavior of the protocol, its fairness system is improved for the fair distribution of resources among other Enhanced-AATP flows. Finally, this evolution is implemented in the network simulator and a set of tests are carried out.
At the end of this thesis, it is concluded that the New Generation Networks have a long way to go and many things to improve due to the digital transformation of society and the appearance of brand-new disruptive technology. Furthermore, it is confirmed that the introduction of specific mechanisms in the conception and operation of transport protocols improves their performance on Heterogeneous Long Fat Networks
Recommended from our members
Performance Evaluation of Classical and Quantum Communication Systems
The Transmission Control Protocol (TCP) is a robust and reliable method used to transport data across a network. Many variants of TCP exist, e.g., Scalable TCP, CUBIC, and H-TCP. While some of them have been studied from empirical and theoretical perspectives, others have been less amenable to a thorough mathematical analysis. Moreover, some of the more popular variants had not been analyzed in the context of the high-speed environments for which they were designed. To address this issue, we develop a generalized modeling technique for TCP congestion control under the assumption of high bandwidth-delay product. In a separate contribution, we develop a versatile fluid model for congestion-window-based and rate-based congestion controllers that can be used to analyze a protocol’s stability. We apply this model to CUBIC – the default implementation of TCP in Linux systems – and discover that under a certain loss probability model, CUBIC is locally asymptotically stable. The contribution of this work is twofold: (i) the first formal stability analysis of CUBIC, and (ii) the fluid model can be easily adapted to other protocols whose window or rate functions are difficult to model. We demonstrate another application of this model by analyzing the stability of H-TCP, another popular variant used in data science networks.
On a different front, a wide range of quantum distributed applications, which either promise to improve on existing classical applications or offer functionality that is entirely unobtainable via classical means, are helping to fuel rapid technological advances in the area of quantum communication. In view of this, it is prudent to model and analyze quantum networks, whose applications range from quantum cryptography to quantum sensing. Several types of quantum distributed applications, such as the E91 protocol for quantum key distribution, make use of entanglement to meet their objectives. Thus, being able to distribute entanglement efficiently is one of the most important and fundamental tasks that must be performed in a quantum network – without this functionality, many quantum distributed applications would be rendered infeasible. Modeling such systems is vital in order to better conceptualize their operation, and more importantly, to discover and address the challenges involved in actualizing them. To this end, we explore the limits of star-topology entanglement switching networks and introduce methods to model the process of entanglement generation, a set of switching policies, memory constraints, link heterogeneity, and quantum state decoherence for a switch that can serve bipartite (and in a specific case, tripartite) entangled states. In one part of this work, we compare two modeling techniques: discrete time Markov chains (DTMCs) and continuous-time Markov chains (CTMCs). We find that while DTMCs are a more accurate way to model the operation of an entanglement distribution switch, they quickly become intractable when one introduces link heterogeneity or state decoherence into the model. In terms of accuracy, we show that not much is lost for the case of homogeneous links, infinite buffer and no decoherence when CTMCs are employed. We then use CTMCs to model more complex systems. In another part of this work, we analyze a switch that can store one or two qubits per link and can serve both bipartite and tripartite entangled states. Through analysis, we discover that randomized policies allow the switch to achieve a better capacity than time-division multiplexing between bipartite and tripartite entangling measurements, but the advantage decreases as the number of links grows
- …