1,442 research outputs found

    Congestion mitigation in LTE base stations using radio resource allocation techniques with TCP end to end transport

    Get PDF
    As of 2019, Long Term Evolution (LTE) is the chosen standard for most mobile and fixed wireless data communication. The next generation of standards known as 5G will encompass the Internet of Things (IoT) which will add more wireless devices to the network. Due to an exponential increase in the number of wireless subscriptions, in the next few years there is also an expected exponential increase in data traffic. Most of these devices will use Transmission Control Protocol (TCP) which is a type of network protocol for delivering internet data to users. Due to its reliability in delivering data payload to users and congestion management, TCP is the most common type of network protocol used. However, the ability for TCP to combat network congestion has certain limitations especially in a wireless network. This is due to wireless networks not being as reliable as fixed line networks for data delivery because of the use of last mile radio interface. LTE uses various error correction techniques for reliable data delivery over the air-interface. These cause other issues such as excessive latency and queuing in the base station leading to degradation in throughput for users and congestion in the network. Traditional methods of dealing with congestion such as tail-drop can be inefficient and cumbersome. Therefore, adequate congestion mitigation mechanisms are required. The LTE standard uses a technique to pre-empt network congestion by a mechanism known as Discard Timer. Additionally, there are other algorithms such as Random Early Detection (RED) that also are used for network congestion mitigation. However, these mechanisms rely on configured parameters and only work well within certain regions of operation. If the parameters are not set correctly then the TCP links can experience congestion collapse. In this thesis, the limitations of using existing LTE congestion mitigation mechanisms such as Discard Timer and RED have been explored. A different mechanism to analyse the effects of using control theory for congestion mitigation has been developed. Finally, congestion mitigation in LTE networks has been addresses using radio resource allocation techniques with non-cooperative game theory being an underlying mathematical framework. In doing so, two key end-to-end performance measurements considered for measuring congestion for the game theoretic models were identified which were the total end-to-end delay and the overall throughput of each individual TCP link. An end to end wireless simulator model with the radio access network using LTE and a TCP based backbone to the end server was developed using MATLAB. This simulator was used as a baseline for testing each of the congestion mitigation mechanisms. This thesis also provides a comparison and performance evaluation between the congestion mitigation models developed using existing techniques (such as Discard Timer and RED), control theory and game theory. As of 2019, Long Term Evolution (LTE) is the chosen standard for most mobile and fixed wireless data communication. The next generation of standards known as 5G will encompass the Internet of Things (IoT) which will add more wireless devices to the network. Due to an exponential increase in the number of wireless subscriptions, in the next few years there is also an expected exponential increase in data traffic. Most of these devices will use Transmission Control Protocol (TCP) which is a type of network protocol for delivering internet data to users. Due to its reliability in delivering data payload to users and congestion management, TCP is the most common type of network protocol used. However, the ability for TCP to combat network congestion has certain limitations especially in a wireless network. This is due to wireless networks not being as reliable as fixed line networks for data delivery because of the use of last mile radio interface. LTE uses various error correction techniques for reliable data delivery over the air-interface. These cause other issues such as excessive latency and queuing in the base station leading to degradation in throughput for users and congestion in the network. Traditional methods of dealing with congestion such as tail-drop can be inefficient and cumbersome. Therefore, adequate congestion mitigation mechanisms are required. The LTE standard uses a technique to pre-empt network congestion by a mechanism known as Discard Timer. Additionally, there are other algorithms such as Random Early Detection (RED) that also are used for network congestion mitigation. However, these mechanisms rely on configured parameters and only work well within certain regions of operation. If the parameters are not set correctly then the TCP links can experience congestion collapse. In this thesis, the limitations of using existing LTE congestion mitigation mechanisms such as Discard Timer and RED have been explored. A different mechanism to analyse the effects of using control theory for congestion mitigation has been developed. Finally, congestion mitigation in LTE networks has been addresses using radio resource allocation techniques with non-cooperative game theory being an underlying mathematical framework. In doing so, two key end-to-end performance measurements considered for measuring congestion for the game theoretic models were identified which were the total end-to-end delay and the overall throughput of each individual TCP link. An end to end wireless simulator model with the radio access network using LTE and a TCP based backbone to the end server was developed using MATLAB. This simulator was used as a baseline for testing each of the congestion mitigation mechanisms. This thesis also provides a comparison and performance evaluation between the congestion mitigation models developed using existing techniques (such as Discard Timer and RED), control theory and game theory

    JTP, an energy-aware transport protocol for mobile ad hoc networks (PhD thesis)

    Full text link
    Wireless ad-hoc networks are based on a cooperative communication model, where all nodes not only generate traffic but also help to route traffic from other nodes to its final destination. In such an environment where there is no infrastructure support the lifetime of the network is tightly coupled with the lifetime of individual nodes. Most of the devices that form such networks are battery-operated, and thus it becomes important to conserve energy so as to maximize the lifetime of a node. In this thesis, we present JTP, a new energy-aware transport protocol, whose goal is to reduce power consumption without compromising delivery requirements of applications. JTP has been implemented within the JAVeLEN system. JAVeLEN [RKM+08], is a new system architecture for ad hoc networks that has been developed to elevate energy efficiency as a first-class optimization metric at all protocol layers, from physical to transport. Thus, energy gains obtained in one layer would not be offset by incompatibilities and/or inefficiencies in other layers. To meet its goal of energy efficiency, JTP (1) contains mechanisms to balance end-toend vs. local retransmissions; (2) minimizes acknowledgment traffic using receiver regulated rate-based flow control combined with selected acknowledgments and in-network caching of packets; and (3) aggressively seeks to avoid any congestion-based packet loss. Within this ultra low-power multi-hop wireless network system, simulations and experimental results demonstrate that our transport protocol meets its goal of preserving the energy efficiency of the underlying network. JTP has been implemented on the actual JAVeLEN nodes and its benefits have been demonstrated on a real system

    JTP, an energy-aware transport protocol for mobile ad hoc networks

    Full text link
    Wireless ad-hoc networks are based on a cooperative communication model, where all nodes not only generate traffic but also help to route traffic from other nodes to its final destination. In such an environment where there is no infrastructure support the lifetime of the network is tightly coupled with the lifetime of individual nodes. Most of the devices that form such networks are battery-operated, and thus it becomes important to conserve energy so as to maximize the lifetime of a node. In this thesis, we present JTP, a new energy-aware transport protocol, whose goal is to reduce power consumption without compromising delivery requirements of applications. JTP has been implemented within the JAVeLEN system. JAVeLEN~\cite{javelen08redi}, is a new system architecture for ad hoc networks that has been developed to elevate energy efficiency as a first-class optimization metric at all protocol layers, from physical to transport. Thus, energy gains obtained in one layer would not be offset by incompatibilities and/or inefficiencies in other layers. To meet its goal of energy efficiency, JTP (1) contains mechanisms to balance end-to-end vs. local retransmissions; (2) minimizes acknowledgment traffic using receiver regulated rate-based flow control combined with selected acknowledgments and in-network caching of packets; and (3) aggressively seeks to avoid any congestion-based packet loss. Within this ultra low-power multi-hop wireless network system, simulations and experimental results demonstrate that our transport protocol meets its goal of preserving the energy efficiency of the underlying network. JTP has been implemented on the actual JAVeLEN nodes and its benefits have been demoed on a real system

    Adaptive protocols for mobile ad hoc networks

    Get PDF
    Recent advances in low-power technologies have resulted in the proliferation of inexpensive handheld mobile computing devices. Soon, just like the Internet empow- ered a whole new world of applications for personal computers, the development and deployment of robust ubiquitous wireless networks will enable many new and exciting futuristic applications. Certain to be an important part of this future is a class of networks known as "mobile ad hoc networks." Mobile ad hoc networks (or simply "ad hoc networks") are local-area networks formed "on the spot" between collocated wireless devices. These devices self-organize by sharing information with their neigh- bors to establish communication pathways whenever and wherever they are. For ad hoc networks to succeed, however, new protocols must be developed that are capable of adapting to their dynamic nature. In this dissertation, we present a number of adaptive protocols that are designed for this purpose. We investigate new link layer mechanisms that dynamically monitor and adapt to changes in link quality, including a protocol that uses common control messages to form a tight feedback control loop for adaptation of the link data rate to best match the channel conditions perceived by the receiver. We also investigate routing protocols that adapt route selection according to network characteristics. In particular, we present two on-demand routing protocols that are designed to take advantage of the presence of multirate links. We then investigate the performance of TCP, showing how communication outages caused by link failures and routing delays can be very detrimental to its performance. In response, we present a solution to this problem that uses explicit feedback messages from the link layer about link failures to adapt TCP's behavior. Finally, we show how link failures in heterogeneous networks containing links with widely varying bandwidth and delay can cause repeated "modal" changes in capacity that TCP is slow to detect. We then present a modifed version of TCP that is capable of more rapidly detecting and adapting to these changes

    An IPsec Compatible Implementation of DBRA and IP-ABR

    Get PDF
    Satellites are some of the most difficult links to exploit in a Quality of Service (QoS) sensitive network, largely due to their high latency, variable-bandwidth and low-bandwidth nature. Central management of shared links has been shown to provide efficiency gains and enhanced QoS by effectively allocating resources according to reservations and dynamic resource availability. In a modern network, segregated by secure gateways and tunnels such as provided by IPsec, central management appears impossible to implement due to the barriers created between a global Dynamic Bandwidth Resource Allocation (DBRA) system and the mediators controlling the individual flows. This thesis explores and evaluates various through-IPsec communications techniques aimed at providing a satellite-to-network control channel, while maintaining data security for all communications involved

    Techniques for End-to-End Tcp Performance Enhancement Over Wireless Networks

    Get PDF
    Today’s wireless network complexity and the new applications from various user devices call for an in-depth understanding of the mutual performance impact of networks and applications. It includes understanding of the application traffic and network layer protocols to enable end-to-end application performance enhancements over wireless networks. Although Transport Control Protocol (TCP) behavior over wireless networks is well known, it remains as one of the main drivers which may significantly impact the user experience through application performance as well as the network resource utilization, since more than 90% of the internet traffic uses TCP in both wireless and wire-line networks. In this dissertation, we employ application traffic measurement and packet analysis over a commercial Long Term Evolution (LTE) network combined with an in-depth LTE protocol simulation to identify three critical problems that may negatively affect the application performance and wireless network resource utilization: (i) impact of the wireless MAC protocol on the TCP throughput performance, (ii) impact of applications on network resource utilization, and (iii) impact of TCP on throughput performance over wireless networks. We further propose four novel mechanisms to improve the end-to-end application and wireless system performance: (i) an enhanced LTE uplink resource allocation mechanism to reduce network delay and help prevent a TCP timeout, (ii) a new TCP snooping mechanism, which according to our experiments, can save about 20% of system resources by preventing unnecessary video packet transmission through the air interface, and (iii) two Split-TCP protocols: an Enhanced Split-TCP (ES-TCP) and an Advanced Split-TCP (AS-TCP), which significantly improve the application throughput without breaking the end-to-end TCP semantics. Experimental results show that the proposed ES-TCP and AS-TCP protocols can boost the TCP throughput by more than 60% in average, when exercised over a 4G LTE network. Furthermore, the TCP throughput performance improvement may be even superior to 200%, depending on network and usage conditions. We expect that these proposed Split-TCP protocol enhancements, together with the new uplink resource allocation enhancement and the new TCP snooping mechanism may provide even greater performance gains when more advanced radio technologies, such as 5G, are deployed. Thanks to their superior resource utilization efficiency, such advanced radio technologies will put to greater use the techniques and protocol enhancements disclosed through this dissertation

    Moving toward the intra-protocol de-ossification of TCP in mobile networks: Start-up and mobility

    Get PDF
    182 p.El uso de las redes móviles de banda ancha ha aumentado significativamente los últimos años y se espera un crecimiento aún mayor con la inclusión de las futuras capacidades 5G. 5G proporcionará unas velocidades de transmisión y reducidos retardos nunca antes vistos. Sin embargo, la posibilidad de alcanzar las mencionadas cuotas está limitada por la gestión y rendimiento de los protocolos de transporte. A este respecto, TCP sigue siendo el protocolo de transporte imperante y sus diferentes algoritmos de control de congestión (CCA) los responsables finales del rendimiento obtenido. Mientras que originalmente los distintos CCAs han sido implementados para hacer frente a diferentes casos de uso en redes fijas, ninguno de los CCAs ha sido diseñado para poder gestionar la variabilidad de throughput y retardos de diferentes condiciones de red redes móviles de una manera fácilmente implantable. Dado que el análisis de TCP sobre redes móviles es complejo debido a los múltiples factores de impacto, nuestro trabajo se centra en dos casos de uso generalizados que resultan significativos en cuanto a afección del rendimiento: movimiento de los usuarios como representación de la característica principal de las redes móviles frente a las redes fijas y el rendimiento de la fase de Start-up de TCP debido a la presencia mayoritaria de flujos cortos en Internet. Diferentes trabajos han sugerido la importancia de una mayor flexibilidad en la capa de transporte, creando servicios de transporte sobre TCP o UDP. Sin embargo, estas propuestas han encontrado limitaciones relativas a las dependencias arquitecturales de los protocolos utilizados como sustrato (p.ej. imposibilidad de cambiar la configuración de la capa de transporte una vez la transmisión a comenzado), experimentando una capa de transporte "osificada". Esta tesis surge como respuesta a fin de abordar la citada limitación y demostrando que existen posibilidades de mejora dentro de la familia de TCP (intra-protocolar), proponiendo un marco para solventar parcialmente la restricción a través de la selección dinámica del CCA más apropiado. Para ello, se evalúan y seleccionan los mayores puntos de impacto en el rendimiento de los casos de uso seleccionados en despliegues de red 4G y en despliegues de baja latencia que emulan las potenciales latencias en las futuras capacidades 5G. Estos puntos de impacto sirven como heurísticas para decidir el CCA más apropiado en el propuesto marco. Por último, se valida la propuesta en entornos de movilidad con dos posibilidades de selección: al comienzo de la transmisión (limitada flexibilidad de la capa de transporte) y dinámicamente durante la transmisión (con una capa de transporte flexible). Se concluye que la propuesta puede acarrear importantes mejoras de rendimiento al seleccionar el CCA más apropiado teniendo en cuenta la situación de red y los requerimientos de la capa de aplicación
    corecore