14 research outputs found
Integração do paradigma de cloud computing com a infraestrutura de rede do operador
Doutoramento em Engenharia InformáticaThe proliferation of Internet access allows that users have the possibility to use
services available directly through the Internet, which translates in a change of
the paradigm of using applications and in the way of communicating,
popularizing in this way the so-called cloud computing paradigm. Cloud
computing brings with it requirements at two different levels: at the cloud level,
usually relying in centralized data centers, where information technology and
network resources must be able to guarantee the demand of such services;
and at the access level, i.e., depending on the service being consumed,
different quality of service is required in the access network, which is a Network
Operator (NO) domain. In summary, there is an obvious network dependency.
However, the network has been playing a relatively minor role, mostly as a
provider of (best-effort) connectivity within the cloud and in the access network.
The work developed in this Thesis enables for the effective integration of cloud
and NO domains, allowing the required network support for cloud. We propose
a framework and a set of associated mechanisms for the integrated
management and control of cloud computing and NO domains to provide endto-
end services. Moreover, we elaborate a thorough study on the embedding of
virtual resources in this integrated environment. The study focuses on
maximizing the host of virtual resources on the physical infrastructure through
optimal embedding strategies (considering the initial allocation of resources as
well as adaptations through time), while at the same time minimizing the costs
associated to energy consumption, in single and multiple domains.
Furthermore, we explore how the NO can take advantage of the integrated
environment to host traditional network functions. In this sense, we study how
virtual network Service Functions (SFs) should be modelled and managed in a
cloud environment and enhance the framework accordingly.
A thorough evaluation of the proposed solutions was performed in the scope of
this Thesis, assessing their benefits. We implemented proof of concepts to
prove the added value, feasibility and easy deployment characteristics of the
proposed framework. Furthermore, the embedding strategies evaluation has
been performed through simulation and Integer Linear Programming (ILP)
solving tools, and it showed that it is possible to reduce the physical
infrastructure energy consumption without jeopardizing the virtual resources
acceptance. This fact can be further increased by allowing virtual resource
adaptation through time. However, one should have in mind the costs
associated to adaptation processes. The costs can be minimized, but the virtual
resource acceptance can be also reduced. This tradeoff has also been subject
of the work in this Thesis.A proliferação do acesso à Internet permite aos utilizadores usar serviços
disponibilizados diretamente através da Internet, o que se traduz numa
mudança de paradigma na forma de usar aplicações e na forma de comunicar,
popularizando desta forma o conceito denominado de cloud computing. Cloud
computing traz consigo requisitos a dois níveis: ao nível da própria cloud,
geralmente dependente de centros de dados centralizados, onde as
tecnologias de informação e recursos de rede têm que ser capazes de garantir
as exigências destes serviços; e ao nível do acesso, ou seja, dependendo do
serviço que esteja a ser consumido, são necessários diferentes níveis de
qualidade de serviço na rede de acesso, um domínio do operador de rede. Em
síntese, existe uma clara dependência da cloud na rede. No entanto, o papel
que a rede tem vindo a desempenhar neste âmbito é reduzido, sendo
principalmente um fornecedor de conectividade (best-effort) tanto no dominio
da cloud como no da rede de acesso.
O trabalho desenvolvido nesta Tese permite uma integração efetiva dos
domínios de cloud e operador de rede, dando assim à cloud o efetivo suporte
da rede. Para tal, apresentamos uma plataforma e um conjunto de
mecanismos associados para gestão e controlo integrado de domínios cloud
computing e operador de rede por forma a fornecer serviços fim-a-fim. Além
disso, elaboramos um estudo aprofundado sobre o mapeamento de recursos
virtuais neste ambiente integrado. O estudo centra-se na maximização da
incorporação de recursos virtuais na infraestrutura física por meio de
estratégias de mapeamento ótimas (considerando a alocação inicial de
recursos, bem como adaptações ao longo do tempo), enquanto que se
minimizam os custos associados ao consumo de energia. Este estudo é feito
para cenários de apenas um domínio e para cenários com múltiplos domínios.
Além disso, exploramos como o operador de rede pode aproveitar o referido
ambiente integrado para suportar funções de rede tradicionais. Neste sentido,
estudamos como as funções de rede virtualizadas devem ser modeladas e
geridas num ambiente cloud e estendemos a plataforma de acordo com este
conceito.
No âmbito desta Tese foi feita uma avaliação extensa das soluções propostas,
avaliando os seus benefícios. Implementámos provas de conceito por forma a
demonstrar as mais-valias, viabilidade e fácil implantação das soluções
propostas. Além disso, a avaliação das estratégias de mapeamento foi
realizada através de ferramentas de simulação e de programação linear inteira,
mostrando que é possível reduzir o consumo de energia da infraestrutura
física, sem comprometer a aceitação de recursos virtuais. Este aspeto pode
ser melhorado através da adaptação de recursos virtuais ao longo do tempo.
No entanto, deve-se ter em mente os custos associados aos processos de
adaptação. Os custos podem ser minimizados, mas isso implica uma redução
na aceitação de recursos virtuais. Esta compensação foi também um tema
abordado nesta Tese
Hybrid SDN Evolution: A Comprehensive Survey of the State-of-the-Art
Software-Defined Networking (SDN) is an evolutionary networking paradigm
which has been adopted by large network and cloud providers, among which are
Tech Giants. However, embracing a new and futuristic paradigm as an alternative
to well-established and mature legacy networking paradigm requires a lot of
time along with considerable financial resources and technical expertise.
Consequently, many enterprises can not afford it. A compromise solution then is
a hybrid networking environment (a.k.a. Hybrid SDN (hSDN)) in which SDN
functionalities are leveraged while existing traditional network
infrastructures are acknowledged. Recently, hSDN has been seen as a viable
networking solution for a diverse range of businesses and organizations.
Accordingly, the body of literature on hSDN research has improved remarkably.
On this account, we present this paper as a comprehensive state-of-the-art
survey which expands upon hSDN from many different perspectives
Routing on the Channel Dependency Graph:: A New Approach to Deadlock-Free, Destination-Based, High-Performance Routing for Lossless Interconnection Networks
In the pursuit for ever-increasing compute power, and with Moore's law slowly coming to an end, high-performance computing started to scale-out to larger systems. Alongside the increasing system size, the interconnection network is growing to accommodate and connect tens of thousands of compute nodes. These networks have a large influence on total cost, application performance, energy consumption, and overall system efficiency of the supercomputer. Unfortunately, state-of-the-art routing algorithms, which define the packet paths through the network, do not utilize this important resource efficiently. Topology-aware routing algorithms become increasingly inapplicable, due to irregular topologies, which either are irregular by design, or most often a result of hardware failures. Exchanging faulty network components potentially requires whole system downtime further increasing the cost of the failure. This management approach becomes more and more impractical due to the scale of today's networks and the accompanying steady decrease of the mean time between failures. Alternative methods of operating and maintaining these high-performance interconnects, both in terms of hardware- and software-management, are necessary to mitigate negative effects experienced by scientific applications executed on the supercomputer. However, existing topology-agnostic routing algorithms either suffer from poor load balancing or are not bounded in the number of virtual channels needed to resolve deadlocks in the routing tables.
Using the fail-in-place strategy, a well-established method for storage systems to repair only critical component failures, is a feasible solution for current and future HPC interconnects as well as other large-scale installations such as data center networks. Although, an appropriate combination of topology and routing algorithm is required to minimize the throughput degradation for the entire system. This thesis contributes a network simulation toolchain to facilitate the process of finding a suitable combination, either during system design or while it is in operation. On top of this foundation, a key contribution is a novel scheduling-aware routing, which reduces fault-induced throughput degradation while improving overall network utilization. The scheduling-aware routing performs frequent property preserving routing updates to optimize the path balancing for simultaneously running batch jobs. The increased deployment of lossless interconnection networks, in conjunction with fail-in-place modes of operation and topology-agnostic, scheduling-aware routing algorithms, necessitates new solutions to solve the routing-deadlock problem. Therefore, this thesis further advances the state-of-the-art by introducing a novel concept of routing on the channel dependency graph, which allows the design of an universally applicable destination-based routing capable of optimizing the path balancing without exceeding a given number of virtual channels, which are a common hardware limitation. This disruptive innovation enables implicit deadlock-avoidance during path calculation, instead of solving both problems separately as all previous solutions
Recommended from our members
Measurement-Driven Algorithm and System Design for Wireless and Datacenter Networks
The growing number of mobile devices and data-intensive applications pose unique challenges for wireless access networks as well as datacenter networks that enable modern cloud-based services. With the enormous increase in volume and complexity of traffic from applications such as video streaming and cloud computing, the interconnection networks have become a major performance bottleneck. In this thesis, we study algorithms and architectures spanning several layers of the networking protocol stack that enable and accelerate novel applications and that are easily deployable and scalable. The design of these algorithms and architectures is motivated by measurements and observations in real world or experimental testbeds.
In the first part of this thesis, we address the challenge of wireless content delivery in crowded areas. We present the AMuSe system, whose objective is to enable scalable and adaptive WiFi multicast. AMuSe is based on accurate receiver feedback and incurs a small control overhead. This feedback information can be used by the multicast sender to optimize multicast service quality, e.g., by dynamically adjusting transmission bitrate. Specifically, we develop an algorithm for dynamic selection of a subset of the multicast receivers as feedback nodes which periodically send information about the channel quality to the multicast sender. Further, we describe the Multicast Dynamic Rate Adaptation (MuDRA) algorithm that utilizes AMuSe's feedback to optimally tune the physical layer multicast rate. MuDRA balances fast adaptation to channel conditions and stability, which is essential for multimedia applications.
We implemented the AMuSe system on the ORBIT testbed and evaluated its performance in large groups with approximately 200 WiFi nodes. Our extensive experiments demonstrate that AMuSe can provide accurate feedback in a dense multicast environment. It outperforms several alternatives even in the case of external interference and changing network conditions. Further, our experimental evaluation of MuDRA on the ORBIT testbed shows that MuDRA outperforms other schemes and supports high throughput multicast flows to hundreds of nodes while meeting quality requirements. As an example application, MuDRA can support multiple high quality video streams, where 90% of the nodes report excellent or very good video quality.
Next, we specifically focus on ensuring high Quality of Experience (QoE) for video streaming over WiFi multicast. We formulate the problem of joint adaptation of multicast transmission rate and video rate for ensuring high video QoE as a utility maximization problem and propose an online control algorithm called DYVR which is based on Lyapunov optimization techniques. We evaluated the performance of DYVR through analysis, simulations, and experiments using a testbed composed of Android devices and o the shelf APs. Our evaluation shows that DYVR can ensure high video rates while guaranteeing a low but acceptable number of segment losses, buffer underflows, and video rate switches.
We leverage the lessons learnt from AMuSe for WiFi to address the performance issues with LTE evolved Multimedia Broadcast/Multicast Service (eMBMS). We present the Dynamic Monitoring (DyMo) system which provides low-overhead and real-time feedback about eMBMS performance. DyMo employs eMBMS for broadcasting instructions which indicate the reporting rates as a function of the observed Quality of Service (QoS) for each UE. This simple feedback mechanism collects very limited QoS reports which can be used for network optimization. We evaluated the performance of DyMo analytically and via simulations. DyMo infers the optimal eMBMS settings with extremely low overhead, while meeting strict QoS requirements under different UE mobility patterns and presence of network component failures.
In the second part of the thesis, we study datacenter networks which are key enablers of the end-user applications such as video streaming and storage. Datacenter applications such as distributed file systems, one-to-many virtual machine migrations, and large-scale data processing involve bulk multicast flows. We propose a hardware and software system for enabling physical layer optical multicast in datacenter networks using passive optical splitters. We built a prototype and developed a simulation environment to evaluate the performance of the system for bulk multicasting. Our evaluation shows that the optical multicast architecture can achieve higher throughput and lower latency than IP multicast and peer-to-peer multicast schemes with lower switching energy consumption.
Finally, we study the problem of congestion control in datacenter networks. Quantized Congestion Control (QCN), a switch-supported standard, utilizes direct multi-bit feedback from the network for hardware rate limiting. Although QCN has been shown to be fast-reacting and effective, being a Layer-2 technology limits its adoption in IP-routed Layer 3 datacenters. We address several design challenges to overcome QCN feedback's Layer- 2 limitation and use it to design window-based congestion control (QCN-CC) and load balancing (QCN-LB) schemes. Our extensive simulations, based on real world workloads, demonstrate the advantages of explicit, multi-bit congestion feedback, especially in a typical environment where intra-datacenter traffic with short Round Trip Times (RTT: tens of s) run in conjunction with web-facing traffic with long RTTs (tens of milliseconds)
Next generation control of transport networks
It is widely understood by telecom operators and industry analysts that bandwidth demand is increasing dramatically, year on year, with typical growth figures of 50% for Internet-based traffic [5]. This trend means that the consumers will have both a wide variety of devices attaching to their networks and a range of high bandwidth service requirements. The corresponding impact is the effect on the traffic engineered network (often referred to as the “transport network”) to ensure that the current rate of growth of network traffic is supported and meets predicted future demands. As traffic demands increase and newer services continuously arise, novel network elements are needed to provide more flexibility, scalability, resilience, and adaptability to today’s transport network. The transport network provides transparent traffic engineered communication of user, application, and device traffic between attached clients (software and hardware) and establishing and maintaining point-to-point or point-to-multipoint connections. The research documented in this thesis was based on three initial research questions posed while performing research at British Telecom research labs and investigating control of transport networks of future transport networks: 1. How can we meet Internet bandwidth growth yet minimise network costs? 2. Which enabling network technologies might be leveraged to control network layers and functions cooperatively, instead of separated network layer and technology control? 3. Is it possible to utilise both centralised and distributed control mechanisms for automation and traffic optimisation? This thesis aims to provide the classification, motivation, invention, and evolution of a next generation control framework for transport networks, and special consideration of delivering broadcast video traffic to UK subscribers. The document outlines pertinent telecoms technology and current art, how requirements I gathered, and research I conducted, and by which the transport control framework functional components are identified and selected, and by which method the architecture was implemented and applied to key research projects requiring next generation control capabilities, both at British Telecom and the wider research community. Finally, in the closing chapters, the thesis outlines the next steps for ongoing research and development of the transport network framework and key areas for further study
Evaluation of data centre networks and future directions
Traffic forecasts predict a more than threefold increase in the global datacentre workload in coming years, caused by the increasing adoption of cloud and data-intensive applications. Consequently, there has been an unprecedented need for ultra-high throughput and minimal latency. Currently deployed hierarchical architectures using electronic packet switching technologies are costly and energy-inefficient. Very high capacity switches are required to satisfy the enormous bandwidth requirements of cloud datacentres and this limits the overall network scalability. With the maturity of photonic components, turning to optical switching in data centres is a viable option to accommodate greater bandwidth and network flexibility while potentially minimising the latency, cost and power consumption.
Various DCN architectures have been proposed to date and this thesis includes a comparative analysis of such electronic and optical topologies to judge their suitability based on network performance parameters and cost/energy effectiveness, while identifying the challenges faced by recent DCN infrastructures. An analytical Layer 2 switching model is introduced that can alleviate the simulation scalability problem and evaluate the performance of the underlying DCN architecture. This model is also used to judge the variation in traffic arrival/offloading at the intermediate queueing stages and the findings are used to derive closed form expressions for traffic arrival rates and delay. The results from the simulated network demonstrate the impact of buffering and versubscription and reveal the potential bottlenecks and network design tradeoffs. TCP traffic forms the bulk of current DCN workload and so the designed network is further modified to include TCP flows generated from a realistic traffic generator for assessing the impact of Layer 4 congestion control on the DCN performance with standard TCP and datacentre specific TCP protocols (DCTCP). Optical DCN architectures mostly concentrate on core-tier switching. However, substantial energy saving is possible by introducing optics in the edge tiers. Hence, a new approach to optical switching is introduced using Optical ToR switches which can offer better delay performance than commodity switches of similiar size, while having far less power dissipation. An all-optical topology has been further outlined for the efficient implementation of the optical switch meeting the future scalability demands
D4.1 Draft air interface harmonization and user plane design
The METIS-II project envisions the design of a new air interface in order to fulfil all the performance requirements of the envisioned 5G use cases including some extreme low latency use cases and ultra-reliable transmission, xMBB requiring additional capacity that is only available in very high frequencies, as well as mMTC with extremely densely distributed sensors and very long battery life requirements. Designing an adaptable and flexible 5G Air Interface (AI), which will tackle these use cases while offering native multi-service support, is one of the key tasks of METIS-II WP4. This deliverable will highlight the challenges of designing an AI required to operate in a wide range of spectrum bands and cell sizes, capable of addressing the diverse services with often diverging requirements, and propose a design and suitability assessment framework for 5G AI candidates.Aydin, O.; Gebert, J.; Belschner, J.; Bazzi, J.; Weitkemper, P.; Kilinc, C.; Leonardo Da Silva, I.... (2016). D4.1 Draft air interface harmonization and user plane design. https://doi.org/10.13140/RG.2.2.24542.0288
Combining SOA and BPM Technologies for Cross-System Process Automation
This paper summarizes the results of an industry case study that introduced a cross-system business process automation solution based on a combination of SOA and BPM standard technologies (i.e., BPMN, BPEL, WSDL). Besides discussing major weaknesses of the existing, custom-built, solution and comparing them against experiences with the developed prototype, the paper presents a course of action for transforming the current solution into the proposed solution. This includes a general approach, consisting of four distinct steps, as well as specific action items that are to be performed for every step. The discussion also covers language and tool support and challenges arising from the transformation