975 research outputs found
Distributed VNF Scaling in Large-scale Datacenters: An ADMM-based Approach
Network Functions Virtualization (NFV) is a promising network architecture
where network functions are virtualized and decoupled from proprietary
hardware. In modern datacenters, user network traffic requires a set of Virtual
Network Functions (VNFs) as a service chain to process traffic demands. Traffic
fluctuations in Large-scale DataCenters (LDCs) could result in overload and
underload phenomena in service chains. In this paper, we propose a distributed
approach based on Alternating Direction Method of Multipliers (ADMM) to jointly
load balance the traffic and horizontally scale up and down VNFs in LDCs with
minimum deployment and forwarding costs. Initially we formulate the targeted
optimization problem as a Mixed Integer Linear Programming (MILP) model, which
is NP-complete. Secondly, we relax it into two Linear Programming (LP) models
to cope with over and underloaded service chains. In the case of small or
medium size datacenters, LP models could be run in a central fashion with a low
time complexity. However, in LDCs, increasing the number of LP variables
results in additional time consumption in the central algorithm. To mitigate
this, our study proposes a distributed approach based on ADMM. The
effectiveness of the proposed mechanism is validated in different scenarios.Comment: IEEE International Conference on Communication Technology (ICCT),
Chengdu, China, 201
Offline and online power aware resource allocation algorithms with migration and delay constraints
© . This manuscript version is made available under the CC-BY-NC-ND 4.0 license http://creativecommons.org/licenses/by-nc-nd/4.0/In order to handle advanced mobile broadband services and Internet of Things (IoT), future Internet and 5G networks are expected to leverage the use of network virtualization, be much faster, have greater capacities, provide lower latencies, and significantly be power efficient than current mobile technologies. Therefore, this paper proposes three power aware algorithms for offline, online, and migration applications, solving the resource allocation problem within the frameworks of network function virtualization (NFV) environments in fractions of a second. The proposed algorithms target minimizing the total costs and power consumptions in the physical network through sufficiently allocating the least physical resources to host the demands of the virtual network services, and put into saving mode all other not utilized physical components. Simulations and evaluations of the offline algorithm compared to the state-of-art resulted on lower total costs by 32%. In addition to that, the online algorithm was tested through four different experiments, and the results argued that the overall power consumption of the physical network was highly dependent on the demands’ lifetimes, and the strictness of the required end-to-end delay. Regarding migrations during online, the results concluded that the proposed algorithms would be most effective when applied for maintenance and emergency conditions.Peer ReviewedPreprin
Efficient Synthesis of Network Updates
Software-defined networking (SDN) is revolutionizing the networking industry,
but current SDN programming platforms do not provide automated mechanisms for
updating global configurations on the fly. Implementing updates by hand is
challenging for SDN programmers because networks are distributed systems with
hundreds or thousands of interacting nodes. Even if initial and final
configurations are correct, naively updating individual nodes can lead to
incorrect transient behaviors, including loops, black holes, and access control
violations. This paper presents an approach for automatically synthesizing
updates that are guaranteed to preserve specified properties. We formalize
network updates as a distributed programming problem and develop a synthesis
algorithm based on counterexample-guided search and incremental model checking.
We describe a prototype implementation, and present results from experiments on
real-world topologies and properties demonstrating that our tool scales to
updates involving over one-thousand nodes
Content Defined Optical Network
Optical interconnection has become one of the key technologies to adapt the needs of large-scale data center networking with the advantages of large capacity, high bandwidth, and high efficiency. Data center optical interconnection has the characteristics of resource and technology heterogeneity. Its networking and control face enormous challenges for the increasing number of users with a high level quality of service requirements. Around different scenarios, there are a series of key networking and control problems in data center optical interconnection, such as multiple layers and stratums resources optimization in inter-data center, and time-aware resource scheduling in intra-data center. To solve these problems and challenges, this chapter mainly researches on content defined optical networking and integrated control for data center. For networking of vertical “multi-layer-carried” and horizontal “heterogeneous-cross-stratum”, the chapter launches research work around application scenarios about inter-data center optical interconnection with optical network, and intra-data center. The model architecture, implementation mechanism and control strategy are analyzed and demonstrated on the experiment and simulation platform of data center optical interconnection. This chapter will provide important references for future diverse applications of data center optical interconnection and software defined networking and control in practice
ICONA: Inter Cluster ONOS Network Application
Several Network Operating Systems (NOS) have been proposed in the last few
years for Software Defined Networks; however, a few of them are currently
offering the resiliency, scalability and high availability required for
production environments. Open Networking Operating System (ONOS) is an open
source NOS, designed to be reliable and to scale up to thousands of managed
devices. It supports multiple concurrent instances (a cluster of controllers)
with distributed data stores. A tight requirement of ONOS is that all instances
must be close enough to have negligible communication delays, which means they
are typically installed within a single datacenter or a LAN network. However in
certain wide area network scenarios, this constraint may limit the speed of
responsiveness of the controller toward network events like failures or
congested links, an important requirement from the point of view of a Service
Provider. This paper presents ICONA, a tool developed on top of ONOS and
designed in order to extend ONOS capability in network scenarios where there
are stringent requirements in term of control plane responsiveness. In
particular the paper describes the architecture behind ICONA and provides some
initial evaluation obtained on a preliminary version of the tool.Comment: Paper submitted to a conferenc
TimeTrader: Exploiting Latency Tail to Save Datacenter Energy for On-line Data-Intensive Applications
Datacenters running on-line, data-intensive applications (OLDIs) consume
significant amounts of energy. However, reducing their energy is challenging
due to their tight response time requirements. A key aspect of OLDIs is that
each user query goes to all or many of the nodes in the cluster, so that the
overall time budget is dictated by the tail of the replies' latency
distribution; replies see latency variations both in the network and compute.
Previous work proposes to achieve load-proportional energy by slowing down the
computation at lower datacenter loads based directly on response times (i.e.,
at lower loads, the proposal exploits the average slack in the time budget
provisioned for the peak load). In contrast, we propose TimeTrader to reduce
energy by exploiting the latency slack in the sub- critical replies which
arrive before the deadline (e.g., 80% of replies are 3-4x faster than the
tail). This slack is present at all loads and subsumes the previous work's
load-related slack. While the previous work shifts the leaves' response time
distribution to consume the slack at lower loads, TimeTrader reshapes the
distribution at all loads by slowing down individual sub-critical nodes without
increasing missed deadlines. TimeTrader exploits slack in both the network and
compute budgets. Further, TimeTrader leverages Earliest Deadline First
scheduling to largely decouple critical requests from the queuing delays of
sub- critical requests which can then be slowed down without hurting critical
requests. A combination of real-system measurements and at-scale simulations
shows that without adding to missed deadlines, TimeTrader saves 15-19% and
41-49% energy at 90% and 30% loading, respectively, in a datacenter with 512
nodes, whereas previous work saves 0% and 31-37%.Comment: 13 page
- …