10,074 research outputs found
Joint Power Scaling of Processing Resources and Consolidation of Virtual Network Functions
With the advent of network 'softwarization', Network Functions Virtualization (NFV) is foreseen to provide flexibility and programmability levels that would essentially help in coping with tomorrow's demands. However, energy efficiency and the resulting complexity of network/service management pose serious sustainability and scalability issues that may hinder NFV's advantages. This paper considers these aspects in the context of datacenter networks. We propose an energy-Aware resource allocation scheme to manage virtual machines, dedicated to perform certain (virtualized) network functions, among a pool of energy-Tunable physical resources (processors/cores). We use online measurements to periodically estimate some statistical features of the offered workloads by considering a fairly general renewal model that captures traffic bustiness and hardware operational settings. Then, resources are dynamically managed by jointly performing power scaling and in-server consolidation according to the actual workload variations. The average power consumption generated by this strategy is evaluated and compared with that of a classical bin-packing consolidation, over processors running always with the highest-performance configuration. Results show that the proposed approach can reduce the average power consumption of the datacenter by up to 10%, suggesting a considerable amount of annual savings
An Algorithm for Network and Data-aware Placement of Multi-Tier Applications in Cloud Data Centers
Today's Cloud applications are dominated by composite applications comprising
multiple computing and data components with strong communication correlations
among them. Although Cloud providers are deploying large number of computing
and storage devices to address the ever increasing demand for computing and
storage resources, network resource demands are emerging as one of the key
areas of performance bottleneck. This paper addresses network-aware placement
of virtual components (computing and data) of multi-tier applications in data
centers and formally defines the placement as an optimization problem. The
simultaneous placement of Virtual Machines and data blocks aims at reducing the
network overhead of the data center network infrastructure. A greedy heuristic
is proposed for the on-demand application components placement that localizes
network traffic in the data center interconnect. Such optimization helps
reducing communication overhead in upper layer network switches that will
eventually reduce the overall traffic volume across the data center. This, in
turn, will help reducing packet transmission delay, increasing network
performance, and minimizing the energy consumption of network components.
Experimental results demonstrate performance superiority of the proposed
algorithm over other approaches where it outperforms the state-of-the-art
network-aware application placement algorithm across all performance metrics by
reducing the average network cost up to 67% and network usage at core switches
up to 84%, as well as increasing the average number of application deployments
up to 18%.Comment: Submitted for publication consideration for the Journal of Network
and Computer Applications (JNCA). Total page: 28. Number of figures: 15
figure
Impact of Processing-Resource Sharing on the Placement of Chained Virtual Network Functions
Network Function Virtualization (NFV) provides higher flexibility for network
operators and reduces the complexity in network service deployment. Using NFV,
Virtual Network Functions (VNF) can be located in various network nodes and
chained together in a Service Function Chain (SFC) to provide a specific
service. Consolidating multiple VNFs in a smaller number of locations would
allow decreasing capital expenditures. However, excessive consolidation of VNFs
might cause additional latency penalties due to processing-resource sharing,
and this is undesirable, as SFCs are bounded by service-specific latency
requirements. In this paper, we identify two different types of penalties
(referred as "costs") related to the processingresource sharing among multiple
VNFs: the context switching costs and the upscaling costs. Context switching
costs arise when multiple CPU processes (e.g., supporting different VNFs) share
the same CPU and thus repeated loading/saving of their context is required.
Upscaling costs are incurred by VNFs requiring multi-core implementations,
since they suffer a penalty due to the load-balancing needs among CPU cores.
These costs affect how the chained VNFs are placed in the network to meet the
performance requirement of the SFCs. We evaluate their impact while considering
SFCs with different bandwidth and latency requirements in a scenario of VNF
consolidation.Comment: Accepted for publication in IEEE Transactions on Cloud Computin
Offline and online power aware resource allocation algorithms with migration and delay constraints
© . This manuscript version is made available under the CC-BY-NC-ND 4.0 license http://creativecommons.org/licenses/by-nc-nd/4.0/In order to handle advanced mobile broadband services and Internet of Things (IoT), future Internet and 5G networks are expected to leverage the use of network virtualization, be much faster, have greater capacities, provide lower latencies, and significantly be power efficient than current mobile technologies. Therefore, this paper proposes three power aware algorithms for offline, online, and migration applications, solving the resource allocation problem within the frameworks of network function virtualization (NFV) environments in fractions of a second. The proposed algorithms target minimizing the total costs and power consumptions in the physical network through sufficiently allocating the least physical resources to host the demands of the virtual network services, and put into saving mode all other not utilized physical components. Simulations and evaluations of the offline algorithm compared to the state-of-art resulted on lower total costs by 32%. In addition to that, the online algorithm was tested through four different experiments, and the results argued that the overall power consumption of the physical network was highly dependent on the demands’ lifetimes, and the strictness of the required end-to-end delay. Regarding migrations during online, the results concluded that the proposed algorithms would be most effective when applied for maintenance and emergency conditions.Peer ReviewedPreprin
- …