1,532 research outputs found

    Balancing the Migration of Virtual Network Functions with Replications in Data Centers

    Full text link
    The Network Function Virtualization (NFV) paradigm is enabling flexibility, programmability and implementation of traditional network functions into generic hardware, in form of the so-called Virtual Network Functions (VNFs). Today, cloud service providers use Virtual Machines (VMs) for the instantiation of VNFs in the data center (DC) networks. To instantiate multiple VNFs in a typical scenario of Service Function Chains (SFCs), many important objectives need to be met simultaneously, such as server load balancing, energy efficiency and service execution time. The well-known \emph{VNF placement} problem requires solutions that often consider \emph{migration} of virtual machines (VMs) to meet this objectives. Ongoing efforts, for instance, are making a strong case for migrations to minimize energy consumption, while showing that attention needs to be paid to the Quality of Service (QoS) due to service interruptions caused by migrations. To balance the server allocation strategies and QoS, we propose using \emph{replications} of VNFs to reduce migrations in DC networks. We propose a Linear Programming (LP) model to study a trade-off between replications, which while beneficial to QoS require additional server resources, and migrations, which while beneficial to server load management can adversely impact the QoS. The results show that, for a given objective, the replications can reduce the number of migrations and can also enable a better server and data center network load balancing

    Providing Transaction Class-Based QoS in In-Memory Data Grids via Machine Learning

    Get PDF
    Elastic architectures and the ”pay-as-you-go” resource pricing model offered by many cloud infrastructure providers may seem the right choice for companies dealing with data centric applications characterized by high variable workload. In such a context, in-memory transactional data grids have demonstrated to be particularly suited for exploiting advantages provided by elastic computing platforms, mainly thanks to their ability to be dynamically (re-)sized and tuned. Anyway, when specific QoS requirements have to be met, this kind of architectures have revealed to be complex to be managed by humans. Particularly, their management is a very complex task without the stand of mechanisms supporting run-time automatic sizing/tuning of the data platform and the underlying (virtual) hardware resources provided by the cloud. In this paper, we present a neural network-based architecture where the system is constantly and automatically re-configured, particularly in terms of computing resources

    Performance-oriented Cloud Provisioning: Taxonomy and Survey

    Full text link
    Cloud computing is being viewed as the technology of today and the future. Through this paradigm, the customers gain access to shared computing resources located in remote data centers that are hosted by cloud providers (CP). This technology allows for provisioning of various resources such as virtual machines (VM), physical machines, processors, memory, network, storage and software as per the needs of customers. Application providers (AP), who are customers of the CP, deploy applications on the cloud infrastructure and then these applications are used by the end-users. To meet the fluctuating application workload demands, dynamic provisioning is essential and this article provides a detailed literature survey of dynamic provisioning within cloud systems with focus on application performance. The well-known types of provisioning and the associated problems are clearly and pictorially explained and the provisioning terminology is clarified. A very detailed and general cloud provisioning classification is presented, which views provisioning from different perspectives, aiding in understanding the process inside-out. Cloud dynamic provisioning is explained by considering resources, stakeholders, techniques, technologies, algorithms, problems, goals and more.Comment: 14 pages, 3 figures, 3 table
    corecore