84,135 research outputs found

    Workload Prediction for Efficient Performance Isolation and System Reliability

    Get PDF
    In large-scaled and distributed systems, like multi-tier storage systems and cloud data centers, resource sharing among workloads brings multiple benefits while introducing many performance challenges. The key to effective workload multiplexing is accurate workload prediction. This thesis focuses on how to capture the salient characteristics of the real-world workloads to develop workload prediction methods and to drive scheduling and resource allocation policies, in order to achieve efficient and in-time resource isolation among applications. For a multi-tier storage system, high-priority user work is often multiplexed with low-priority background work. This brings the challenge of how to strike a balance between maintaining the user performance and maximizing the amount of finished background work. In this thesis, we propose two resource isolation policies based on different workload prediction methods: one is a Markovian model-based and the other is a neural networks-based. These policies aim at, via workload prediction, discovering the opportune time to schedule background work with minimum impact on user performance. Trace-driven simulations verify the efficiency of the two pro- posed resource isolation policies. The Markovian model-based policy successfully schedules the background work at the appropriate periods with small impact on the user performance. The neural networks-based policy adaptively schedules user and background work, resulting in meeting both performance requirements consistently. This thesis also proposes an accurate while efficient neural networks-based pre- diction method for data center usage series, called PRACTISE. Different from the traditional neural networks for time series prediction, PRACTISE selects the most informative features from the past observations of the time series itself. Testing on a large set of usage series in production data centers illustrates the accuracy (e.g., prediction error) and efficiency (e.g., time cost) of PRACTISE. The superiority of the usage prediction also allows a proactive resource management in the highly virtualized cloud data centers. In this thesis, we analyze on the performance tickets in the cloud data centers, and propose an active sizing algorithm, named ATM, that predicts the usage workloads and re-allocates capacity to work- loads to avoid VM performance tickets. Moreover, driven by cheap prediction of usage tails, we also present TailGuard in this thesis, which dynamically clones VMs among co-located boxes, in order to efficiently reduce the performance violations of physical boxes in cloud data centers

    Analysis of Software Aging in a Web Server

    Get PDF
    A number of recent studies have reported the phenomenon of “software aging”, characterized by progressive performance degradation and/or an increased occurrence rate of hang/crash failures of a software system due to the exhaustion of operating system resources or the accumulation of errors. To counteract this phenomenon, a proactive technique called 'software rejuvenation' has been proposed. It essentially involves stopping the running software, cleaning its internal state and/or its environment and then restarting it. Software rejuvenation, being preventive in nature, begs the question as to when to schedule it. Periodic rejuvenation, while straightforward to implement, may not yield the best results, because the rate at which software ages is not constant, but it depends on the time-varying system workload. Software rejuvenation should therefore be planned and initiated in the face of the actual system behavior. This requires the measurement, analysis and prediction of system resource usage. In this paper, we study the development of resource usage in a web server while subjecting it to an artificial workload. We first collect data on several system resource usage and activity parameters. Non-parametric statistical methods are then applied for detecting and estimating trends in the data sets. Finally, we fit time series models to the data collected. Unlike the models used previously in the research on software aging, these time series models allow for seasonal patterns, and we show how the exploitation of the seasonal variation can help in adequately predicting the future resource usage. Based on the models employed here, proactive management techniques like software rejuvenation triggered by actual measurements can be built. --Software aging,software rejuvenation,Linux,Apache,web server,performance monitoring,prediction of resource utilization,non-parametric trend analysis,time series analysis

    A study on performance measures for auto-scaling CPU-intensive containerized applications

    Get PDF
    Autoscaling of containers can leverage performance measures from the different layers of the computational stack. This paper investigate the problem of selecting the most appropriate performance measure to activate auto-scaling actions aiming at guaranteeing QoS constraints. First, the correlation between absolute and relative usage measures and how a resource allocation decision can be influenced by them is analyzed in different workload scenarios. Absolute and relative measures could assume quite different values. The former account for the actual utilization of resources in the host system, while the latter account for the share that each container has of the resources used. Then, the performance of a variant of Kubernetes’ auto-scaling algorithm, that transparently uses the absolute usage measures to scale-in/out containers, is evaluated through a wide set of experiments. Finally, a detailed analysis of the state-of-the-art is presented

    Workload utilization dissemination on grid resources for simulation environment

    Get PDF
    This paper discusses the workload utilization dissemination for grid computing. The CPU is a well-known resource item and it is an integral part in most literatures while other RI's may include memory, network and I/O overhead. The selection of resource variables and the number of RI's involved will result in different definitions of the workload. Various combination of computer RI's have been explored for studying the style of usage, techniques embedded and their capabilities. In contemplating the exploration, this study successfully describe the pattern of workload dissemination through the usage of the RI's and elicited the enhancement factors for systems performance. Among these factors are the manipulation of computer RI's, type of workload information with method of use, the workload dissemination direction along with implementation method and using certain algorithm to come out with new integrated scheduling with load balancing capability. A combination of these factors will help in developing an optimized scheduling or load balancing algorithm

    A rough-cut capacity planning model with overlapping

    Get PDF
    In the early phases of projects, capacity planning is performed to assess the feasibility of the project in terms of delivery date, resource usage and cost. This tactical approach relies on an aggregated representation of tasks in work packages. At this level, aggressive project duration objectives are achieved by adopting work package overlapping policies that affect both workload and resource usage. In this article, we propose a mixed-time MILP model for project capacity planning with different possibilities for overlapping levels between work packages. In the model, the planning time horizon is divided into time buckets used to evaluate resource usage, while starting and ending times for work packages are continuous. The model was tested on a benchmark of 5 sets of 450 theoretical instances each. More than half of the tested instances were solved to optimality within 500 s. Results also show that, while overlapping is more beneficial for accelerating project delivery times, it can still have a positive impact on project cost by allowing a better distribution of workload. Finally, overlapping options seem to have less influence on the performance of the model than project slack or number of work packages

    A rough-cut capacity planning model with overlapping

    Get PDF
    In the early phases of projects, capacity planning is performed to assess the feasibility of the project in terms of delivery date, resource usage and cost. This tactical approach relies on an aggregated representation of tasks in work packages. At this level, aggressive project duration objectives are achieved by adopting work package overlapping policies that affect both workload and resource usage. In this article, we propose a mixed-time MILP model for project capacity planning with different possibilities for overlapping levels between work packages. In the model, the planning time horizon is divided into time buckets used to evaluate resource usage, while starting and ending times for work packages are continuous. The model was tested on a benchmark of 5 sets of 450 theoretical instances each. More than half of the tested instances were solved to optimality within 500 s. Results also show that, while overlapping is more beneficial for accelerating project delivery times, it can still have a positive impact on project cost by allowing a better distribution of workload. Finally, overlapping options seem to have less influence on the performance of the model than project slack or number of work packages
    corecore