954 research outputs found
Scalable Performance Analysis of Massively Parallel Stochastic Systems
The accurate performance analysis of large-scale computer and communication systems is directly
inhibited by an exponential growth in the state-space of the underlying Markovian performance
model. This is particularly true when considering massively-parallel architectures
such as cloud or grid computing infrastructures. Nevertheless, an ability to extract quantitative
performance measures such as passage-time distributions from performance models of
these systems is critical for providers of these services. Indeed, without such an ability, they
remain unable to offer realistic end-to-end service level agreements (SLAs) which they can have
any confidence of honouring. Additionally, this must be possible in a short enough period of
time to allow many different parameter combinations in a complex system to be tested. If we
can achieve this rapid performance analysis goal, it will enable service providers and engineers
to determine the cost-optimal behaviour which satisfies the SLAs.
In this thesis, we develop a scalable performance analysis framework for the grouped PEPA
stochastic process algebra. Our approach is based on the approximation of key model quantities
such as means and variances by tractable systems of ordinary differential equations (ODEs).
Crucially, the size of these systems of ODEs is independent of the number of interacting entities
within the model, making these analysis techniques extremely scalable. The reliability of our
approach is directly supported by convergence results and, in some cases, explicit error bounds.
We focus on extracting passage-time measures from performance models since these are very
commonly the language in which a service level agreement is phrased. We design scalable analysis
techniques which can handle passages defined both in terms of entire component populations
as well as individual or tagged members of a large population.
A precise and straightforward specification of a passage-time service level agreement is as important
to the performance engineering process as its evaluation. This is especially true of
large and complex models of industrial-scale systems. To address this, we introduce the unified
stochastic probe framework. Unified stochastic probes are used to generate a model augmentation
which exposes explicitly the SLA measure of interest to the analysis toolkit. In this thesis,
we deploy these probes to define many detailed and derived performance measures that can
be automatically and directly analysed using rapid ODE techniques. In this way, we tackle
applicable problems at many levels of the performance engineering process: from specification
and model representation to efficient and scalable analysis
CloudMedia: When cloud on demand meets video on demand
Internet-based cloud computing is a new computing paradigm aiming to provide agile and scalable resource access in a utility-like fashion. Other than being an ideal platform for computation-intensive tasks, clouds are believed to be also suitable to support large-scale applications with periods of flash crowds by providing elastic amounts of bandwidth and other resources on the fly. The fundamental question is how to configure the cloud utility to meet the highly dynamic demands of such applications at a modest cost. In this paper, we address this practical issue with solid theoretical analysis and efficient algorithm design using Video on Demand (VoD) as the example application. Having intensive bandwidth and storage demands in real time, VoD applications are purportedly ideal candidates to be supported on a cloud platform, where the on-demand resource supply of the cloud meets the dynamic demands of the VoD applications. We introduce a queueing network based model to characterize the viewing behaviors of users in a multichannel VoD application, and derive the server capacities needed to support smooth playback in the channels for two popular streaming models: client-server and P2P. We then propose a dynamic cloud resource provisioning algorithm which, using the derived capacities and instantaneous network statistics as inputs, can effectively support VoD streaming with low cloud utilization cost. Our analysis and algorithm design are verified and extensively evaluated using large-scale experiments under dynamic realistic settings on a home-built cloud platform. © 2011 IEEE.published_or_final_versionThe 31st International Conference on Distributed Computing Systems (ICDCS 2011), Minneapolis, MN., 20-24 June 2011. In Proceedings of 31st ICDCS, 2011, p. 268-27
Application of learning algorithms to traffic management in integrated services networks.
SIGLEAvailable from British Library Document Supply Centre-DSC:DXN027131 / BLDSC - British Library Document Supply CentreGBUnited Kingdo
Estimating customer impatience in a service system with unobserved balking
This paper studies a service system in which arriving customers are provided
with information about the delay they will experience. Based on this
information they decide to wait for service or to leave the system. The main
objective is to estimate the customers' patience-level distribution and the
corresponding potential arrival rate, using knowledge of the actual
queue-length process only. The main complication, and distinguishing feature of
our setup, lies in the fact that customers who decide not to join are not
observed, but, remarkably, we manage to devise a procedure to estimate the load
they would generate. We express our system in terms of a multi-server queue
with a Poisson stream of customers, which allows us to evaluate the
corresponding likelihood function. Estimating the unknown parameters relying on
a maximum likelihood procedure, we prove strong consistency and derive the
asymptotic distribution of the estimation error. Several applications and
extensions of the method are discussed. The performance of our approach is
further assessed through a series of numerical experiments. By fitting
parameters of hyperexponential and generalized-hyperexponential distributions
our method provides a robust estimation framework for any continuous
patience-level distribution
- …