11,717 research outputs found
Cloud-based Content Distribution on a Budget
To leverage the elastic nature of cloud computing, a solution provider must be able to accurately gauge demand for its offering. For applications that involve swarm-to-cloud interactions, gauging such demand is not straightforward. In this paper, we propose a general framework, analyze a mathematical model, and present a prototype implementation of a canonical swarm-to-cloud application, namely peer-assisted content delivery. Our system – called Cyclops – dynamically adjusts the off-cloud bandwidth consumed by content servers (which represents the bulk of the provider's cost) to feed a set of swarming clients, based on a feedback signal that gauges the real-time health of the swarm. Our extensive evaluation of Cyclops in a variety of settings – including controlled PlanetLab and live Internet experiments involving thousands of users – show significant reduction in content distribution costs (by as much as two orders of magnitude) when compared to non-feedback-based swarming solutions, with minor impact on content delivery times
Collaborative Data Distribution with BitTorrent for Computational Desktop Grids
Data-centric applications are still a challenging issue for Large Scale Distributed Computing Systems. The emergence of new protocols and softwares for collaborative content distribution over Internet offers a new opportunity for efficient and fast delivery of high volume of data. This paper presents an evaluation of the BitTorrent protocol for Computational Desktop Grids. We first present a prototype of a generic subsystem dedicated to data management and designed to serve as a building block for any Desktop Grid System. Based on this prototype we conduct experimentations to evaluate the potential of BitTorrent compared to a classical approach based on FTP data server. The preliminary results obtained with a 65-nodes cluster measure the basic characteristics of BitTorrent in terms of latency and bandwidth and evaluate the scalability of BitTorrent for the delivery of large input files. Moreover, we show that BitTorrent has a considerable latency overhead compared to FTP but clearly outperforms FTP when distributing large files or files to a high number of nodes. Tests on a synthetic application show that BitTorrent significantly increases the communication/computation ratio of the applications eligible to run on a Desktop Grid System
The Anatomy of the Grid - Enabling Scalable Virtual Organizations
"Grid" computing has emerged as an important new field, distinguished from
conventional distributed computing by its focus on large-scale resource
sharing, innovative applications, and, in some cases, high-performance
orientation. In this article, we define this new field. First, we review the
"Grid problem," which we define as flexible, secure, coordinated resource
sharing among dynamic collections of individuals, institutions, and
resources-what we refer to as virtual organizations. In such settings, we
encounter unique authentication, authorization, resource access, resource
discovery, and other challenges. It is this class of problem that is addressed
by Grid technologies. Next, we present an extensible and open Grid
architecture, in which protocols, services, application programming interfaces,
and software development kits are categorized according to their roles in
enabling resource sharing. We describe requirements that we believe any such
mechanisms must satisfy, and we discuss the central role played by the
intergrid protocols that enable interoperability among different Grid systems.
Finally, we discuss how Grid technologies relate to other contemporary
technologies, including enterprise integration, application service provider,
storage service provider, and peer-to-peer computing. We maintain that Grid
concepts and technologies complement and have much to contribute to these other
approaches.Comment: 24 pages, 5 figure
Estimating Self-Sustainability in Peer-to-Peer Swarming Systems
Peer-to-peer swarming is one of the \emph{de facto} solutions for distributed
content dissemination in today's Internet. By leveraging resources provided by
clients, swarming systems reduce the load on and costs to publishers. However,
there is a limit to how much cost savings can be gained from swarming; for
example, for unpopular content peers will always depend on the publisher in
order to complete their downloads. In this paper, we investigate this
dependence. For this purpose, we propose a new metric, namely \emph{swarm
self-sustainability}. A swarm is referred to as self-sustaining if all its
blocks are collectively held by peers; the self-sustainability of a swarm is
the fraction of time in which the swarm is self-sustaining. We pose the
following question: how does the self-sustainability of a swarm vary as a
function of content popularity, the service capacity of the users, and the size
of the file? We present a model to answer the posed question. We then propose
efficient solution methods to compute self-sustainability. The accuracy of our
estimates is validated against simulation. Finally, we also provide closed-form
expressions for the fraction of time that a given number of blocks is
collectively held by peers.Comment: 27 pages, 5 figure
- …