6,812 research outputs found
Correlated Resource Models of Internet End Hosts
Understanding and modelling resources of Internet end hosts is essential for
the design of desktop software and Internet-distributed applications. In this
paper we develop a correlated resource model of Internet end hosts based on
real trace data taken from the SETI@home project. This data covers a 5-year
period with statistics for 2.7 million hosts. The resource model is based on
statistical analysis of host computational power, memory, and storage as well
as how these resources change over time and the correlations between them. We
find that resources with few discrete values (core count, memory) are well
modeled by exponential laws governing the change of relative resource
quantities over time. Resources with a continuous range of values are well
modeled with either correlated normal distributions (processor speed for
integer operations and floating point operations) or log-normal distributions
(available disk space). We validate and show the utility of the models by
applying them to a resource allocation problem for Internet-distributed
applications, and demonstrate their value over other models. We also make our
trace data and tool for automatically generating realistic Internet end hosts
publicly available
PFS: A Productivity Forecasting System For Desktop Computers To Improve Grid Applications Performance In Enterprise Desktop Grid
An Enterprise Desktop Grid (EDG) is a low cost platform that gathers desktop computers spread over different institutions. This platform uses desktop computers idle time to run Grid applications. We argue that computers in these environments have a predictable productivity that affects a Grid application execution time. In this paper, we propose a system called PFS for computer productivity forecasting that improves Grid applications performance. We simulated 157.500 applications and compared the performance achieved by our proposal against two recent strategies. Our experiments show that a Grid scheduler based on PFS runs applications faster than schedulers based on other selection strategies.Fil: Salinas, Sergio Ariel. Universidad Nacional de Cuyo; ArgentinaFil: Garcia Garino, Carlos Gabriel. Universidad Nacional de Cuyo; ArgentinaFil: Zunino Suarez, Alejandro Octavio. Consejo Nacional de Investigaciones CientĂficas y TĂ©cnicas. Centro CientĂfico TecnolĂłgico Tandil. Instituto Superior de Ingenieria del Software; Argentin
PFS: A Productivity Forecasting System for Desktop Computers to Improve Grid Applications Performance in Enterprise Desktop Grid
An Enterprise Desktop Grid (EDG) is a low cost platform that gathers desktop computers spread over different institutions. This platform uses desktop computers idle time to run Grid applications. We argue that computers in these environments have a predictable productivity that affects a Grid application execution time. In this paper, we propose a system called PFS for computer productivity forecasting that improves Grid applications performance. We simulated 157.500 applications and compared the performance achieved by our proposal against two recent strategies. Our experiments show that a Grid scheduler based on PFS runs applications faster than schedulers based on other selection strategies
Flexible distributed computing with volunteered resources
PhDNowadays, computational grids have evolved to a stage where they can comprise many
volunteered resources owned by different individual users and/or institutions, such as desktop
grids and volunteered computing grids. This brings benefits for large-scale computing, as more
resources are available to exploit. On the other hand, the inherent characteristics of the
volunteered resources bring some challenges for efficiently exploiting them. For example, jobs
may not be able to be executed by some resources, as the computing resources can be
heterogeneous. Furthermore, the resources can be volatile as the resource owners usually have
the right to decide when and how to donate the idle Central Processing Unit (CPU) cycles of
their computers.
Therefore, in order to utilise volunteered resources efficiently, this research investigated
solutions from different aspects. Firstly, this research proposes a new computational Grid
architecture based on Java and Java application migration technologies to provide fundamental
support for coping with these challenges. This proposed architecture supports heterogeneous
resources, ensuring local activities are not affected by Grid jobs and enabling resources to carry
out live and automatic Java application migration.
Secondly, this research work proposes some job-scheduling and migration algorithms based
on resource availability prediction and/or artificial intelligence techniques. To examine the
proposed algorithms, this work includes a series of experiments in both synthetic and practical
scenarios and compares the performance of the proposed algorithms with existing ones across a
variety of scenarios. According to the critical assessment, each algorithm has its own distinct
advantages and performs well when certain conditions are met.
In addition, this research analyses the characteristics of resources in terms of the availability
pattern of practical volunteer-based grids. The analysis shows that each environment has its own
characteristics and each volunteered resource’s availability tends to possess weak correlations
across different days and times-of-day.British Telco
Enhancing reliability with Latin Square redundancy on desktop grids.
Computational grids are some of the largest computer systems in existence today. Unfortunately they are also, in many cases, the least reliable. This research examines the use of redundancy with permutation as a method of improving reliability in computational grid applications. Three primary avenues are explored - development of a new redundancy model, the Replication and Permutation Paradigm (RPP) for computational grids, development of grid simulation software for testing RPP against other redundancy methods and, finally, running a program on a live grid using RPP. An important part of RPP involves distributing data and tasks across the grid in Latin Square fashion. Two theorems and subsequent proofs regarding Latin Squares are developed. The theorems describe the changing position of symbols between the rows of a standard Latin Square. When a symbol is missing because a column is removed the theorems provide a basis for determining the next row and column where the missing symbol can be found. Interesting in their own right, the theorems have implications for redundancy. In terms of the redundancy model, the theorems allow one to state the maximum makespan in the face of missing computational hosts when using Latin Square redundancy. The simulator software was developed and used to compare different data and task distribution schemes on a simulated grid. The software clearly showed the advantage of running RPP, which resulted in faster completion times in the face of computational host failures. The Latin Square method also fails gracefully in that jobs complete with massive node failure while increasing makespan. Finally an Inductive Logic Program (ILP) for pharmacophore search was executed, using a Latin Square redundancy methodology, on a Condor grid in the Dahlem Lab at the University of Louisville Speed School of Engineering. All jobs completed, even in the face of large numbers of randomly generated computational host failures
Recommended from our members
VPeak: Exploiting Volunteer Energy Resources for Flexible Peak Shaving
Traditionally, utility companies have employed demand response for large loads or deployed centralized energy storage to alleviate the effects of peak demand on the grid. The advent of Internet of Things (IoT) and the proliferation of networked energy devices have opened up new opportunities for coordinated control of smaller residential loads at large scales to achieve similar benefits. In this paper, we present VPeak, an approach that uses residential loads volunteered by their owners for coordinated control by a utility for grid optimizations. Since the use of volunteer resources comes with hard limits on how frequently they can be used by a remote utility, we present machine learning techniques for carefully selecting which days to operate these loads based on expected peak demand. VPeak uses a distributed and heterogeneous pool of volunteer loads to implement flexible peak shaving that can either selectively target hotspots within the distribution network or perform grid-wide peak shaving. Our results show that VPeak is able to shave up to 26% of the total demand when selectively shaving peaks at local hotspots and up to 46.7% of the demand for grid-wide peak shaving
Survey and Analysis of Production Distributed Computing Infrastructures
This report has two objectives. First, we describe a set of the production
distributed infrastructures currently available, so that the reader has a basic
understanding of them. This includes explaining why each infrastructure was
created and made available and how it has succeeded and failed. The set is not
complete, but we believe it is representative.
Second, we describe the infrastructures in terms of their use, which is a
combination of how they were designed to be used and how users have found ways
to use them. Applications are often designed and created with specific
infrastructures in mind, with both an appreciation of the existing capabilities
provided by those infrastructures and an anticipation of their future
capabilities. Here, the infrastructures we discuss were often designed and
created with specific applications in mind, or at least specific types of
applications. The reader should understand how the interplay between the
infrastructure providers and the users leads to such usages, which we call
usage modalities. These usage modalities are really abstractions that exist
between the infrastructures and the applications; they influence the
infrastructures by representing the applications, and they influence the ap-
plications by representing the infrastructures
- …