82 research outputs found
InterCloud: Utility-Oriented Federation of Cloud Computing Environments for Scaling of Application Services
Cloud computing providers have setup several data centers at different
geographical locations over the Internet in order to optimally serve needs of
their customers around the world. However, existing systems do not support
mechanisms and policies for dynamically coordinating load distribution among
different Cloud-based data centers in order to determine optimal location for
hosting application services to achieve reasonable QoS levels. Further, the
Cloud computing providers are unable to predict geographic distribution of
users consuming their services, hence the load coordination must happen
automatically, and distribution of services must change in response to changes
in the load. To counter this problem, we advocate creation of federated Cloud
computing environment (InterCloud) that facilitates just-in-time,
opportunistic, and scalable provisioning of application services, consistently
achieving QoS targets under variable workload, resource and network conditions.
The overall goal is to create a computing environment that supports dynamic
expansion or contraction of capabilities (VMs, services, storage, and database)
for handling sudden variations in service demands.
This paper presents vision, challenges, and architectural elements of
InterCloud for utility-oriented federation of Cloud computing environments. The
proposed InterCloud environment supports scaling of applications across
multiple vendor clouds. We have validated our approach by conducting a set of
rigorous performance evaluation study using the CloudSim toolkit. The results
demonstrate that federated Cloud computing model has immense potential as it
offers significant performance gains as regards to response time and cost
saving under dynamic workload scenarios.Comment: 20 pages, 4 figures, 3 tables, conference pape
Workflow scheduling for service oriented cloud computing
Service Orientation (SO) and grid computing are two computing paradigms that when put together using Internet technologies promise to provide a scalable yet flexible computing platform for a diverse set of distributed computing applications. This practice gives rise to the notion of a computing cloud that addresses some previous limitations of interoperability, resource sharing and utilization within distributed computing. In such a Service Oriented Computing Cloud (SOCC), applications are formed by composing a set of services together. In addition, hierarchical service layers are also possible where general purpose services at lower layers are composed to deliver more domain specific services at the higher layer. In general an SOCC is a horizontally scalable computing platform that offers its resources as services in a standardized fashion. Workflow based applications are a suitable target for SOCC where workflow tasks are executed via service calls within the cloud. One or more workflows can be deployed over an SOCC and their execution requires scheduling of services to workflow tasks as the task become ready following their interdependencies. In this thesis heuristics based scheduling policies are evaluated for scheduling workflows over a collection of services offered by the SOCC. Various execution scenarios and workflow characteristics are considered to understand the implication of the heuristic based workflow scheduling
Performance control of internet-based engineering applications.
2006/2007Grazie alle tecnologie capaci di semplificare l'integrazione tra
programmi remoti ospitati da differenti organizzazioni,
le comunitĂ scientifica ed ingegneristica stanno adottando
architetture orientate ai servizi per: aggregare, condividere e
distribuire le loro risorse di calcolo, per gestire grandi
quantitĂ di dati e per eseguire simulazioni attraverso Internet.
I Web Service, per esempio, permettono ad un'organizzazione di
esporre, in Internet, le funzionalitĂ dei loro sistemi e di
renderle scopribili ed accessibili in un modo controllato.
Questo progresso tecnologico può permettere nuove applicazioni
anche nell'area dell'ottimizzazione di progetti. Gli attuali
sistemi di ottimizzazione di progetti sono di solito confinati
all'interno di una singola organizzazione o dipartimento.
D'altra parte, i moderni prodotti manifatturieri sono
l'assemblaggio di componenti provenienti da diverse
organizzazioni. Componendo i servizi delle organizzazioni
coinvolte, si può creare un workflow che descrive il modello del
prodotto composto. Questo servizio composto puo a sua volta
essere usato da un sistema di ottimizzazione
inter-organizzazione.
I compromessi progettuali che sono implicitamente incorporati
per architetture locali, devono essere riconsiderati quando
questi sistemi sono messi in opera su scala globale in Internet.
Ad esempio: i) la qualità delle connessioni tra i nodi può
variare in modo impredicibile; ii) i nodi di terze parti
mantengono il pieno controllo delle loro risorse, incluso, per
esempio, il diritto di diminuire le risorse in modo temporaneo
ed impredicibile.
Dal punto di vista del sistema come un'entitĂ unica, si
vorrebbero massimizzare le prestazioni, cioè, per esempio, il
throughput inteso come numero di progetti candidati valutati per
unitĂ di tempo. Dal punto di vista delle organizzazioni
partecipanti al workflow si vorrebbe, invece, minimizzare il
costo associato ad ogni valutazione. Questo costo può essere un
ostacolo all'adozione del paradigma distribuito, perché le
organizzazioni partecipanti condividono le loro risorse (cioè
CPU, connessioni, larghezza di banda e licenze software) con
altre organizzazioni potenzialmente sconosciute. Minimizzare
questo costo, mentre si mantengono le prestazioni fornite ai
clienti ad un livello accettabile, può essere un potente fattore
per incoraggiare le organizzazioni a condividere effettvivamente
le proprie risorse.
Lo scheduling di istanze di workflows, ovvero stabilire quando
e dove eseguire un certo workflow, in un tale ambiente
multi-organizzazione, multi-livello e geograficamente disperso,
ha un forte impatto sulle prestazioni. Questo lavoro investiga
alcuni dei problemi essenziali di prestazioni e di costo legati
a questo nuovo scenario. Per risolvere i problemi inviduati,
si propone un sistema di controllo dell'accesso adattativo
davanti al workflow engine che limita il numero di esecuzioni
concorrenti. Questa proposta può essere implementata in modo
molto semplice: tratta i servizi come black-box e non richiede
alcuna interazione da parte delle organizzazioni partecipanti.
La tecnica è stata valutata in un ampio spettro di scenari,
attraverso simulazione ad eventi discreti. I risultati
sperimentali suggeriscono che questa tecnica può fornire dei
significativi benefici garantendo alti livelli di throughput
e bassi costi.Thanks to technologies able to simplifying the integration
among remote programs hosted by different organizations,
engineering and scientific communities are embodying service
oriented architectures to aggregate, share and distribute their
computing resources to process and manage large data sets, and
to execute simulations through Internet. Web Service, for
example, allow an organization to expose the functionality of
its internal systems on the Internet and to make it
discoverable and accessible in a controlled manner.
Such a technological advance may enable novel applications also
in the area of design optimization. Current design optimization
systems are usually confined within the boundary of a single
organization or department. Modern engineering products, on the
other hand, are assembled out of components developed by
several organizations. Composing services from the involved
organizations, a model of the composite product can be
described by an appropriate workflow. Such composite service
can then be used by a inter-organizational design optimization
system.
The design trade-offs that have been implicitly incorporated
within local environments, may have to be reconsidered when
deploying these systems on a global scale on the Internet. For
example: i) node-to-node links may vary their service
quality in an unpredictable manner; ii) third party
nodes retains full control over their resources including, e.g.,
the right to decrease the resource amount temporarily and
unpredictably.
From the point of view of the system as a whole, one would like
to maximize the performance, i.e. throughput the number of
candidate design evaluations performed per unit of time. From
the point of view of a participant organization, however, one
would like to minimize the cost associated with each
evaluation. This cost can be an obstacle to the adoption of
this distributed paradigm, because organizations participating
in the composite service share they resources (e.g. CPU, link
bandwidth and software licenses) with other, potentially
unknown, organizations. Minimizing such cost while keeping
performance delivered to clients at an acceptable level can be
a powerful factor for encouraging organizations to indeed share
their services.
The scheduling of workflow instances in such a
multi-organization, multi-tiered and geographically dispersed
environment have strong impacts on performance. This work
investigates some of the fundamental performance and cost
related issues involved in such a novel scenario. We propose an
adaptive admission control to be deployed at the workflow
engine level that limits the number of concurrent jobs. Our
proposal can be implemented very simply: it handles the service
as black-boxes, and it does not require any hook from the
participating organizations.
We evaluated our technique in a broad range of scenarios, by
means of discrete event simulation. Experimental results
suggest that it can provide significant benefits guaranteeing
high level of throughput and low costs.XX Ciclo197
3rd EGEE User Forum
We have organized this book in a sequence of chapters, each chapter associated with an application or technical theme introduced by an overview of the contents, and a summary of the main conclusions coming from the Forum for the chapter topic. The first chapter gathers all the plenary session keynote addresses, and following this there is a sequence of chapters covering the application flavoured sessions. These are followed by chapters with the flavour of Computer Science and Grid Technology. The final chapter covers the important number of practical demonstrations and posters exhibited at the Forum. Much of the work presented has a direct link to specific areas of Science, and so we have created a Science Index, presented below. In addition, at the end of this book, we provide a complete list of the institutes and countries involved in the User Forum
A Model for Scientific Workflows with Parallel and Distributed Computing
In the last decade we witnessed an immense evolution of the computing infrastructures
in terms of processing, storage and communication. On one hand, developments in hardware architectures have made it possible to run multiple virtual machines on a single physical machine. On the other hand, the increase of the available network communication bandwidth has enabled the widespread use of distributed computing infrastructures, for example based on clusters, grids and clouds. The above factors enabled different scientific communities to aim for the development and implementation of complex scientific applications possibly involving large amounts of data. However, due to their structural complexity, these applications require decomposition models to allow multiple tasks running in parallel and distributed environments.
The scientific workflow concept arises naturally as a way to model applications composed of multiple activities. In fact, in the past decades many initiatives have been
undertaken to model application development using the workflow paradigm, both in
the business and in scientific domains. However, despite such intensive efforts, current
scientific workflow systems and tools still have limitations, which pose difficulties to the
development of emerging large-scale, distributed and dynamic applications.
This dissertation proposes the AWARD model for scientific workflows with parallel
and distributed computing. AWARD is an acronym for Autonomic Workflow Activities
Reconfigurable and Dynamic.
The AWARD model has the following main characteristics.
It is based on a decentralized execution control model where multiple autonomic
workflow activities interact by exchanging tokens through input and output ports. The
activities can be executed separately in diverse computing environments, such as in a
single computer or on multiple virtual machines running on distributed infrastructures,
such as clusters and clouds.
It provides basic workflow patterns for parallel and distributed application decomposition and other useful patterns supporting feedback loops and load balancing. The model is suitable to express applications based on a finite or infinite number of iterations, thus allowing to model long-running workflows, which are typical in scientific experimention. A distintive contribution of the AWARD model is the support for dynamic reconfiguration
of long-running workflows. A dynamic reconfiguration allows to modify the
structure of the workflow, for example, to introduce new activities, modify the connections
between activity input and output ports. The activity behavior can also be modified,
for example, by dynamically replacing the activity algorithm.
In addition to the proposal of a new workflow model, this dissertation presents the
implementation of a fully functional software architecture that supports the AWARD
model. The implemented prototype was used to validate and refine the model across
multiple workflow scenarios whose usefulness has been demonstrated in practice clearly, through experimental results, demonstrating the advantages of the major characteristics and contributions of the AWARD model. The implemented prototype was also used to develop application cases, such as a workflow to support the implementation of the MapReduce model and a workflow to support a text mining application developed by an external user.
The extensive experimental work confirmed the adequacy of the AWARD model and
its implementation for developing applications that exploit parallelism and distribution
using the scientific workflows paradigm
Technologies and Applications for Big Data Value
This open access book explores cutting-edge solutions and best practices for big data and data-driven AI applications for the data-driven economy. It provides the reader with a basis for understanding how technical issues can be overcome to offer real-world solutions to major industrial areas. The book starts with an introductory chapter that provides an overview of the book by positioning the following chapters in terms of their contributions to technology frameworks which are key elements of the Big Data Value Public-Private Partnership and the upcoming Partnership on AI, Data and Robotics. The remainder of the book is then arranged in two parts. The first part “Technologies and Methods” contains horizontal contributions of technologies and methods that enable data value chains to be applied in any sector. The second part “Processes and Applications” details experience reports and lessons from using big data and data-driven approaches in processes and applications. Its chapters are co-authored with industry experts and cover domains including health, law, finance, retail, manufacturing, mobility, and smart cities. Contributions emanate from the Big Data Value Public-Private Partnership and the Big Data Value Association, which have acted as the European data community's nucleus to bring together businesses with leading researchers to harness the value of data to benefit society, business, science, and industry. The book is of interest to two primary audiences, first, undergraduate and postgraduate students and researchers in various fields, including big data, data science, data engineering, and machine learning and AI. Second, practitioners and industry experts engaged in data-driven systems, software design and deployment projects who are interested in employing these advanced methods to address real-world problems
A service broker for Intercloud computing
This thesis aims at assisting users in finding the most suitable Cloud resources taking into account their functional and non-functional SLA requirements. A key feature of the work is a Cloud service broker acting as mediator between consumers and Clouds. The research involves the implementation and evaluation of two SLA-aware match-making algorithms by use of a simulation environment. The work investigates also the optimal deployment of Multi-Cloud workflows on Intercloud environments
- …