2,977 research outputs found
Resource provisioning in Science Clouds: Requirements and challenges
Cloud computing has permeated into the information technology industry in the
last few years, and it is emerging nowadays in scientific environments. Science
user communities are demanding a broad range of computing power to satisfy the
needs of high-performance applications, such as local clusters,
high-performance computing systems, and computing grids. Different workloads
are needed from different computational models, and the cloud is already
considered as a promising paradigm. The scheduling and allocation of resources
is always a challenging matter in any form of computation and clouds are not an
exception. Science applications have unique features that differentiate their
workloads, hence, their requirements have to be taken into consideration to be
fulfilled when building a Science Cloud. This paper will discuss what are the
main scheduling and resource allocation challenges for any Infrastructure as a
Service provider supporting scientific applications
Business-driven resource allocation and management for data centres in cloud computing markets
Cloud Computing markets arise as an efficient way to allocate resources for the execution of tasks and services within a set of geographically dispersed providers from different organisations. Client applications and service providers meet in a market and negotiate for the sales of services by means of the signature of a Service Level Agreement that contains the Quality of Service terms that the Cloud provider has to guarantee by managing properly its resources.
Current implementations of Cloud markets suffer from a lack of information flow between the negotiating agents, which sell the resources, and the resource managers that allocate the resources to fulfil the agreed Quality of Service. This thesis establishes an intermediate layer between the market agents and the resource managers. In consequence, agents can perform accurate negotiations by considering the status of the resources in their negotiation models, and providers can manage their resources considering both the performance and the business objectives. This thesis defines a set of policies for the negotiation and enforcement of Service Level Agreements. Such policies deal with different Business-Level Objectives: maximisation of the revenue, classification of clients, trust and reputation maximisation, and risk minimisation. This thesis demonstrates the effectiveness of such policies by means of fine-grained simulations.
A pricing model may be influenced by many parameters. The weight of such parameters within the final model is not always known, or it can change as the market environment evolves. This thesis models and evaluates how the providers can self-adapt to changing environments by means of genetic algorithms. Providers that rapidly adapt to changes in the environment achieve higher revenues than providers that do not.
Policies are usually conceived for the short term: they model the behaviour of the system by considering the current status and the expected immediate after their application. This thesis defines and evaluates a trust and reputation system that enforces providers to consider the impact of their decisions in the long term. The trust and reputation system expels providers and clients with dishonest behaviour, and providers that consider the impact of their reputation in their actions improve on the achievement of their Business-Level Objectives.
Finally, this thesis studies the risk as the effects of the uncertainty over the expected outcomes of cloud providers. The particularities of cloud appliances as a set of interconnected resources are studied, as well as how the risk is propagated through the linked nodes. Incorporating risk models helps providers differentiate Service Level Agreements according to their risk, take preventive actions in the focus of the risk, and pricing accordingly. Applying risk management raises the fulfilment rate of the Service-Level Agreements and increases the profit of the providerPostprint (published version
On Allocation Policies for Power and Performance
With the increasing popularity of Internet-based services and applications,
power efficiency is becoming a major concern for data center operators, as high
electricity consumption not only increases greenhouse gas emissions, but also
increases the cost of running the server farm itself. In this paper we address
the problem of maximizing the revenue of a service provider by means of dynamic
allocation policies that run the minimum amount of servers necessary to meet
user's requirements in terms of performance. The results of several experiments
executed using Wikipedia traces are described, showing that the proposed
schemes work well, even if the workload is non-stationary. Since any resource
allocation policy requires the use of forecasting mechanisms, various schemes
allowing compensating errors in the load forecasts are presented and evaluated.Comment: 8 pages, 11 figures, 2010 11th IEEE/ACM International Conference on
Grid Computing (GRID), pp 313 - 320 (E2GC2-2010 workshop
Datacenter Traffic Control: Understanding Techniques and Trade-offs
Datacenters provide cost-effective and flexible access to scalable compute
and storage resources necessary for today's cloud computing needs. A typical
datacenter is made up of thousands of servers connected with a large network
and usually managed by one operator. To provide quality access to the variety
of applications and services hosted on datacenters and maximize performance, it
deems necessary to use datacenter networks effectively and efficiently.
Datacenter traffic is often a mix of several classes with different priorities
and requirements. This includes user-generated interactive traffic, traffic
with deadlines, and long-running traffic. To this end, custom transport
protocols and traffic management techniques have been developed to improve
datacenter network performance.
In this tutorial paper, we review the general architecture of datacenter
networks, various topologies proposed for them, their traffic properties,
general traffic control challenges in datacenters and general traffic control
objectives. The purpose of this paper is to bring out the important
characteristics of traffic control in datacenters and not to survey all
existing solutions (as it is virtually impossible due to massive body of
existing research). We hope to provide readers with a wide range of options and
factors while considering a variety of traffic control mechanisms. We discuss
various characteristics of datacenter traffic control including management
schemes, transmission control, traffic shaping, prioritization, load balancing,
multipathing, and traffic scheduling. Next, we point to several open challenges
as well as new and interesting networking paradigms. At the end of this paper,
we briefly review inter-datacenter networks that connect geographically
dispersed datacenters which have been receiving increasing attention recently
and pose interesting and novel research problems.Comment: Accepted for Publication in IEEE Communications Surveys and Tutorial
Cooperative Multi-Bitrate Video Caching and Transcoding in Multicarrier NOMA-Assisted Heterogeneous Virtualized MEC Networks
Cooperative video caching and transcoding in mobile edge computing (MEC)
networks is a new paradigm for future wireless networks, e.g., 5G and 5G
beyond, to reduce scarce and expensive backhaul resource usage by prefetching
video files within radio access networks (RANs). Integration of this technique
with other advent technologies, such as wireless network virtualization and
multicarrier non-orthogonal multiple access (MC-NOMA), provides more flexible
video delivery opportunities, which leads to enhancements both for the
network's revenue and for the end-users' service experience. In this regard, we
propose a two-phase RAF for a parallel cooperative joint multi-bitrate video
caching and transcoding in heterogeneous virtualized MEC networks. In the cache
placement phase, we propose novel proactive delivery-aware cache placement
strategies (DACPSs) by jointly allocating physical and radio resources based on
network stochastic information to exploit flexible delivery opportunities.
Then, for the delivery phase, we propose a delivery policy based on the user
requests and network channel conditions. The optimization problems
corresponding to both phases aim to maximize the total revenue of network
slices, i.e., virtual networks. Both problems are non-convex and suffer from
high-computational complexities. For each phase, we show how the problem can be
solved efficiently. We also propose a low-complexity RAF in which the
complexity of the delivery algorithm is significantly reduced. A Delivery-aware
cache refreshment strategy (DACRS) in the delivery phase is also proposed to
tackle the dynamically changes of network stochastic information. Extensive
numerical assessments demonstrate a performance improvement of up to 30% for
our proposed DACPSs and DACRS over traditional approaches.Comment: 53 pages, 24 figure
- …