2,822 research outputs found
Learning a Partitioning Advisor with Deep Reinforcement Learning
Commercial data analytics products such as Microsoft Azure SQL Data Warehouse
or Amazon Redshift provide ready-to-use scale-out database solutions for
OLAP-style workloads in the cloud. While the provisioning of a database cluster
is usually fully automated by cloud providers, customers typically still have
to make important design decisions which were traditionally made by the
database administrator such as selecting the partitioning schemes.
In this paper we introduce a learned partitioning advisor for analytical
OLAP-style workloads based on Deep Reinforcement Learning (DRL). The main idea
is that a DRL agent learns its decisions based on experience by monitoring the
rewards for different workloads and partitioning schemes. We evaluate our
learned partitioning advisor in an experimental evaluation with different
databases schemata and workloads of varying complexity. In the evaluation, we
show that our advisor is not only able to find partitionings that outperform
existing approaches for automated partitioning design but that it also can
easily adjust to different deployments. This is especially important in cloud
setups where customers can easily migrate their cluster to a new set of
(virtual) machines
Towards Energy-Proportional Computing for Enterprise-Class Server Workloads
Massive data centers housing thousands of computing nodes
have become commonplace in enterprise computing, and the
power consumption of such data centers is growing at an
unprecedented rate. Adding to the problem is the inability
of the servers to exhibit energy proportionality, i.e., provide
energy-ecient execution under all levels of utilization,
which diminishes the overall energy eciency of the data
center. It is imperative that we realize eective strategies
to control the power consumption of the server and improve
the energy eciency of data centers. With the advent of
Intel Sandy Bridge processors, we have the ability to specify
a limit on power consumption during runtime, which creates
opportunities to design new power-management techniques
for enterprise workloads and make the systems that they run
on more energy-proportional.
In this paper, we investigate whether it is possible to achieve
energy proportionality for an enterprise-class server workload,
namely SPECpower ssj2008 benchmark, by using Intel's
Running Average Power Limit (RAPL) interfaces. First,
we analyze the power consumption and characterize the instantaneous
power prole of the SPECpower benchmark at
a subsystem-level using the on-chip energy meters exposed
via the RAPL interfaces. We then analyze the impact of
RAPL power limiting on the performance, per-transaction
response time, power consumption, and energy eciency of
the benchmark under dierent load levels. Our observations
and results shed light on the ecacy of the RAPL interfaces
and provide guidance for designing power-management techniques
for enterprise-class workloads
Controlling Network Latency in Mixed Hadoop Clusters: Do We Need Active Queue Management?
With the advent of big data, data center applications are processing vast amounts of unstructured and semi-structured data, in parallel on large clusters, across hundreds to thousands of nodes. The highest performance for these batch big data workloads is achieved using expensive network equipment with large buffers, which accommodate bursts in network traffic and allocate bandwidth fairly even when the network is congested. Throughput-sensitive big data applications are, however, often executed in the same data center as latency-sensitive workloads. For both workloads to be supported well, the network must provide both maximum throughput and low latency. Progress has been made in this direction, as modern network switches support Active Queue Management (AQM) and Explicit Congestion Notifications (ECN), both mechanisms to control the level of queue occupancy, reducing the total network latency. This paper is the first study of the effect of Active Queue Management on both throughput and latency, in the context of Hadoop and the MapReduce programming model. We give a quantitative comparison of four different approaches for controlling buffer occupancy and latency: RED and CoDel, both standalone and also combined with ECN and DCTCP network protocol, and identify the AQM configurations that maintain Hadoop execution time gains from larger buffers within 5%, while reducing network packet latency caused by bufferbloat by up to 85%. Finally, we provide recommendations to administrators of Hadoop clusters as to how to improve latency without degrading the throughput of batch big data workloads.The research leading to these results has received funding from the European Unions Seventh Framework Programme (FP7/2007–2013) under grant agreement number 610456 (Euroserver).
The research was also supported by the Ministry of Economy and Competitiveness of Spain under the contracts TIN2012-34557 and TIN2015-65316-P, Generalitat de Catalunya (contracts 2014-SGR-1051 and 2014-SGR-1272), HiPEAC-3 Network of Excellence (ICT- 287759), and the Severo Ochoa Program (SEV-2011-00067) of the Spanish Government.Peer ReviewedPostprint (author's final draft
- …