2 research outputs found
Development details and computational benchmarking of DEPAM
In the big data era of observational oceanography, passive acoustics datasets
are becoming too high volume to be processed on local computers due to their
processor and memory limitations. As a result there is a current need for our
community to turn to cloud-based distributed computing. We present a scalable
computing system for FFT (Fast Fourier Transform)-based features (e.g., Power
Spectral Density) based on the Apache distributed frameworks Hadoop and Spark.
These features are at the core of many different types of acoustic analysis
where the need of processing data at scale with speed is evident, e.g. serving
as long-term averaged learning representations of soundscapes to identify
periods of acoustic interest. In addition to provide a complete description of
our system implementation, we also performed a computational benchmark
comparing our system to three other Scala-only, Matlab and Python based systems
in standalone executions, and evaluated its scalability using the speed up
metric. Our current results are very promising in terms of computational
performance, as we show that our proposed Hadoop/Spark system performs
reasonably well on a single node setup comparatively to state-of-the-art
processing tools used by the PAM community, and that it could also fully
leverage more intensive cluster resources with a almost-linear scalability
behaviour above a certain dataset volume
Addressing Algorithmic Bottlenecks in Elastic Machine Learning with Chicle
Distributed machine learning training is one of the most common and important
workloads running on data centers today, but it is rarely executed alone.
Instead, to reduce costs, computing resources are consolidated and shared by
different applications. In this scenario, elasticity and proper load balancing
are vital to maximize efficiency, fairness, and utilization. Currently, most
distributed training frameworks do not support the aforementioned properties. A
few exceptions that do support elasticity, imitate generic distributed
frameworks and use micro-tasks. In this paper we illustrate that micro-tasks
are problematic for machine learning applications, because they require a high
degree of parallelism which hinders the convergence of distributed training at
a pure algorithmic level (i.e., ignoring overheads and scalability
limitations). To address this, we propose Chicle, a new elastic distributed
training framework which exploits the nature of machine learning algorithms to
implement elasticity and load balancing without micro-tasks. We use Chicle to
train deep neural network as well as generalized linear models, and show that
Chicle achieves performance competitive with state of the art rigid frameworks,
while efficiently enabling elastic execution and dynamic load balancing