7 research outputs found
Don't Repeat Yourself: Seamless Execution and Analysis of Extensive Network Experiments
This paper presents MACI, the first bespoke framework for the management, the
scalable execution, and the interactive analysis of a large number of network
experiments. Driven by the desire to avoid repetitive implementation of just a
few scripts for the execution and analysis of experiments, MACI emerged as a
generic framework for network experiments that significantly increases
efficiency and ensures reproducibility. To this end, MACI incorporates and
integrates established simulators and analysis tools to foster rapid but
systematic network experiments.
We found MACI indispensable in all phases of the research and development
process of various communication systems, such as i) an extensive DASH video
streaming study, ii) the systematic development and improvement of Multipath
TCP schedulers, and iii) research on a distributed topology graph pattern
matching algorithm. With this work, we make MACI publicly available to the
research community to advance efficient and reproducible network experiments
In Datacenter Performance, The Only Constant Is Change
All computing infrastructure suffers from performance variability, be it
bare-metal or virtualized. This phenomenon originates from many sources: some
transient, such as noisy neighbors, and others more permanent but sudden, such
as changes or wear in hardware, changes in the underlying hypervisor stack, or
even undocumented interactions between the policies of the computing resource
provider and the active workloads. Thus, performance measurements obtained on
clouds, HPC facilities, and, more generally, datacenter environments are almost
guaranteed to exhibit performance regimes that evolve over time, which leads to
undesirable nonstationarities in application performance. In this paper, we
present our analysis of performance of the bare-metal hardware available on the
CloudLab testbed where we focus on quantifying the evolving performance regimes
using changepoint detection. We describe our findings, backed by a dataset with
nearly 6.9M benchmark results collected from over 1600 machines over a period
of 2 years and 9 months. These findings yield a comprehensive characterization
of real-world performance variability patterns in one computing facility, a
methodology for studying such patterns on other infrastructures, and contribute
to a better understanding of performance variability in general.Comment: To be presented at the 20th IEEE/ACM International Symposium on
Cluster, Cloud and Internet Computing (CCGrid,
http://cloudbus.org/ccgrid2020/) on May 11-14, 2020 in Melbourne, Victoria,
Australi
Recommended from our members
Optimizing Computations and Allocations on High Performance and Cloud Computing Systems
Over the last decade, many research and development projects have focused on Cloud Computing systems. After forming around the early research papers and the first commercial cloud offerings in 2006-2008, the field has seen a tremendous progress and has provided the primary infrastructure and technology for applications at small, medium, and large scales. Cloud Computing systems have provided diverse on-demand resources to individual researchers and developers, groups and entire institutions, as well as commercial companies and government organizations. Clouds have also found their niche in scientific computing applications, offering attractive alternatives to High Performance Computing models and systems. While cloud economics and technologies have significantly matured recently, there is much active research revolving around topics such as optimality, usability, manageability, and reproducibility in the latest studies. This dissertation presents our findings and relevant developments at the intersection of Cloud Computing and such “flavors” of computing as High Performance Computing and High Throughput Computing. We primarily focus on optimality issues in this area and propose solutions that address the needs of individual researchers and research groups with limited computational and financial resources
Recommended from our members
Optimizing Computations and Allocations on High Performance and Cloud Computing Systems
Over the last decade, many research and development projects have focused on Cloud Computing systems. After forming around the early research papers and the first commercial cloud offerings in 2006-2008, the field has seen a tremendous progress and has provided the primary infrastructure and technology for applications at small, medium, and large scales. Cloud Computing systems have provided diverse on-demand resources to individual researchers and developers, groups and entire institutions, as well as commercial companies and government organizations. Clouds have also found their niche in scientific computing applications, offering attractive alternatives to High Performance Computing models and systems. While cloud economics and technologies have significantly matured recently, there is much active research revolving around topics such as optimality, usability, manageability, and reproducibility in the latest studies. This dissertation presents our findings and relevant developments at the intersection of Cloud Computing and such “flavors” of computing as High Performance Computing and High Throughput Computing. We primarily focus on optimality issues in this area and propose solutions that address the needs of individual researchers and research groups with limited computational and financial resources
Modeling for inversion in exploration geophysics
Seismic inversion, and more generally geophysical exploration, aims at better understanding the earth's subsurface, which is one of today's most important challenges. Firstly, it contains natural resources that are critical to our technologies such as water, minerals and oil and gas. Secondly, monitoring the subsurface in the context of CO2 sequestration, earthquake detection and global seismology are of major interests with regard to safety and the environment hazards. However, the technologies to monitor the subsurface or find resources are scientifically extremely challenging. Seismic inversion can be formulated as a mathematical optimization problem that minimizes the difference between field recorded data and numerically modeled synthetic data. The process of solving this optimization problem then requires to numerically model, thousands of times, wave-propagation in large three-dimensional representations of part of the earth subsurface. The mathematical and computational complexity of this problem, therefore, calls for software design that abstracts these requirements and facilitates algorithm and software development. My thesis addresses some of the challenges that arise from these problems; mainly the computational cost and access to the right software for research and development. In the first part, I will discuss a performance metric that improves the current runtime-only benchmarks in exploration geophysics. This metric, the roofline model, first provides insight at the hardware level of the performance of a given implementation relative to the maximum achievable performance. Second, this study demonstrates that the choice of numerical discretization has a major impact on the achievable performance depending on the hardware at hand and shows that a flexible framework with respect to the discretization parameters is necessary. In the second part, I will introduce and describe Devito, a symbolic finite-difference DSL that provides a high-level interface to the definition of partial differential equations (PDE) such as the wave equation. Devito, from the symbolic definition of PDEs, then generates and compiles highly optimized C code on-the-fly to compute the solution of the PDE. The combination of the high-level abstractions and the just-in-time compiler enable research for geophysical exploration and PDE-constrainted optimization based on the paradigm of separation of concerns. This allows researchers to concentrate on their respective field of study while having access to computationally performant solvers with a flexible and easy to use interface to successfully implement complex representations of the physics. The second part of my thesis will be split into two sub-parts; first describing the symbolic application programming interface (API), before describing and benchmarking the just-in-time compiler. I will end my thesis with concluding remarks, the latest developments and a brief description of projects that were enabled by Devito.Ph.D