12 research outputs found

    Objcache: An Elastic Filesystem over External Persistent Storage for Container Clusters

    Full text link
    Container virtualization enables emerging AI workloads such as model serving, highly parallelized training, machine learning pipelines, and so on, to be easily scaled on demand on the elastic cloud infrastructure. Particularly, AI workloads require persistent storage to store data such as training inputs, models, and checkpoints. An external storage system like cloud object storage is a common choice because of its elasticity and scalability. To mitigate access latency to external storage, caching at a local filesystem is an essential technique. However, building local caches on scaling clusters must cope with explosive disk usage, redundant networking, and unexpected failures. We propose objcache, an elastic filesystem over external storage. Objcache introduces an internal transaction protocol over Raft logging to enable atomic updates of distributed persistent states with consistent hashing. The proposed transaction protocol can also manage inode dirtiness by maintaining the consistency between the local cache and external storage. Objcache supports scaling down to zero by automatically evicting dirty files to external storage. Our evaluation reports that objcache speeded up model serving startup by 98.9% compared to direct copies via S3 interfaces. Scaling up with dirty files completed from 2 to 14 seconds with 1024 dirty files.Comment: 13 page

    DPFS:DPU-Powered File System Virtualization

    No full text
    As we move towards hyper-converged cloud solutions, the efficiency and overheads of distributed file systems at the cloud tenant side (i.e., client) become of paramount importance. Often, the clientside driver of a cloud file system is complex and CPU intensive, deeply coupled with the backend implementation, and requires optimizing multiple intrusive knobs. In this work, we propose to decouple the file system client from its backend implementation by virtualizing it with an off-the-shelf DPU using the Linux virtio-fs software stack. The decoupling allows us to offload the file system client execution to a DPU, which is managed and optimized by the cloud provider, while freeing the host CPU cycles. DPFS, our proposed framework, is 4.4Ă— more host CPU efficient per I/O, delivers comparable performance to a tenant with zero-configuration and without modification of their host software stack, while allowing workload and hardware specific backend optimizations. The DPFS framework and its artifacts are publically available at https://github.com/IBM/DPFS.</p

    An Elastic Ephemeral Datastore using Cheap, Transient Cloud Resources

    Full text link
    Spot instances are virtual machines offered at 60-90% lower cost that can be reclaimed at any time, with only a short warning period. Spot instances have already been used to significantly reduce the cost of processing workloads in the cloud. However, leveraging spot instances to reduce the cost of stateful cloud applications is much more challenging, as the sudden preemptions lead to data loss. In this work, we propose leveraging spot instances to decrease the cost of ephemeral data management in distributed data analytics applications. We specifically target ephemeral data as this large class of data in modern analytics workloads has low durability requirements; if lost, the data can be regenerated by re-executing compute tasks. We design an elastic, distributed ephemeral datastore that handles node preemptions transparently to user applications and minimizes data loss by redistributing data during node preemption warning periods. We implement our elastic datastore on top of the Apache Crail datastore and evaluate the system with various workloads and VM types. By leveraging spot instances, we show that we can run TPC-DS queries with 60\% lower cost compared to using on-demand VMs for the datastore, while only increasing end-to-end execution time by 2.1%

    Understanding Modern Storage APIs: A systematic study of libaio, SPDK, and io-uring

    No full text
    Recent high-performance storage devices have exposed software inefficiencies in existing storage stacks, leading to a new breed of I/O stacks. The newest storage API of the Linux kernel is io-uring. We perform one of the first in-depth studies of io-uring, and compare its performance and dis-/advantages with the established libaio and SPDK APIs. Our key findings reveal that (i) polling design significantly impacts performance; (ii) with enough CPU cores io-uring can deliver performance close to that of SPDK; and (iii) performance scalability over multiple CPU cores and devices requires careful consideration and necessitates a hybrid approach. Last, we provide design guidelines for developers of storage intensive applications

    Unification of temporary storage in the nodekernel architecture

    No full text
    Efficiently exchanging temporary data between tasks is critical to the end-to-end performance of many data processing frameworks and applications. Unfortunately, the diverse nature of temporary data creates storage demands that often fall between the sweet spots of traditional storage platforms, such as file systems or key-value stores. We present NodeKernel, a novel distributed storage architecture that offers a convenient new point in the design space by fusing file system and key-value semantics in a common storage kernel while leveraging modern networking and storage hardware to achieve high performance and cost-efficiency. NodeKernel provides hierarchical naming, high scalability, and close to bare-metal performance for a wide range of data sizes and access patterns that are characteristic of temporary data. We show that storing temporary data in Crail, our concrete implementation of the NodeKernel architecture which uses RDMA networking with tiered DRAM/NVMe-Flash storage, improves NoSQL workload performance by up to 4.8Ă— and Spark application performance by up to 3.4Ă—. Furthermore, by storing data across NVMe Flash and DRAM storage tiers, Crail reduces storage cost by up to 8Ă— compared to DRAM-only storage systems

    Unification of temporary storage in the nodekernel architecture

    No full text
    Efficiently exchanging temporary data between tasks is critical to the end-to-end performance of many data processing frameworks and applications. Unfortunately, the diverse nature of temporary data creates storage demands that often fall between the sweet spots of traditional storage platforms, such as file systems or key-value stores. We present NodeKernel, a novel distributed storage architecture that offers a convenient new point in the design space by fusing file system and key-value semantics in a common storage kernel while leveraging modern networking and storage hardware to achieve high performance and cost-efficiency. NodeKernel provides hierarchical naming, high scalability, and close to bare-metal performance for a wide range of data sizes and access patterns that are characteristic of temporary data. We show that storing temporary data in Crail, our concrete implementation of the NodeKernel architecture which uses RDMA networking with tiered DRAM/NVMe-Flash storage, improves NoSQL workload performance by up to 4.8Ă— and Spark application performance by up to 3.4Ă—. Furthermore, by storing data across NVMe Flash and DRAM storage tiers, Crail reduces storage cost by up to 8Ă— compared to DRAM-only storage systems

    The SARS-CoV-2 main protease M-pro causes microvascular brain pathology by cleaving NEMO in brain endothelial cells

    No full text
    Coronavirus disease 2019 (COVID-19) can damage cerebral small vessels and cause neurological symptoms. Here we describe structural changes in cerebral small vessels of patients with COVID-19 and elucidate potential mechanisms underlying the vascular pathology. In brains of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2)-infected individuals and animal models, we found an increased number of empty basement membrane tubes, so-called string vessels representing remnants of lost capillaries. We obtained evidence that brain endothelial cells are infected and that the main protease of SARS-CoV-2 (M-pro) cleaves NEMO, the essential modulator of nuclear factor-kappa B. By ablating NEMO, M-pro induces the death of human brain endothelial cells and the occurrence of string vessels in mice. Deletion of receptor-interacting protein kinase (RIPK) 3, a mediator of regulated cell death, blocks the vessel rarefaction and disruption of the blood-brain barrier due to NEMO ablation. Importantly, a pharmacological inhibitor of RIPK signaling prevented the M-pro-induced microvascular pathology. Our data suggest RIPK as a potential therapeutic target to treat the neuropathology of COVID-19. A novel study led by scientists in Lubeck, Germany, shows that SARS-CoV-2-infected brain endothelial cells undergo cell death due to the cleavage of NEMO by the viral protease M-pro, potentially causing cerebral COVID-19 and 'long COVID' symptoms
    corecore