346 research outputs found

    Workflow scheduling for service oriented cloud computing

    Get PDF
    Service Orientation (SO) and grid computing are two computing paradigms that when put together using Internet technologies promise to provide a scalable yet flexible computing platform for a diverse set of distributed computing applications. This practice gives rise to the notion of a computing cloud that addresses some previous limitations of interoperability, resource sharing and utilization within distributed computing. In such a Service Oriented Computing Cloud (SOCC), applications are formed by composing a set of services together. In addition, hierarchical service layers are also possible where general purpose services at lower layers are composed to deliver more domain specific services at the higher layer. In general an SOCC is a horizontally scalable computing platform that offers its resources as services in a standardized fashion. Workflow based applications are a suitable target for SOCC where workflow tasks are executed via service calls within the cloud. One or more workflows can be deployed over an SOCC and their execution requires scheduling of services to workflow tasks as the task become ready following their interdependencies. In this thesis heuristics based scheduling policies are evaluated for scheduling workflows over a collection of services offered by the SOCC. Various execution scenarios and workflow characteristics are considered to understand the implication of the heuristic based workflow scheduling

    Improving the Elasticity of Services with a P2P Infrastructure

    Get PDF
    Web Services (WS) are one of the most popular approaches for building distributed systems due to their widespread acceptance within industry and academia, established and open standards, a wealth of development tools/infrastructures (e.g. Apache Axis [1]), and maybe most importantly the number of successful deployments. A key issue in the deployment of WS is the elasticity of services e.g. their ability to respond to changes in the demand by dynamically adding or removing WS provider nodes. This paper presents a novel P2P based management approach for WS providers, that allows organizations to dynamically add or remove replicated providers at runtime and thus increases their elasticity. Unlike the current centralized approaches, this scalable P2P approach allows for a fully distributed and fault-tolerant management of replicated provider and thus enables organizations to respond faster to changes in WS consumer behavior

    PADLL: Taming Metadata-intensive HPC Jobs Through Dynamic, Application-agnostic QoS Control

    Full text link
    Modern I/O applications that run on HPC infrastructures are increasingly becoming read and metadata intensive. However, having multiple concurrent applications submitting large amounts of metadata operations can easily saturate the shared parallel file system's metadata resources, leading to overall performance degradation and I/O unfairness. We present PADLL, an application and file system agnostic storage middleware that enables QoS control of data and metadata workflows in HPC storage systems. It adopts ideas from Software-Defined Storage, building data plane stages that mediate and rate limit POSIX requests submitted to the shared file system, and a control plane that holistically coordinates how all I/O workflows are handled. We demonstrate its performance and feasibility under multiple QoS policies using synthetic benchmarks, real-world applications, and traces collected from a production file system. Results show that PADLL can enforce complex storage QoS policies over concurrent metadata-aggressive jobs, ensuring fairness and prioritization.Comment: To appear at 23rd IEEE/ACM International Symposium on Cluster, Cloud and Internet Computing (CCGrid'23

    An Uncertainty Visual Analytics Framework for Functional Magnetic Resonance Imaging

    Get PDF
    Improving understanding of the human brain is one of the leading pursuits of modern scientific research. Functional magnetic resonance imaging (fMRI) is a foundational technique for advanced analysis and exploration of the human brain. The modality scans the brain in a series of temporal frames which provide an indication of the brain activity either at rest or during a task. The images can be used to study the workings of the brain, leading to the development of an understanding of healthy brain function, as well as characterising diseases such as schizophrenia and bipolar disorder. Extracting meaning from fMRI relies on an analysis pipeline which can be broadly categorised into three phases: (i) data acquisition and image processing; (ii) image analysis; and (iii) visualisation and human interpretation. The modality and analysis pipeline, however, are hampered by a range of uncertainties which can greatly impact the study of the brain function. Each phase contains a set of required and optional steps, containing inherent limitations and complex parameter selection. These aspects lead to the uncertainty that impacts the outcome of studies. Moreover, the uncertainties that arise early in the pipeline, are compounded by decisions and limitations further along in the process. While a large amount of research has been undertaken to examine the limitations and variable parameter selection, statistical approaches designed to address the uncertainty have not managed to mitigate the issues. Visual analytics, meanwhile, is a research domain which seeks to combine advanced visual interfaces with specialised interaction and automated statistical processing designed to exploit human expertise and understanding. Uncertainty visual analytics (UVA) tools, which aim to minimise and mitigate uncertainties, have been proposed for a variety of data, including astronomical, financial, weather and crime. Importantly, UVA approaches have also seen success in medical imaging and analysis. However, there are many challenges surrounding the application of UVA to each research domain. Principally, these involve understanding what the uncertainties are and the possible effects so they may be connected to visualisation and interaction approaches. With fMRI, the breadth of uncertainty arising in multiple stages along the pipeline and the compound effects, make it challenging to propose UVAs which meaningfully integrate into pipeline. In this thesis, we seek to address this challenge by proposing a unified UVA framework for fMRI. To do so, we first examine the state-of-the-art landscape of fMRI uncertainties, including the compound effects, and explore how they are currently addressed. This forms the basis of a field we term fMRI-UVA. We then present our overall framework, which is designed to meet the requirements of fMRI visual analysis, while also providing an indication and understanding of the effects of uncertainties on the data. Our framework consists of components designed for the spatial, temporal and processed imaging data. Alongside the framework, we propose two visual extensions which can be used as standalone UVA applications or be integrated into the framework. Finally, we describe a conceptual algorithmic approach which incorporates more data into an existing measure used in the fMRI analysis pipeline

    Workshop Report: Container Based Analysis Environments for Research Data Access and Computing

    Get PDF
    Report of the first workshop on Container Based Analysis Environments for Research Data Access and Computing supported by the National Data Service and Data Exploration Lab and held at the National Center for Supercomputing Applications (NCSA) at the University of Illinois at Urbana-Champaign

    Image informatics strategies for deciphering neuronal network connectivity

    Get PDF
    Brain function relies on an intricate network of highly dynamic neuronal connections that rewires dramatically under the impulse of various external cues and pathological conditions. Among the neuronal structures that show morphologi- cal plasticity are neurites, synapses, dendritic spines and even nuclei. This structural remodelling is directly connected with functional changes such as intercellular com- munication and the associated calcium-bursting behaviour. In vitro cultured neu- ronal networks are valuable models for studying these morpho-functional changes. Owing to the automation and standardisation of both image acquisition and image analysis, it has become possible to extract statistically relevant readout from such networks. Here, we focus on the current state-of-the-art in image informatics that enables quantitative microscopic interrogation of neuronal networks. We describe the major correlates of neuronal connectivity and present workflows for analysing them. Finally, we provide an outlook on the challenges that remain to be addressed, and discuss how imaging algorithms can be extended beyond in vitro imaging studies

    MOSTO: A toolkit to facilitate security auditing of ICS devices using Modbus/TCP

    Get PDF
    The integration of the Internet into industrial plants has connected Industrial Control Systems (ICS) worldwide, resulting in an increase in the number of attack surfaces and the exposure of software and devices not originally intended for networking. In addition, the heterogeneity and technical obsolescence of ICS architectures, legacy hardware, and outdated software pose significant challenges. Since these systems control essential infrastructure such as power grids, water treatment plants, and transportation networks, security is of the utmost importance. Unfortunately, current methods for evaluating the security of ICS are often ad-hoc and difficult to formalize into a systematic evaluation methodology with predictable results. In this paper, we propose a practical method supported by a concrete toolkit for performing penetration testing in an industrial setting. The primary focus is on the Modbus/TCP protocol as the field control protocol. Our approach relies on a toolkit, named MOSTO, which is licensed under GNU GPL and enables auditors to assess the security of existing industrial control settings without interfering with ICS workflows. Furthermore, we present a model-driven framework that combines formal methods, testing techniques, and simulation to (formally) test security properties in ICS networks

    Datacenter Traffic Control: Understanding Techniques and Trade-offs

    Get PDF
    Datacenters provide cost-effective and flexible access to scalable compute and storage resources necessary for today's cloud computing needs. A typical datacenter is made up of thousands of servers connected with a large network and usually managed by one operator. To provide quality access to the variety of applications and services hosted on datacenters and maximize performance, it deems necessary to use datacenter networks effectively and efficiently. Datacenter traffic is often a mix of several classes with different priorities and requirements. This includes user-generated interactive traffic, traffic with deadlines, and long-running traffic. To this end, custom transport protocols and traffic management techniques have been developed to improve datacenter network performance. In this tutorial paper, we review the general architecture of datacenter networks, various topologies proposed for them, their traffic properties, general traffic control challenges in datacenters and general traffic control objectives. The purpose of this paper is to bring out the important characteristics of traffic control in datacenters and not to survey all existing solutions (as it is virtually impossible due to massive body of existing research). We hope to provide readers with a wide range of options and factors while considering a variety of traffic control mechanisms. We discuss various characteristics of datacenter traffic control including management schemes, transmission control, traffic shaping, prioritization, load balancing, multipathing, and traffic scheduling. Next, we point to several open challenges as well as new and interesting networking paradigms. At the end of this paper, we briefly review inter-datacenter networks that connect geographically dispersed datacenters which have been receiving increasing attention recently and pose interesting and novel research problems.Comment: Accepted for Publication in IEEE Communications Surveys and Tutorial
    corecore