507 research outputs found

    A Taxonomy of Data Grids for Distributed Data Sharing, Management and Processing

    Full text link
    Data Grids have been adopted as the platform for scientific communities that need to share, access, transport, process and manage large data collections distributed worldwide. They combine high-end computing technologies with high-performance networking and wide-area storage management techniques. In this paper, we discuss the key concepts behind Data Grids and compare them with other data sharing and distribution paradigms such as content delivery networks, peer-to-peer networks and distributed databases. We then provide comprehensive taxonomies that cover various aspects of architecture, data transportation, data replication and resource allocation and scheduling. Finally, we map the proposed taxonomy to various Data Grid systems not only to validate the taxonomy but also to identify areas for future exploration. Through this taxonomy, we aim to categorise existing systems to better understand their goals and their methodology. This would help evaluate their applicability for solving similar problems. This taxonomy also provides a "gap analysis" of this area through which researchers can potentially identify new issues for investigation. Finally, we hope that the proposed taxonomy and mapping also helps to provide an easy way for new practitioners to understand this complex area of research.Comment: 46 pages, 16 figures, Technical Repor

    Methodology for modeling high performance distributed and parallel systems

    Get PDF
    Performance modeling of distributed and parallel systems is of considerable importance to the high performance computing community. To achieve high performance, proper task or process assignment and data or file allocation among processing sites is essential. This dissertation describes an elegant approach to model distributed and parallel systems, which combines the optimal static solutions for data allocation with dynamic policies for task assignment. A performance-efficient system model is developed using analytical tools and techniques. The system model is accomplished in three steps. First, the basic client-server model which allows only data transfer is evaluated. A prediction and evaluation method is developed to examine the system behavior and estimate performance measures. The method is based on known product form queueing networks. The next step extends the model so that each site of the system behaves as both client and server. A data-allocation strategy is designed at this stage which optimally assigns the data to the processing sites. The strategy is based on flow deviation technique in queueing models. The third stage considers process-migration policies. A novel on-line adaptive load-balancing algorithm is proposed which dynamically migrates processes and transfers data among different sites to minimize the job execution cost. The gradient-descent rule is used to optimize the cost function, which expresses the cost of process execution at different processing sites. The accuracy of the prediction method and the effectiveness of the analytical techniques is established by the simulations. The modeling procedure described here is general and applicable to any message-passing distributed and parallel system. The proposed techniques and tools can be easily utilized in other related areas such as networking and operating systems. This work contributes significantly towards the design of distributed and parallel systems where performance is critical

    Infrastructure-as-a-Service Usage Determinants in Enterprises

    Get PDF
    The thesis focuses on the research question, what the determinants of Infrastructure-as-a-Service usage of enterprises are. A wide range of IaaS determinants is collected for an IaaS adoption model of enterprises, which is evaluated in a Web survey. As the economical determinants are especially important, they are separately investigated using a cost-optimizing decision support model. This decision support model is then applied to a potential IaaS use case of a large automobile manufacturer

    From Epidemic to Pandemic Modelling

    Get PDF
    We present a methodology for systematically extending epidemic models to multilevel and multiscale spatio-temporal pandemic ones. Our approach builds on the use of coloured stochastic and continuous Petri nets facilitating the sound component-based extension of basic SIR models to include population stratification and also spatio-geographic information and travel connections, represented as graphs, resulting in robust stratified pandemic metapopulation models. This method is inherently easy to use, producing scalable and reusable models with a high degree of clarity and accessibility which can be read either in a deterministic or stochastic paradigm. Our method is supported by a publicly available platform PetriNuts; it enables the visual construction and editing of models; deterministic, stochastic and hybrid simulation as well as structural and behavioural analysis. All the models are available as supplementary material, ensuring reproducibility.Comment: 79 pages (with Appendix), 23 figures, 7 table

    BUSINESS MODELS IN TWO-SIDED MARKETS: AN ASSESSMENT OF STRATEGIES FOR APP PLATFORMS

    Get PDF
    „App platforms” are electronic software distribution markets for mobile devices like smartphones or tablets. They have gained popularity after Apple launched its AppStore in 2008. Since then, app platforms have transformed the entire mobile communication industry including mobile network operators, device producers, software suppliers, content providers, advertisers, etc. Platforms (like AppStore) that intermediate between two distinct groups of customers connected through indirect network effects can be analyzed effectively using the theory of two-sided markets. The interdependencies of customers, platforms and developers require consideration of strategic issues not present in traditional models. These issues may pertain to all development phases, including platform design, launch and competition and thus, have an effect on existing and new business models in this sector. Economics literature on two-sided markets focuses on theoretical analysis, paying not much attention to managerial implications. Strategic management literature, on the other hand, rather provides practical guidelines. Within this paper, we discuss strategic issues arising in the app platform industry, combining these streams of literature. Based on a thorough analysis of the key stakeholders in the app platform industry (platform owner, developers, and users), we use our findings to provide management recommendations and discuss probable evolutions of the industry

    Development of a seamlessly integrated factory planning software tool (prototype) to evaluate and optimize surface mount manufacturing lines

    Get PDF
    Thesis (M.S.)--Massachusetts Institute of Technology, Sloan School of Management, 1995.Includes bibliographical references (p. 182).by Vijay Mehra.M.S

    Improving Usability And Scalability Of Big Data Workflows In The Cloud

    Get PDF
    Big data workflows have recently emerged as the next generation of data-centric workflow technologies to address the five “V” challenges of big data: volume, variety, velocity, veracity, and value. More formally, a big data workflow is the computerized modeling and automation of a process consisting of a set of computational tasks and their data interdependencies to process and analyze data of ever increasing in scale, complexity, and rate of acquisition. The convergence of big data and workflows creates new challenges in workflow community. First, the variety of big data results in a need for integrating large number of remote Web services and other heterogeneous task components that can consume and produce data in various formats and models into a uniform and interoperable workflow. Existing approaches fall short in addressing the so-called shimming problem only in an adhoc manner and unable to provide a generic solution. We automatically insert a piece of code called shims or adaptors in order to resolve the data type mismatches. Second, the volume of big data results in a large number of datasets that needs to be queried and analyzed in an effective and personalized manner. Further, there is also a strong need for sharing, reusing, and repurposing existing tasks and workflows across different users and institutes. To overcome such limitations, we propose a folksonomy- based social workflow recommendation system to improve workflow design productivity and efficient dataset querying and analyzing. Third, the volume of big data results in the need to process and analyze data of ever increasing in scale, complexity, and rate of acquisition. But a scalable distributed data model is still missing that abstracts and automates data distribution, parallelism, and scalable processing. We propose a NoSQL collectional data model that addresses this limitation. Finally, the volume of big data combined with the unbound resource leasing capability foreseen in the cloud, facilitates data scientists to wring actionable insights from the data in a time and cost efficient manner. We propose BARENTS scheduler that supports high-performance workflow scheduling in a heterogeneous cloud-computing environment with a single objective to minimize the workflow makespan under a user provided budget constraint

    VALIDATION OF DETACHED EDDY SIMULATION USING LESTOOL FOR HOMOGENEOUS TURBULENCE

    Get PDF
    Detached Eddy Simulation (DES) is a hybrid turbulence model, a modification to the one-equation model proposed by Spalart and Allmaras (1997) [26]. It combines the advantages of both the RANS and LES models to predict any fluid flow. Presently, the focus is on using Homogeneous Turbulence to test the DES model. In an attempt to scrutinize this model, many cases are considered involving the variance of DES grid spacing parameter, CDES, the grid density, Reynolds number and cases with different initial conditions. Choosing Homogeneous Turbulence for our study alienates complications related to the geometry, boundary conditions and other flow characteristics helping us in studying the behavior of the model thoroughly. Also, the interdependencies of the model grid spacing parameter, grid density and the numerical scheme used are also investigated. Many previous implementations of the DES model have taken the value of CDES=0.65. Through this work, many issues including the sensitivity of CDES will be made clear. The code used in running the test cases is called LESTool, developed at University of Kentucky, Lexington. The two main test cases considered are based on the benchmark experimental study by Comte Bellot and Corrsin (1971) [12] and the Direct Numerical Scheme (DNS) simulation by Blaisdell et al. (1991) [10]
    • …
    corecore