507 research outputs found
A Taxonomy of Data Grids for Distributed Data Sharing, Management and Processing
Data Grids have been adopted as the platform for scientific communities that
need to share, access, transport, process and manage large data collections
distributed worldwide. They combine high-end computing technologies with
high-performance networking and wide-area storage management techniques. In
this paper, we discuss the key concepts behind Data Grids and compare them with
other data sharing and distribution paradigms such as content delivery
networks, peer-to-peer networks and distributed databases. We then provide
comprehensive taxonomies that cover various aspects of architecture, data
transportation, data replication and resource allocation and scheduling.
Finally, we map the proposed taxonomy to various Data Grid systems not only to
validate the taxonomy but also to identify areas for future exploration.
Through this taxonomy, we aim to categorise existing systems to better
understand their goals and their methodology. This would help evaluate their
applicability for solving similar problems. This taxonomy also provides a "gap
analysis" of this area through which researchers can potentially identify new
issues for investigation. Finally, we hope that the proposed taxonomy and
mapping also helps to provide an easy way for new practitioners to understand
this complex area of research.Comment: 46 pages, 16 figures, Technical Repor
Methodology for modeling high performance distributed and parallel systems
Performance modeling of distributed and parallel systems is of considerable importance to the high performance computing community. To achieve high performance, proper task or process assignment and data or file allocation among processing sites is essential. This dissertation describes an elegant approach to model distributed and parallel systems, which combines the optimal static solutions for data allocation with dynamic policies for task assignment. A performance-efficient system model is developed using analytical tools and techniques.
The system model is accomplished in three steps. First, the basic client-server model which allows only data transfer is evaluated. A prediction and evaluation method is developed to examine the system behavior and estimate performance measures. The method is based on known product form queueing networks. The next step extends the model so that each site of the system behaves as both client and server. A data-allocation strategy is designed at this stage which optimally assigns the data to the processing sites. The strategy is based on flow deviation technique in queueing models. The third stage considers process-migration policies. A novel on-line adaptive load-balancing algorithm is proposed which dynamically migrates processes and transfers data among different sites to minimize the job execution cost. The gradient-descent rule is used to optimize the cost function, which expresses the cost of process execution at different processing sites.
The accuracy of the prediction method and the effectiveness of the analytical techniques is established by the simulations. The modeling procedure described here is general and applicable to any message-passing distributed and parallel system. The proposed techniques and tools can be easily utilized in other related areas such as networking and operating systems. This work contributes significantly towards the design of distributed and parallel systems where performance is critical
Infrastructure-as-a-Service Usage Determinants in Enterprises
The thesis focuses on the research question, what the determinants of Infrastructure-as-a-Service usage of enterprises are. A wide range of IaaS determinants is collected for an IaaS adoption model of enterprises, which is evaluated in a Web survey. As the economical determinants are especially important, they are separately investigated using a cost-optimizing decision support model. This decision support model is then applied to a potential IaaS use case of a large automobile manufacturer
From Epidemic to Pandemic Modelling
We present a methodology for systematically extending epidemic models to
multilevel and multiscale spatio-temporal pandemic ones. Our approach builds on
the use of coloured stochastic and continuous Petri nets facilitating the sound
component-based extension of basic SIR models to include population
stratification and also spatio-geographic information and travel connections,
represented as graphs, resulting in robust stratified pandemic metapopulation
models. This method is inherently easy to use, producing scalable and reusable
models with a high degree of clarity and accessibility which can be read either
in a deterministic or stochastic paradigm. Our method is supported by a
publicly available platform PetriNuts; it enables the visual construction and
editing of models; deterministic, stochastic and hybrid simulation as well as
structural and behavioural analysis. All the models are available as
supplementary material, ensuring reproducibility.Comment: 79 pages (with Appendix), 23 figures, 7 table
BUSINESS MODELS IN TWO-SIDED MARKETS: AN ASSESSMENT OF STRATEGIES FOR APP PLATFORMS
„App platforms” are electronic software distribution markets for mobile devices like smartphones or tablets. They have gained popularity after Apple launched its AppStore in 2008. Since then, app platforms have transformed the entire mobile communication industry including mobile network operators, device producers, software suppliers, content providers, advertisers, etc. Platforms (like AppStore) that intermediate between two distinct groups of customers connected through indirect network effects can be analyzed effectively using the theory of two-sided markets. The interdependencies of customers, platforms and developers require consideration of strategic issues not present in traditional models. These issues may pertain to all development phases, including platform design, launch and competition and thus, have an effect on existing and new business models in this sector. Economics literature on two-sided markets focuses on theoretical analysis, paying not much attention to managerial implications. Strategic management literature, on the other hand, rather provides practical guidelines. Within this paper, we discuss strategic issues arising in the app platform industry, combining these streams of literature. Based on a thorough analysis of the key stakeholders in the app platform industry (platform owner, developers, and users), we use our findings to provide management recommendations and discuss probable evolutions of the industry
Development of a seamlessly integrated factory planning software tool (prototype) to evaluate and optimize surface mount manufacturing lines
Thesis (M.S.)--Massachusetts Institute of Technology, Sloan School of Management, 1995.Includes bibliographical references (p. 182).by Vijay Mehra.M.S
Recommended from our members
From Petascale to Exascale: Eight Focus Areas of R&D Challenges for HPC Simulation Environments
Programming models bridge the gap between the underlying hardware architecture and the supporting layers of software available to applications. Programming models are different from both programming languages and application programming interfaces (APIs). Specifically, a programming model is an abstraction of the underlying computer system that allows for the expression of both algorithms and data structures. In comparison, languages and APIs provide implementations of these abstractions and allow the algorithms and data structures to be put into practice - a programming model exists independently of the choice of both the programming language and the supporting APIs. Programming models are typically focused on achieving increased developer productivity, performance, and portability to other system designs. The rapidly changing nature of processor architectures and the complexity of designing an exascale platform provide significant challenges for these goals. Several other factors are likely to impact the design of future programming models. In particular, the representation and management of increasing levels of parallelism, concurrency and memory hierarchies, combined with the ability to maintain a progressive level of interoperability with today's applications are of significant concern. Overall the design of a programming model is inherently tied not only to the underlying hardware architecture, but also to the requirements of applications and libraries including data analysis, visualization, and uncertainty quantification. Furthermore, the successful implementation of a programming model is dependent on exposed features of the runtime software layers and features of the operating system. Successful use of a programming model also requires effective presentation to the software developer within the context of traditional and new software development tools. Consideration must also be given to the impact of programming models on both languages and the associated compiler infrastructure. Exascale programming models must reflect several, often competing, design goals. These design goals include desirable features such as abstraction and separation of concerns. However, some aspects are unique to large-scale computing. For example, interoperability and composability with existing implementations will prove critical. In particular, performance is the essential underlying goal for large-scale systems. A key evaluation metric for exascale models will be the extent to which they support these goals rather than merely enable them
Improving Usability And Scalability Of Big Data Workflows In The Cloud
Big data workflows have recently emerged as the next generation of data-centric workflow technologies to address the five “V” challenges of big data: volume, variety, velocity, veracity, and value. More formally, a big data workflow is the computerized modeling and automation of a process consisting of a set of computational tasks and their data interdependencies to process and analyze data of ever increasing in scale, complexity, and rate of acquisition. The convergence of big data and workflows creates new challenges in workflow community.
First, the variety of big data results in a need for integrating large number of remote Web services and other heterogeneous task components that can consume and produce data in various formats and models into a uniform and interoperable workflow. Existing approaches fall short in addressing the so-called shimming problem only in an adhoc manner and unable to provide a generic solution. We automatically insert a piece of code called shims or adaptors in order to resolve the data type mismatches.
Second, the volume of big data results in a large number of datasets that needs to be queried and analyzed in an effective and personalized manner. Further, there is also a strong need for sharing, reusing, and repurposing existing tasks and workflows across different users and institutes. To overcome such limitations, we propose a folksonomy- based social workflow recommendation system to improve workflow design productivity and efficient dataset querying and analyzing.
Third, the volume of big data results in the need to process and analyze data of ever increasing in scale, complexity, and rate of acquisition. But a scalable distributed data model is still missing that abstracts and automates data distribution, parallelism, and scalable processing. We propose a NoSQL collectional data model that addresses this limitation.
Finally, the volume of big data combined with the unbound resource leasing capability foreseen in the cloud, facilitates data scientists to wring actionable insights from the data in a time and cost efficient manner. We propose BARENTS scheduler that supports high-performance workflow scheduling in a heterogeneous cloud-computing environment with a single objective to minimize the workflow makespan under a user provided budget constraint
VALIDATION OF DETACHED EDDY SIMULATION USING LESTOOL FOR HOMOGENEOUS TURBULENCE
Detached Eddy Simulation (DES) is a hybrid turbulence model, a modification to the one-equation model proposed by Spalart and Allmaras (1997) [26]. It combines the advantages of both the RANS and LES models to predict any fluid flow. Presently, the focus is on using Homogeneous Turbulence to test the DES model. In an attempt to scrutinize this model, many cases are considered involving the variance of DES grid spacing parameter, CDES, the grid density, Reynolds number and cases with different initial conditions. Choosing Homogeneous Turbulence for our study alienates complications related to the geometry, boundary conditions and other flow characteristics helping us in studying the behavior of the model thoroughly. Also, the interdependencies of the model grid spacing parameter, grid density and the numerical scheme used are also investigated. Many previous implementations of the DES model have taken the value of CDES=0.65. Through this work, many issues including the sensitivity of CDES will be made clear. The code used in running the test cases is called LESTool, developed at University of Kentucky, Lexington. The two main test cases considered are based on the benchmark experimental study by Comte Bellot and Corrsin (1971) [12] and the Direct Numerical Scheme (DNS) simulation by Blaisdell et al. (1991) [10]
- …