304,687 research outputs found

    PISCES: An environment for parallel scientific computation

    Get PDF
    The parallel implementation of scientific computing environment (PISCES) is a project to provide high-level programming environments for parallel MIMD computers. Pisces 1, the first of these environments, is a FORTRAN 77 based environment which runs under the UNIX operating system. The Pisces 1 user programs in Pisces FORTRAN, an extension of FORTRAN 77 for parallel processing. The major emphasis in the Pisces 1 design is in providing a carefully specified virtual machine that defines the run-time environment within which Pisces FORTRAN programs are executed. Each implementation then provides the same virtual machine, regardless of differences in the underlying architecture. The design is intended to be portable to a variety of architectures. Currently Pisces 1 is implemented on a network of Apollo workstations and on a DEC VAX uniprocessor via simulation of the task level parallelism. An implementation for the Flexible Computing Corp. FLEX/32 is under construction. An introduction to the Pisces 1 virtual computer and the FORTRAN 77 extensions is presented. An example of an algorithm for the iterative solution of a system of equations is given. The most notable features of the design are the provision for several granularities of parallelism in programs and the provision of a window mechanism for distributed access to large arrays of data

    Parallel and Distributed Computing for High-Performance Applications

    Get PDF
    The study of parallel and distributed computing has become an important area in computer science because it makes it possible to create high-performance software that can effectively handle challenging computational tasks. In terms of their use in the world of high-performance applications, parallel and distributed computing techniques are given a thorough introduction in this study. The partitioning of computational processes into smaller subtasks that may be completed concurrently on numerous processors or computers is the core idea underpinning parallel and distributed computing. This strategy enables quicker execution times and enhanced performance in general. Parallel and distributed computing are essential for high-performance applications like scientific simulations, data analysis, and artificial intelligence since they frequently call for significant computational resources. High-performance apps are able to effectively handle computationally demanding tasks thanks in large part to parallel and distributed computing. This article offers a thorough review of the theories, methods, difficulties, and developments in parallel and distributed computing for high-performance applications. Researchers and practitioners may fully utilize the potential of parallel and distributed computing to open up new vistas in computational science and engineering by comprehending the underlying concepts and utilizing the most recent breakthroughs

    Towards Intelligent Runtime Framework for Distributed Heterogeneous Systems

    Get PDF
    Scientific applications strive for increased memory and computing performance, requiring massive amounts of data and time to produce results. Applications utilize large-scale, parallel computing platforms with advanced architectures to accommodate their needs. However, developing performance-portable applications for modern, heterogeneous platforms requires lots of effort and expertise in both the application and systems domains. This is more relevant for unstructured applications whose workflow is not statically predictable due to their heavily data-dependent nature. One possible solution for this problem is the introduction of an intelligent Domain-Specific Language (iDSL) that transparently helps to maintain correctness, hides the idiosyncrasies of lowlevel hardware, and scales applications. An iDSL includes domain-specific language constructs, a compilation toolchain, and a runtime providing task scheduling, data placement, and workload balancing across and within heterogeneous nodes. In this work, we focus on the runtime framework. We introduce a novel design and extension of a runtime framework, the Parallel Runtime Environment for Multicore Applications. In response to the ever-increasing intra/inter-node concurrency, the runtime system supports efficient task scheduling and workload balancing at both levels while allowing the development of custom policies. Moreover, the new framework provides abstractions supporting the utilization of heterogeneous distributed nodes consisting of CPUs and GPUs and is extensible to other devices. We demonstrate that by utilizing this work, an application (or the iDSL) can scale its performance on heterogeneous exascale-era supercomputers with minimal effort. A future goal for this framework (out of the scope of this thesis) is to be integrated with machine learning to improve its decision-making and performance further. As a bridge to this goal, since the framework is under development, we experiment with data from Nuclear Physics Particle Accelerators and demonstrate the significant improvements achieved by utilizing machine learning in the hit-based track reconstruction process

    USE OF HIGH PERFORMANCE COMPUTING TECHNOLOGIES AND SCRIPT RUN MEDIATOR MIDDLEWARE FOR EDUCATIONAL PROCESS IN LIEPAJA UNIVERSITY

    Get PDF
    Abstract. In this paper the authors describe a solution for implementation of the SRM (ScriptRunMediatorCSP) middleware to resolve trusting and secure connection problems to HPC (High Performance Computing) cluster resources in the Liepaja University. The main idea is to develop a specific middleware tool called SRM, which would allow connection to the parallel computing cluster system, namely to the main server masternode, to provide the easiest solution for using HPC technologies in educational processes. Cluster computing resources could be used for solution and analysis of various scientific problems applying mathematical methods and high performance technological solutions for learning and performing parallel computations in the educational process thus increasing the students' theoretical knowledge in practical applications. Keywords: cluster, HPC technologies, Lazarus, SRM middleware, parallel calculations, parallel compilers. Introduction Rapid development of High Performance Computing (HPC) systems in the field of IT is determined by the necessity to solve society and science problems more effectively, for instance, to optimize gas and oil research and acquisition, to forecast climate changes, and also to introduce innovation products and services in industrial and financial fields. Therefore, it is important to provide the HPC technological environment, which is implemented with the help of parallel data processing and computation technologies, applications of cluster networks and information technologies, in order to promote involvement of students as young experts to real problem solution in the field of scientific research. In the academic and research field, HPC utilizes multiple processors to perform concurrent tasks in parallel for solving large cutting-edge research computation problems. At this moment authors provide centralized high-performance computing resources and support to Liepaja University students and researchers in large-scale computing disciplines with the use of advanced computing cluster hardware infrastructure, software, tools and programming techniques.The goal of centralized supercomputing centres and conventional distributed computing systems, such as grids and clusters, and their support service is to enhance research capabilities and effectiveness. The current efforts in the grid computing research focus on a design of the new grid schedulers, that can efficiently optimize the standard scheduling objectives, such as makespan, flowtime and resource utilization, but also can fulfil the security requirements of the grid users and can minimize the energy consumed by all of the system components. These schedulers should also capture the complexity of the whole system and provide meaningful measures for a wide range of grid applications and services. Therefore, energy-efficient and security-aware scheduling in CGs (Computational Grids) becomes a complex research and engineering endeavour mainly due to different priorities and preferences of the grid users and resource owners Several parallel computation clusters with various configurations have been created and tested, where the special middleware SRM (ScriptRunMediatorCSP (ClusterSecurityPackage)) was used in order to implement solutions of direct, indirect and optimisation problems using data parallel processing technologies -computation automation software. Within the research a Beowulf cluster is used, and the operating system adopted is Linux, which is widely available; software tools for message passing such as MP

    Developing High Performance Computing Resources for Teaching Cluster and Grid Computing courses

    Get PDF
    High-Performance Computing (HPC) and the ability to process large amounts of data are of paramount importance for UK business and economy as outlined by Rt Hon David Willetts MP at the HPC and Big Data conference in February 2014. However there is a shortage of skills and available training in HPC to prepare and expand the workforce for the HPC and Big Data research and development. Currently, HPC skills are acquired mainly by students and staff taking part in HPC-related research projects, MSc courses, and at the dedicated training centres such as Edinburgh University’s EPCC. There are few UK universities teaching the HPC, Clusters and Grid Computing courses at the undergraduate level. To address the issue of skills shortages in the HPC it is essential to provide teaching and training as part of both postgraduate and undergraduate courses. The design and development of such courses is challenging since the technologies and software in the fields of large scale distributed systems such as Cluster, Cloud and Grid computing are undergoing continuous change. The students completing the HPC courses should be proficient in these evolving technologies and equipped with practical and theoretical skills for future jobs in this fast developing area. In this paper we present our experience in developing the HPC, Cluster and Grid modules including a review of existing HPC courses offered at the UK universities. The topics covered in the modules are described, as well as the coursework projects based on practical laboratory work. We conclude with an evaluation based on our experience over the last ten years in developing and delivering the HPC modules on the undergraduate courses, with suggestions for future work

    Performance Reproduction and Prediction of Selected Dynamic Loop Scheduling Experiments

    Full text link
    Scientific applications are complex, large, and often exhibit irregular and stochastic behavior. The use of efficient loop scheduling techniques in computationally-intensive applications is crucial for improving their performance on high-performance computing (HPC) platforms. A number of dynamic loop scheduling (DLS) techniques have been proposed between the late 1980s and early 2000s, and efficiently used in scientific applications. In most cases, the computing systems on which they have been tested and validated are no longer available. This work is concerned with the minimization of the sources of uncertainty in the implementation of DLS techniques to avoid unnecessary influences on the performance of scientific applications. Therefore, it is important to ensure that the DLS techniques employed in scientific applications today adhere to their original design goals and specifications. The goal of this work is to attain and increase the trust in the implementation of DLS techniques in present studies. To achieve this goal, the performance of a selection of scheduling experiments from the 1992 original work that introduced factoring is reproduced and predicted via both, simulative and native experimentation. The experiments show that the simulation reproduces the performance achieved on the past computing platform and accurately predicts the performance achieved on the present computing platform. The performance reproduction and prediction confirm that the present implementation of the DLS techniques considered both, in simulation and natively, adheres to their original description. The results confirm the hypothesis that reproducing experiments of identical scheduling scenarios on past and modern hardware leads to an entirely different behavior from expected
    • …
    corecore