185 research outputs found

    Autonomic Execution of Computational Workflows

    Get PDF
    This paper describes the application of anautonomic paradigm to manage the complexity of softwaresystems such as computational workflows. To demonstrate ourapproach, the workflow and the services comprising it aretreated as managed resources controlled by hierarchicallyorganized autonomic managers. By applying service-orientedsoftware engineering principles, in particular enterpriseintegration patterns, we have developed a scalable, agile, selfhealingenvironment for execution of dynamic, data-drivenworkflows which are capable of assuring scientific fidelitydespite unavoidable faults and without human intervention

    An Integrated Software Development Model for Heterogeneous High Performance Computing

    Get PDF
    The last few decades have seen an impressive developments in every aspect of parallel computing technology; viz. processing and storage technology, interconnect technology and software technology. Although these systems incorporate large amount of computing power, they are not general enough to efficiently support today's computation-intensive problems (e.g. the Grand Challenges), that warrant multiple computational models and levels of parallelism. We believe that the future of parallel computing lies in the integration of the plethora of "specialized" architectures into a single Heterogeneous High Performance Computing (HHPC) environment that allows them to cooperate in solving complex problems. Software development in any Parallel/Distributed environment is a non-trivial process and requires a thorough understanding of the application and the architecture. This problem further intensifies as systems evolve into HHPC environments. The objective of this paper is to formalize the softw..

    The Gateway System: Uniform Web Based Access to Remote Resources

    Get PDF
    Exploiting our experience developing the WebFlow system, we designed the Gateway system to provide seamless and secure access to computational resources at ASC MSRC. The Gateway follows our commodity components strategy, and it is implemented as a modern three-tier system. Tier 1 is a high-level front-end for visual programming, steering, run-time data analysis and visualization, built on top of the Web and OO commodity standards. Distributed object-based, scalable, and reusable Web server and Object broker middleware forms Tier 2. Back-end services comprise Tier 3. In particular, access to high performance computational resources is provided by implementing the emerging standard for metacomputing API

    Building Distributed Systems for the Pragmatic Object Web

    Get PDF
    We review the growing power and capability of commodity computing and communication technologies largely driven by commercial distributed information systems. These systems are built from CORBA, Microsoft\u27s COM, JavaBeans, and rapidly advancing Web approaches. One can abstract these to a three-tier model with largely independent clients connected to a distributed network of servers. The latter host various services including object and relational databases and of course parallel and sequential computing. High performance can be obtained by combining concurrency at the middle server tier with optimized parallel back end services. The resultant system combines the needed performance for large-scale HPCC applications with the rich functionality of commodity systems. Further the architecture with distinct interface, server and specialized service implementation layers, naturally allows advances in each area to be easily incorporated. We illustrate how performance can be obtained within a commodity architecture and we propose a middleware integration approach based on JWORB (Java Web Object Broker) multi-protocol server technology. We illustrate our approach on a set of prototype applications in areas such as collaborative systems, support of multidisciplinary interactions, WebFlow based visual metacomputing, WebFlow over Globus, Quantum Monte Carlo and distributed interactive simulations

    Which Applications Can Use High Performance Fortran and Fortran-D: Industry Standard Data Parallel Languages?

    Get PDF
    In this paper, we present the first, preliminary results of HPF/Fortran-D language analysis based on compiling and running benchmark applications using a prototype implementation of HPF/Fortran-D compiler. The analysis indicate that the HPF is a very convenient tool for programming many applications on massively parallel and/or distributed systems. In addition, we cumulate experience on how to parallelize irregular problems to extend the scope of Fortran-D beyond HPF and suggest future extensions to the Fortran standard

    Interpreting the Performance of HPF/Fortran 90D

    Get PDF
    In this paper we present a novel interpretive approach for accurate and cost-effective performance prediction in a high performance computing environment, and describe the design of a source-driven HPF/Fortran 90D performance prediction framework based on this approach. The performance prediction framework has been implemented as part of a HPF/Fortran 90D application development environment. A set of benchmarking kernels and application codes are used to validate the accuracy, utility, usability, and cost-effectiveness of the performance prediction framework. The use of the framework for selecting appropriate compiler directives and for application performance debugging is demonstrated. Keywords: Performance prediction, HPF/Fortran 90D application development, System & Application characterization

    Compiling Fortran 90D/HPF for distributed memory MIMD computers

    Get PDF
    This paper describes the design of the Fortran90D/HPF compiler, a source-to-source parallel compiler for distributed memory systems being developed at Syracuse University. Fortran 90D/HPF is a data parallel language with special directives to specify data alignment and distributions. A systematic methodology to process distribution directives of Fortran 90D/HPF is presented. Furthermore, techniques for data and computation partitioning, communication detection and generation, and the run-time support for the compiler are discussed. Finally, initial performance results for the compiler are presented. We believe that the methodology to process data distribution, computation partitioning, communication system design and the overall compiler design can be used by the implementors of compilers for HPF

    A Study of Software Development for High Performance Computing

    Get PDF
    Software development in a High Performance Computing (HPC) environment is non-trivial and requires a thorough understanding of the application and the architecture. The objective of this paper is to study the software development process in a high performance computing environment and to outline the stages typically encountered in this process. Support required at each stage is also highlighted. The modeling of stock option pricing is used as a running example in the study

    Facile detection of proteins on a solid-phase membrane by direct binding of dextran-based luminol?biotin chemiluminescent polymer

    Get PDF
    Facile and non-radioactive methods are desired for the sensitive detection and quantification of various proteins. Herein we describe a novel chemiluminescence (CL)-detection method of particular proteins based on direct binding of a dextran-luminol-biotin (DLB) CL polymer to the proteins on a poly(vinylidene difluoride) membrane. Among 32 kinds of the proteins screened, several proteins such as drug-metabolizing enzymes, cytochrome p450 (CYP)1A2, CYP2E1, and CYP3A4 had the ability to bind directly to the DLB polymer. The binding site in the polymer was owing to the framework of the modified dextran, which underwent oxidation and reduction procedures. This interaction might be the comprehensive effect of both electrostatic interaction and steric complementarities. CL intensity of the proteins detected by the polymer could be further enlarged by the mediation of avidin. The proposed CL-imaging method possesses potential as a rapid, facile, inexpensive and selective detection of the proteins
    corecore