384,565 research outputs found

    Investigation of cluster and cluster queuing system

    Get PDF
    Cluster became main platform as parallel and distributed computing structure for high performance computing. Following the development of high performance computer architecture more and more different branches of natural science benefit fromhuge and efficient computational power. For instance bio-informatics, climate science, computational physics, computational chemistry, marine science, etc. Efficient and reliable computing powermay not only expending demand of existing high performance computing users but also attracting more and more different users. Efficiency and performance are main factors on high performance computing. Most of the high performance computer exists as computer cluster. Computer clustering is the popular and main stream of high-performance computing. Discover the efficiency of high performance computing or cluster is very interesting and never enough as it is really depending on different users. Monitoring and tuning high performance or cluster facilities are always necessary. This project focuses on high performance computer monitoring. Comparing queuing status and work load on different computing nodes on the cluster. As the power consumption is main issue nowadays, our project will also try to estimate power consumption on these special sites and also try to support our way of doing estimation.Master i nettverks- og systemadministrasjo

    High Performance Computing on Cluster and Multicore Architecture

    Get PDF
    High Performance Computing have several issues on architecture, resources, computational model and data. The challenge is establishing the mature architecture with scalable resources. The cluster architecture and multicore architecture implement to produce high performance on computation and process. This research works on architecture development and performance analysis. The cluster architecture build on Raspberry Pi, a single board computer, implement MPICH2. Raspberry Pi cluster build on Raspbian Wheezy operating system and test by metrics computation applications. The multicore architecture build on single computer with Core i5 and Core i7 architecture. The research use himeno98 and himeno16Large tools to analysis the processor and memory allocation. The test run on 1000x1000 matrices and benchmarked with OpenMP. The analysis focuses on CPU Time, FLOPS, and score. The result show on cluster architecture have 2576,07 sec in CPU Time, 86,96 MLPOS, and 2,69 score. The result on Core i5 architecture have 55,57 sec in CPU time, 76,30 MLOPS, and 0,92 score. The result in Core i7 architecture have 59,56 sec CPU Time, 1427,61 MLOPS, and 17,23 score. The cluster and multicore architecture results show that computing process are effected by architecture models. High performance computing architecture that has been built on this result can give learn on the development of HPC architecture models, and baseline performance. In the future it will use for determine the delivery architecture model on HPC and can be test by more variation of load

    Single-Board-Computer Clusters for Cloudlet Computing in Internet of Things

    Get PDF
    The number of connected sensors and devices is expected to increase to billions in the near future. However, centralised cloud-computing data centres present various challenges to meet the requirements inherent to Internet of Things (IoT) workloads, such as low latency, high throughput and bandwidth constraints. Edge computing is becoming the standard computing paradigm for latency-sensitive real-time IoT workloads, since it addresses the aforementioned limitations related to centralised cloud-computing models. Such a paradigm relies on bringing computation close to the source of data, which presents serious operational challenges for large-scale cloud-computing providers. In this work, we present an architecture composed of low-cost Single-Board-Computer clusters near to data sources, and centralised cloud-computing data centres. The proposed cost-efficient model may be employed as an alternative to fog computing to meet real-time IoT workload requirements while keeping scalability. We include an extensive empirical analysis to assess the suitability of single-board-computer clusters as cost-effective edge-computing micro data centres. Additionally, we compare the proposed architecture with traditional cloudlet and cloud architectures, and evaluate them through extensive simulation. We finally show that acquisition costs can be drastically reduced while keeping performance levels in data-intensive IoT use cases.Ministerio de Economía y Competitividad TIN2017-82113-C2-1-RMinisterio de Economía y Competitividad RTI2018-098062-A-I00European Union’s Horizon 2020 No. 754489Science Foundation Ireland grant 13/RC/209

    Air pollution modelling using a graphics processing unit with CUDA

    Get PDF
    The Graphics Processing Unit (GPU) is a powerful tool for parallel computing. In the past years the performance and capabilities of GPUs have increased, and the Compute Unified Device Architecture (CUDA) - a parallel computing architecture - has been developed by NVIDIA to utilize this performance in general purpose computations. Here we show for the first time a possible application of GPU for environmental studies serving as a basement for decision making strategies. A stochastic Lagrangian particle model has been developed on CUDA to estimate the transport and the transformation of the radionuclides from a single point source during an accidental release. Our results show that parallel implementation achieves typical acceleration values in the order of 80-120 times compared to CPU using a single-threaded implementation on a 2.33 GHz desktop computer. Only very small differences have been found between the results obtained from GPU and CPU simulations, which are comparable with the effect of stochastic transport phenomena in atmosphere. The relatively high speedup with no additional costs to maintain this parallel architecture could result in a wide usage of GPU for diversified environmental applications in the near future.Comment: 5 figure

    Archer: A Community Distributed Computing Infrastructure for Computer Architecture Research and Education

    Full text link
    This paper introduces Archer, a community-based computing resource for computer architecture research and education. The Archer infrastructure integrates virtualization and batch scheduling middleware to deliver high-throughput computing resources aggregated from resources distributed across wide-area networks and owned by different participating entities in a seamless manner. The paper discusses the motivations leading to the design of Archer, describes its core middleware components, and presents an analysis of the functionality and performance of a prototype wide-area deployment running a representative computer architecture simulation workload.Comment: 11 pages, 2 figures. Describes the Archer project, http://archer-project.or

    Evaluating Performance of Components

    Get PDF
    Parallel Component Performance Benchmarks is a computer program developed to aid the evaluation of the Common Component Architecture (CCA) - a software architecture, based on a component model, that was conceived to foster high-performance computing, including parallel computing. More specifically, this program compares the performances (principally by measuring computing times) of componentized versus conventional versions of the Parallel Pyramid 2D Adaptive Mesh Refinement library - a software library that is used to generate computational meshes for solving physical problems and that is typical of software libraries in use at NASA s Jet Propulsion Laboratory
    corecore