3 research outputs found

    MPI vs OpenMP: Un caso de estudio sobre la generaci贸n del conjunto de Mandelbrot

    Get PDF
    Nowadays, some of the most popular tools for parallel programming are Message Passing Interface and Open Multi-Processing. It is of interest to compare these tools in solving the same kind of problems, because of the use of different approaches to inter-task communication. This work attempts to contribute to this goal by running trials in a centralized shared memory architecture in the case of problems with an entirely parallel solution. The selected case study was the parallel computation of Mandelbrot set. Trials were conducted for different iteration limits, processors amount, and C++ implementation variants. The results show better performance in the case of Open Multi-Processing.Algunas de las herramientas m谩s populares hoy en d铆a para la programaci贸n paralela son Interfaz de Paso de Mensajes y Multiprocesamiento Abierto. Es de inter茅s comparar estas herramientas en la resoluci贸n de los mismos tipos de problemas, debido a la utilizaci贸n de diferentes enfoques en la comunicaci贸n entre tareas. Este trabajo tiene como objetivo contribuir a este empe帽o al ejecutar pruebas en una arquitectura de memoria compartida y centralizada en el caso de problemas con una soluci贸n completamente paralela. El caso de estudio seleccionado fue la computaci贸n paralela del conjunto de Mandelbrot. Las pruebas se realizaron para diferentes l铆mites de iteraci贸n, cantidad de procesadores y variantes de implementaci贸n en C++. Los resultados muestran un mejor desempe帽o en el caso de Multiprocesamiento Abierto

    Investigation of Parallel Data Processing Using Hybrid High Performance CPU + GPU Systems and CUDA Streams

    Get PDF
    The paper investigates parallel data processing in a hybrid CPU+GPU(s) system using multiple CUDA streams for overlapping communication and computations. This is crucial for efficient processing of data, in particular incoming data stream processing that would naturally be forwarded using multiple CUDA streams to GPUs. Performance is evaluated for various compute time to host-device communication time ratios, numbers of CUDA streams, for various numbers of threads managing computations on GPUs. Tests also reveal benefits of using CUDA MPS for overlapping communication and computations when using multiple processes. Furthermore, using standard memory allocation on a GPU and Unified Memory versions are compared, the latter including programmer added prefetching. Performance of a hybrid CPU+GPU version as well as scaling across multiple GPUs are demonstrated showing good speed-ups of the approach. Finally, the performance per power consumption of selected configurations are presented for various numbers of streams and various relative performances of GPUs and CPUs

    Parallel programming for modern high performance computing systems

    No full text
    corecore