1,347 research outputs found
Hardware Parallelization of Cores Accessing Memory with Irregular Access Patterns
This project studies FPGA-based heterogeneous computing architectures with the objective of
discovering their ability to optimize the performances of algorithms characterized by irregular
memory access patterns. The example used to achieve this is a graph algorithm known as Triad
Census Algorithm, whose implementation has been developed and tested.
First of all, the triad census algorithm is presented, explaining the possible variants and
reviewing the existing implementations upon different architectures. The analysis focuses on
the parallelization techniques which have allowed to boost performance, thus reducing execution
time. Besides, the study tackles the OpenCL programming model, the standard used to develop
the final application. Special attention is paid to the language details that have motivated some
of the most important design decisions.
The dissertation continues with the description of the project implementation, including
the application objectives, the system design, and the different variants developed to enhance
algorithm performance.
Finally, some of the experimental results are presented and discussed. All implemented
versions are evaluated and compared to decide which is the best in terms of scalability and
execution time
Domain-Specific Acceleration and Auto-Parallelization of Legacy Scientific Code in FORTRAN 77 using Source-to-Source Compilation
Massively parallel accelerators such as GPGPUs, manycores and FPGAs represent
a powerful and affordable tool for scientists who look to speed up simulations
of complex systems. However, porting code to such devices requires a detailed
understanding of heterogeneous programming tools and effective strategies for
parallelization. In this paper we present a source to source compilation
approach with whole-program analysis to automatically transform single-threaded
FORTRAN 77 legacy code into OpenCL-accelerated programs with parallelized
kernels.
The main contributions of our work are: (1) whole-source refactoring to allow
any subroutine in the code to be offloaded to an accelerator. (2) Minimization
of the data transfer between the host and the accelerator by eliminating
redundant transfers. (3) Pragmatic auto-parallelization of the code to be
offloaded to the accelerator by identification of parallelizable maps and
reductions.
We have validated the code transformation performance of the compiler on the
NIST FORTRAN 78 test suite and several real-world codes: the Large Eddy
Simulator for Urban Flows, a high-resolution turbulent flow model; the shallow
water component of the ocean model Gmodel; the Linear Baroclinic Model, an
atmospheric climate model and Flexpart-WRF, a particle dispersion simulator.
The automatic parallelization component has been tested on as 2-D Shallow
Water model (2DSW) and on the Large Eddy Simulator for Urban Flows (UFLES) and
produces a complete OpenCL-enabled code base. The fully OpenCL-accelerated
versions of the 2DSW and the UFLES are resp. 9x and 20x faster on GPU than the
original code on CPU, in both cases this is the same performance as manually
ported code.Comment: 12 pages, 5 figures, submitted to "Computers and Fluids" as full
paper from ParCFD conference entr
Hardware and Software Task Scheduling for ARM-FPGA Platforms
ARM-FPGA coupled platforms allow accelerating the computation of specific algorithms by executing them in the FPGA fabric. Several computation steps of our case study for a stereo vision application have been accelerated by hardware implementations. Dynamic Partial Reconfiguration places these hardware tasks in the programmable logic at appropriate times. For an efficient scheduling, it needs to be decided when and where to execute a task. Although there already exist hardware/software scheduling strategies and algorithms, none exploit all possible optimization techniques: re-use, prefetching, parallelization, and pipelining of hardware tasks. The scheduling algorithm proposed in this paper takes this into account and optimizes for the objectives latency/throughput and power/energy
- …