213,330 research outputs found
Efficiency analysis methodology of FPGAs based on lost frequencies, area and cycles
We propose a methodology to study and to quantify efficiency and the impact of overheads on runtime performance. Most work on High-Performance Computing (HPC) for FPGAs only studies runtime performance or cost, while we are interested in how far we are from peak performance and, more importantly, why. The efficiency of runtime performance is defined with respect to the ideal computational runtime in absence of inefficiencies. The analysis of the difference between actual and ideal runtime reveals the overheads and bottlenecks. A formal approach is proposed to decompose the efficiency into three components: frequency, area and cycles. After quantification of the efficiencies, a detailed analysis has to reveal the reasons for the lost frequencies, lost area and lost cycles. We propose a taxonomy of possible causes and practical methods to identify and quantify the overheads. The proposed methodology is applied on a number of use cases to illustrate the methodology. We show the interaction between the three components of efficiency and show how bottlenecks are revealed
Reducing branch delay to zero in pipelined processors
A mechanism to reduce the cost of branches in pipelined processors is described and evaluated. It is based on the use of multiple prefetch, early computation of the target address, delayed branch, and parallel execution of branches. The implementation of this mechanism using a branch target instruction memory is described. An analytical model of the performance of this implementation makes it possible to measure the efficiency of the mechanism with a very low computational cost. The model is used to determine the size of cache lines that maximizes the processor performance, to compare the performance of the mechanism with that of other schemes, and to analyze the performance of the mechanism with two alternative cache organizations.Peer ReviewedPostprint (published version
Adaptive control in rollforward recovery for extreme scale multigrid
With the increasing number of compute components, failures in future
exa-scale computer systems are expected to become more frequent. This motivates
the study of novel resilience techniques. Here, we extend a recently proposed
algorithm-based recovery method for multigrid iterations by introducing an
adaptive control. After a fault, the healthy part of the system continues the
iterative solution process, while the solution in the faulty domain is
re-constructed by an asynchronous on-line recovery. The computations in both
the faulty and healthy subdomains must be coordinated in a sensitive way, in
particular, both under and over-solving must be avoided. Both of these waste
computational resources and will therefore increase the overall
time-to-solution. To control the local recovery and guarantee an optimal
re-coupling, we introduce a stopping criterion based on a mathematical error
estimator. It involves hierarchical weighted sums of residuals within the
context of uniformly refined meshes and is well-suited in the context of
parallel high-performance computing. The re-coupling process is steered by
local contributions of the error estimator. We propose and compare two criteria
which differ in their weights. Failure scenarios when solving up to
unknowns on more than 245\,766 parallel processes will be
reported on a state-of-the-art peta-scale supercomputer demonstrating the
robustness of the method
Improving optimal control of grid-connected lithium-ion batteries through more accurate battery and degradation modelling
The increased deployment of intermittent renewable energy generators opens up
opportunities for grid-connected energy storage. Batteries offer significant
flexibility but are relatively expensive at present. Battery lifetime is a key
factor in the business case, and it depends on usage, but most techno-economic
analyses do not account for this. For the first time, this paper quantifies the
annual benefits of grid-connected batteries including realistic physical
dynamics and nonlinear electrochemical degradation. Three lithium-ion battery
models of increasing realism are formulated, and the predicted degradation of
each is compared with a large-scale experimental degradation data set
(Mat4Bat). A respective improvement in RMS capacity prediction error from 11\%
to 5\% is found by increasing the model accuracy. The three models are then
used within an optimal control algorithm to perform price arbitrage over one
year, including degradation. Results show that the revenue can be increased
substantially while degradation can be reduced by using more realistic models.
The estimated best case profit using a sophisticated model is a 175%
improvement compared with the simplest model. This illustrates that using a
simplistic battery model in a techno-economic assessment of grid-connected
batteries might substantially underestimate the business case and lead to
erroneous conclusions
- …