19,604 research outputs found
A Rate-Distortion Exponent Approach to Multiple Decoding Attempts for Reed-Solomon Codes
Algorithms based on multiple decoding attempts of Reed-Solomon (RS) codes
have recently attracted new attention. Choosing decoding candidates based on
rate-distortion (R-D) theory, as proposed previously by the authors, currently
provides the best performance-versus-complexity trade-off. In this paper, an
analysis based on the rate-distortion exponent (RDE) is used to directly
minimize the exponential decay rate of the error probability. This enables
rigorous bounds on the error probability for finite-length RS codes and leads
to modest performance gains. As a byproduct, a numerical method is derived that
computes the rate-distortion exponent for independent non-identical sources.
Analytical results are given for errors/erasures decoding.Comment: accepted for presentation at 2010 IEEE International Symposium on
Information Theory (ISIT 2010), Austin TX, US
A computationally tractable version of the collective model
A computationally tractable version of the Bohr-Mottelson collective model is
presented which makes it possible to diagonalize realistic collective models
and obtain convergent results in relatively small appropriately chosen
subspaces of the collective model Hilbert space. Special features of the
proposed model is that it makes use of the beta wave functions given
analytically by the softened-beta version of the Wilets-Jean model, proposed by
Elliott et al., and a simple algorithm for computing SO(5) > SO(3) spherical
harmonics. The latter has much in common with the methods of Chacon, Moshinsky,
and Sharp but is conceptually and computationally simpler. Results are
presented for collective models ranging from the sherical vibrator to the
Wilets-Jean and axially symmetric rotor-vibrator models.Comment: 16 pages, 9 figure
Adaptive control in rollforward recovery for extreme scale multigrid
With the increasing number of compute components, failures in future
exa-scale computer systems are expected to become more frequent. This motivates
the study of novel resilience techniques. Here, we extend a recently proposed
algorithm-based recovery method for multigrid iterations by introducing an
adaptive control. After a fault, the healthy part of the system continues the
iterative solution process, while the solution in the faulty domain is
re-constructed by an asynchronous on-line recovery. The computations in both
the faulty and healthy subdomains must be coordinated in a sensitive way, in
particular, both under and over-solving must be avoided. Both of these waste
computational resources and will therefore increase the overall
time-to-solution. To control the local recovery and guarantee an optimal
re-coupling, we introduce a stopping criterion based on a mathematical error
estimator. It involves hierarchical weighted sums of residuals within the
context of uniformly refined meshes and is well-suited in the context of
parallel high-performance computing. The re-coupling process is steered by
local contributions of the error estimator. We propose and compare two criteria
which differ in their weights. Failure scenarios when solving up to
unknowns on more than 245\,766 parallel processes will be
reported on a state-of-the-art peta-scale supercomputer demonstrating the
robustness of the method
On Multiple Decoding Attempts for Reed-Solomon Codes: A Rate-Distortion Approach
One popular approach to soft-decision decoding of Reed-Solomon (RS) codes is
based on using multiple trials of a simple RS decoding algorithm in combination
with erasing or flipping a set of symbols or bits in each trial. This paper
presents a framework based on rate-distortion (RD) theory to analyze these
multiple-decoding algorithms. By defining an appropriate distortion measure
between an error pattern and an erasure pattern, the successful decoding
condition, for a single errors-and-erasures decoding trial, becomes equivalent
to distortion being less than a fixed threshold. Finding the best set of
erasure patterns also turns into a covering problem which can be solved
asymptotically by rate-distortion theory. Thus, the proposed approach can be
used to understand the asymptotic performance-versus-complexity trade-off of
multiple errors-and-erasures decoding of RS codes.
This initial result is also extended a few directions. The rate-distortion
exponent (RDE) is computed to give more precise results for moderate
blocklengths. Multiple trials of algebraic soft-decision (ASD) decoding are
analyzed using this framework. Analytical and numerical computations of the RD
and RDE functions are also presented. Finally, simulation results show that
sets of erasure patterns designed using the proposed methods outperform other
algorithms with the same number of decoding trials.Comment: to appear in the IEEE Transactions on Information Theory (Special
Issue on Facets of Coding Theory: from Algorithms to Networks
Modernizing PHCpack through phcpy
PHCpack is a large software package for solving systems of polynomial
equations. The executable phc is menu driven and file oriented. This paper
describes the development of phcpy, a Python interface to PHCpack. Instead of
navigating through menus, users of phcpy solve systems in the Python shell or
via scripts. Persistent objects replace intermediate files.Comment: Part of the Proceedings of the 6th European Conference on Python in
Science (EuroSciPy 2013), Pierre de Buyl and Nelle Varoquaux editors, (2014
Research and Education in Computational Science and Engineering
Over the past two decades the field of computational science and engineering
(CSE) has penetrated both basic and applied research in academia, industry, and
laboratories to advance discovery, optimize systems, support decision-makers,
and educate the scientific and engineering workforce. Informed by centuries of
theory and experiment, CSE performs computational experiments to answer
questions that neither theory nor experiment alone is equipped to answer. CSE
provides scientists and engineers of all persuasions with algorithmic
inventions and software systems that transcend disciplines and scales. Carried
on a wave of digital technology, CSE brings the power of parallelism to bear on
troves of data. Mathematics-based advanced computing has become a prevalent
means of discovery and innovation in essentially all areas of science,
engineering, technology, and society; and the CSE community is at the core of
this transformation. However, a combination of disruptive
developments---including the architectural complexity of extreme-scale
computing, the data revolution that engulfs the planet, and the specialization
required to follow the applications to new frontiers---is redefining the scope
and reach of the CSE endeavor. This report describes the rapid expansion of CSE
and the challenges to sustaining its bold advances. The report also presents
strategies and directions for CSE research and education for the next decade.Comment: Major revision, to appear in SIAM Revie
Numerical Algebraic Geometry: A New Perspective on String and Gauge Theories
The interplay rich between algebraic geometry and string and gauge theories
has recently been immensely aided by advances in computational algebra.
However, these symbolic (Gr\"{o}bner) methods are severely limited by
algorithmic issues such as exponential space complexity and being highly
sequential. In this paper, we introduce a novel paradigm of numerical algebraic
geometry which in a plethora of situations overcomes these short-comings. Its
so-called 'embarrassing parallelizability' allows us to solve many problems and
extract physical information which elude the symbolic methods. We describe the
method and then use it to solve various problems arising from physics which
could not be otherwise solved.Comment: 36 page
- …