109,018 research outputs found
Statistical mechanics of typical set decoding
The performance of ``typical set (pairs) decoding'' for ensembles of
Gallager's linear code is investigated using statistical physics. In this
decoding, error happens when the information transmission is corrupted by an
untypical noise or two or more typical sequences satisfy the parity check
equation provided by the received codeword for which a typical noise is added.
We show that the average error rate for the latter case over a given code
ensemble can be tightly evaluated using the replica method, including the
sensitivity to the message length. Our approach generally improves the existing
analysis known in information theory community, which was reintroduced by
MacKay (1999) and believed as most accurate to date.Comment: 7 page
Statistical Pruning for Near-Maximum Likelihood Decoding
In many communications problems, maximum-likelihood (ML) decoding reduces to finding the closest (skewed) lattice point in N-dimensions to a given point xisin CN. In its full generality, this problem is known to be NP-complete. Recently, the expected complexity of the sphere decoder, a particular algorithm that solves the ML problem exactly, has been computed. An asymptotic analysis of this complexity has also been done where it is shown that the required computations grow exponentially in N for any fixed SNR. At the same time, numerical computations of the expected complexity show that there are certain ranges of rates, SNRs and dimensions N for which the expected computation (counted as the number of scalar multiplications) involves no more than N3 computations. However, when the dimension of the problem grows too large, the required computations become prohibitively large, as expected from the asymptotic exponential complexity. In this paper, we propose an algorithm that, for large N, offers substantial computational savings over the sphere decoder, while maintaining performance arbitrarily close to ML. We statistically prune the search space to a subset that, with high probability, contains the optimal solution, thereby reducing the complexity of the search. Bounds on the error performance of the new method are proposed. The complexity of the new algorithm is analyzed through an upper bound. The asymptotic behavior of the upper bound for large N is also analyzed which shows that the upper bound is also exponential but much lower than the sphere decoder. Simulation results show that the algorithm is much more efficient than the original sphere decoder for smaller dimensions as well, and does not sacrifice much in terms of performance
Critical Noise Levels for LDPC decoding
We determine the critical noise level for decoding low density parity check
error correcting codes based on the magnetization enumerator (\cM), rather
than on the weight enumerator (\cW) employed in the information theory
literature. The interpretation of our method is appealingly simple, and the
relation between the different decoding schemes such as typical pairs decoding,
MAP, and finite temperature decoding (MPM) becomes clear. In addition, our
analysis provides an explanation for the difference in performance between MN
and Gallager codes. Our results are more optimistic than those derived via the
methods of information theory and are in excellent agreement with recent
results from another statistical physics approach.Comment: 9 pages, 5 figure
LEDAkem: a post-quantum key encapsulation mechanism based on QC-LDPC codes
This work presents a new code-based key encapsulation mechanism (KEM) called
LEDAkem. It is built on the Niederreiter cryptosystem and relies on
quasi-cyclic low-density parity-check codes as secret codes, providing high
decoding speeds and compact keypairs. LEDAkem uses ephemeral keys to foil known
statistical attacks, and takes advantage of a new decoding algorithm that
provides faster decoding than the classical bit-flipping decoder commonly
adopted in this kind of systems. The main attacks against LEDAkem are
investigated, taking into account quantum speedups. Some instances of LEDAkem
are designed to achieve different security levels against classical and quantum
computers. Some performance figures obtained through an efficient C99
implementation of LEDAkem are provided.Comment: 21 pages, 3 table
Statistical framework for video decoding complexity modeling and prediction
Video decoding complexity modeling and prediction is an increasingly important issue for efficient resource utilization in a variety of applications, including task scheduling, receiver-driven complexity shaping, and adaptive dynamic voltage scaling. In this paper we present a novel view of this problem based on a statistical framework perspective. We explore the statistical structure (clustering) of the execution time required by each video decoder module (entropy decoding, motion compensation, etc.) in conjunction with complexity features that are easily extractable at encoding time (representing the properties of each module's input source data). For this purpose, we employ Gaussian mixture models (GMMs) and an expectation-maximization algorithm to estimate the joint execution-time - feature probability density function (PDF). A training set of typical video sequences is used for this purpose in an offline estimation process. The obtained GMM representation is used in conjunction with the complexity features of new video sequences to predict the execution time required for the decoding of these sequences. Several prediction approaches are discussed and compared. The potential mismatch between the training set and new video content is addressed by adaptive online joint-PDF re-estimation. An experimental comparison is performed to evaluate the different approaches and compare the proposed prediction scheme with related resource prediction schemes from the literature. The usefulness of the proposed complexity-prediction approaches is demonstrated in an application of rate-distortion-complexity optimized decoding
Belief propagation decoding of quantum channels by passing quantum messages
Belief propagation is a powerful tool in statistical physics, machine
learning, and modern coding theory. As a decoding method, it is ubiquitous in
classical error correction and has also been applied to stabilizer-based
quantum error correction. The algorithm works by passing messages between nodes
of the factor graph associated with the code and enables efficient decoding, in
some cases even up to the Shannon capacity of the channel. Here we construct a
belief propagation algorithm which passes quantum messages on the factor graph
and is capable of decoding the classical-quantum channel with pure state
outputs. This gives explicit decoding circuits whose number of gates is
quadratic in the blocklength of the code. We also show that this decoder can be
modified to work with polar codes for the pure state channel and as part of a
polar decoder for transmitting quantum information over the amplitude damping
channel. These represent the first explicit capacity-achieving decoders for
non-Pauli channels.Comment: v3: final version for publication; v2: improved discussion of the
algorithm; 7 pages & 2 figures. v1: 6 pages, 1 figur
- …