41 research outputs found
An Efficient Algorithm for Counting Cycles in QC and APM LDPC Codes
In this paper, a new method is given for counting cycles in the Tanner graph
of a (Type-I) quasi-cyclic (QC) low-density parity-check (LDPC) code which the
complexity mainly is dependent on the base matrix, independent from the
CPM-size of the constructed code. Interestingly, for large CPM-sizes, in
comparison of the existing methods, this algorithm is the first approach which
efficiently counts the cycles in the Tanner graphs of QC-LDPC codes. In fact,
the algorithm recursively counts the cycles in the parity-check matrix
column-by-column by finding all non-isomorph tailless backtrackless closed
(TBC) walks in the base graph and enumerating theoretically their corresponding
cycles in the same equivalent class. Moreover, this approach can be modified in
few steps to find the cycle distributions of a class of LDPC codes based on
Affine permutation matrices (APM-LDPC codes). Interestingly, unlike the
existing methods which count the cycles up to , where is the girth,
the proposed algorithm can be used to enumerate the cycles of arbitrary length
in the Tanner graph. Moreover, the proposed cycle searching algorithm improves
upon various previously known methods, in terms of computational complexity and
memory requirements.Comment: 18 pages, 4 figure
Minimum Pseudoweight Analysis of 3-Dimensional Turbo Codes
In this work, we consider pseudocodewords of (relaxed) linear programming
(LP) decoding of 3-dimensional turbo codes (3D-TCs). We present a relaxed LP
decoder for 3D-TCs, adapting the relaxed LP decoder for conventional turbo
codes proposed by Feldman in his thesis. We show that the 3D-TC polytope is
proper and -symmetric, and make a connection to finite graph covers of the
3D-TC factor graph. This connection is used to show that the support set of any
pseudocodeword is a stopping set of iterative decoding of 3D-TCs using maximum
a posteriori constituent decoders on the binary erasure channel. Furthermore,
we compute ensemble-average pseudoweight enumerators of 3D-TCs and perform a
finite-length minimum pseudoweight analysis for small cover degrees. Also, an
explicit description of the fundamental cone of the 3D-TC polytope is given.
Finally, we present an extensive numerical study of small-to-medium block
length 3D-TCs, which shows that 1) typically (i.e., in most cases) when the
minimum distance and/or the stopping distance is
high, the minimum pseudoweight (on the additive white Gaussian noise channel)
is strictly smaller than both the and the , and 2)
the minimum pseudoweight grows with the block length, at least for
small-to-medium block lengths.Comment: To appear in IEEE Transactions on Communication
Diagnosis of weaknesses in modern error correction codes: a physics approach
One of the main obstacles to the wider use of the modern error-correction
codes is that, due to the complex behavior of their decoding algorithms, no
systematic method which would allow characterization of the Bit-Error-Rate
(BER) is known. This is especially true at the weak noise where many systems
operate and where coding performance is difficult to estimate because of the
diminishingly small number of errors. We show how the instanton method of
physics allows one to solve the problem of BER analysis in the weak noise range
by recasting it as a computationally tractable minimization problem.Comment: 9 pages, 8 figure
A STUDY OF LINEAR ERROR CORRECTING CODES
Since Shannon's ground-breaking work in 1948, there have been two main development streams
of channel coding in approaching the limit of communication channels, namely classical coding
theory which aims at designing codes with large minimum Hamming distance and probabilistic
coding which places the emphasis on low complexity probabilistic decoding using long codes built
from simple constituent codes. This work presents some further investigations in these two channel
coding development streams.
Low-density parity-check (LDPC) codes form a class of capacity-approaching codes with sparse
parity-check matrix and low-complexity decoder Two novel methods of constructing algebraic binary
LDPC codes are presented. These methods are based on the theory of cyclotomic cosets, idempotents
and Mattson-Solomon polynomials, and are complementary to each other. The two methods
generate in addition to some new cyclic iteratively decodable codes, the well-known Euclidean and
projective geometry codes. Their extension to non binary fields is shown to be straightforward.
These algebraic cyclic LDPC codes, for short block lengths, converge considerably well under iterative
decoding. It is also shown that for some of these codes, maximum likelihood performance may
be achieved by a modified belief propagation decoder which uses a different subset of 7^ codewords
of the dual code for each iteration.
Following a property of the revolving-door combination generator, multi-threaded minimum
Hamming distance computation algorithms are developed. Using these algorithms, the previously
unknown, minimum Hamming distance of the quadratic residue code for prime 199 has been evaluated.
In addition, the highest minimum Hamming distance attainable by all binary cyclic codes
of odd lengths from 129 to 189 has been determined, and as many as 901 new binary linear codes
which have higher minimum Hamming distance than the previously considered best known linear
code have been found.
It is shown that by exploiting the structure of circulant matrices, the number of codewords
required, to compute the minimum Hamming distance and the number of codewords of a given
Hamming weight of binary double-circulant codes based on primes, may be reduced. A means
of independently verifying the exhaustively computed number of codewords of a given Hamming
weight of these double-circulant codes is developed and in coiyunction with this, it is proved that
some published results are incorrect and the correct weight spectra are presented. Moreover, it is
shown that it is possible to estimate the minimum Hamming distance of this family of prime-based
double-circulant codes.
It is shown that linear codes may be efficiently decoded using the incremental correlation Dorsch
algorithm. By extending this algorithm, a list decoder is derived and a novel, CRC-less error detection
mechanism that offers much better throughput and performance than the conventional ORG
scheme is described. Using the same method it is shown that the performance of conventional CRC
scheme may be considerably enhanced. Error detection is an integral part of an incremental redundancy
communications system and it is shown that sequences of good error correction codes,
suitable for use in incremental redundancy communications systems may be obtained using the
Constructions X and XX. Examples are given and their performances presented in comparison to
conventional CRC schemes
On Lowering the Error Floor of Short-to-Medium Block Length Irregular Low Density Parity Check Codes
Edited version embargoed until 22.03.2019
Full version: Access restricted permanently due to 3rd party copyright restrictions. Restriction set on 22.03.2018 by SE, Doctoral CollegeGallager proposed and developed low density parity check (LDPC) codes in the early 1960s. LDPC codes were rediscovered in the early 1990s and shown to be capacity approaching over the additive white Gaussian noise (AWGN) channel. Subsequently, density evolution (DE) optimized symbol node degree distributions were used to significantly improve the decoding performance of short to medium length irregular LDPC codes. Currently, the short to medium length LDPC codes with the lowest error floor are DE optimized irregular LDPC codes constructed using progressive edge growth (PEG) algorithm modifications which are designed to increase the approximate cycle extrinsic message degrees (ACE) in the LDPC code graphs constructed.
The aim of the present work is to find efficient means to improve on the error floor performance published for short to medium length irregular LDPC codes over AWGN channels in the literature. An efficient algorithm for determining the girth and ACE distributions in short to medium length LDPC code Tanner graphs has been proposed. A cyclic PEG (CPEG) algorithm which uses an edge connections sequence that results in LDPC codes with improved girth and ACE distributions is presented. LDPC codes with DE optimized/’good’ degree distributions which have larger minimum distances and stopping distances than previously published for LDPC codes of similar length and rate have been found. It is shown that increasing the minimum distance of LDPC codes lowers their error floor performance over AWGN channels; however, there are threshold minimum distances values above which there is no further lowering of the error floor performance. A minimum local girth (edge skipping) (MLG (ES)) PEG algorithm is presented; the algorithm controls the minimum local girth (global girth) connected in the Tanner graphs of LDPC codes constructed by forfeiting some edge connections. A technique for constructing optimal low correlated edge density (OED) LDPC codes based on modified DE optimized symbol node degree distributions and the MLG (ES) PEG algorithm modification is presented. OED rate-½ (n, k)=(512, 256) LDPC codes have been shown to have lower error floor over the AWGN channel than previously published for LDPC codes of similar length and rate. Similarly, consequent to an improved symbol node degree distribution, rate ½ (n, k)=(1024, 512) LDPC codes have been shown to have lower error floor over the AWGN channel than previously published for LDPC codes of similar length and rate.
An improved BP/SPA (IBP/SPA) decoder, obtained by making two simple modifications to the standard BP/SPA decoder, has been shown to result in an unprecedented generalized improvement in the performance of short to medium length irregular LDPC codes under iterative message passing decoding. The superiority of the Slepian Wolf distributed source coding model over other distributed source coding models based on LDPC codes has been shown
Pseudocodewords of Parity-Check Codes
The success of modern algorithms for the decoding problem such as message-passing iterative decoding and linear programming decoding lies in their local nature. This feature allows the algorithms to be extremely fast and capable of correcting more errors than guaranteed by the classical minimum distance of the code. Nonetheless, the performance of these decoders depends crucially on the Tanner graph representation of the code. In order to understand this choice of representation, we need to analyze the pseudocodewords of the Tanner graph of a code. These pseudocodewords are outputs of local decoding algorithms which may not be legitimate codewords. In this dissertation, we introduce a lifted fundamental cone and show that there is a one-to-one correspondence between graph cover pseudocodewords of a binary code and integer points in the lifted fundamental cone. We use this fact to prove the rationality of the generating function of the pseudocodewords for a general binary parity-check code. Our approach also yields algorithms for producing this generating function and provides tools for studying the irreducible pseudocodewords. Understanding irreducible pseudocodewords is crucial to determining the best representation of a code. Moreover, combining these techniques with the recent characterization of fundamental cone over F_3, we can analyze ternary parity-check codes. Finally, we make progress in the study of more general nonbinary codes by determining constraints satisfied by all pseudocodewords of a code over F_p where p is prime
Tree-Structure Expectation Propagation for LDPC Decoding over the BEC
We present the tree-structure expectation propagation (Tree-EP) algorithm to
decode low-density parity-check (LDPC) codes over discrete memoryless channels
(DMCs). EP generalizes belief propagation (BP) in two ways. First, it can be
used with any exponential family distribution over the cliques in the graph.
Second, it can impose additional constraints on the marginal distributions. We
use this second property to impose pair-wise marginal constraints over pairs of
variables connected to a check node of the LDPC code's Tanner graph. Thanks to
these additional constraints, the Tree-EP marginal estimates for each variable
in the graph are more accurate than those provided by BP. We also reformulate
the Tree-EP algorithm for the binary erasure channel (BEC) as a peeling-type
algorithm (TEP) and we show that the algorithm has the same computational
complexity as BP and it decodes a higher fraction of errors. We describe the
TEP decoding process by a set of differential equations that represents the
expected residual graph evolution as a function of the code parameters. The
solution of these equations is used to predict the TEP decoder performance in
both the asymptotic regime and the finite-length regime over the BEC. While the
asymptotic threshold of the TEP decoder is the same as the BP decoder for
regular and optimized codes, we propose a scaling law (SL) for finite-length
LDPC codes, which accurately approximates the TEP improved performance and
facilitates its optimization