202 research outputs found
The Trapping Redundancy of Linear Block Codes
We generalize the notion of the stopping redundancy in order to study the
smallest size of a trapping set in Tanner graphs of linear block codes. In this
context, we introduce the notion of the trapping redundancy of a code, which
quantifies the relationship between the number of redundant rows in any
parity-check matrix of a given code and the size of its smallest trapping set.
Trapping sets with certain parameter sizes are known to cause error-floors in
the performance curves of iterative belief propagation decoders, and it is
therefore important to identify decoding matrices that avoid such sets. Bounds
on the trapping redundancy are obtained using probabilistic and constructive
methods, and the analysis covers both general and elementary trapping sets.
Numerical values for these bounds are computed for the [2640,1320] Margulis
code and the class of projective geometry codes, and compared with some new
code-specific trapping set size estimates.Comment: 12 pages, 4 tables, 1 figure, accepted for publication in IEEE
Transactions on Information Theor
Density Evolution and Functional Threshold for the Noisy Min-Sum Decoder
This paper investigates the behavior of the Min-Sum decoder running on noisy
devices. The aim is to evaluate the robustness of the decoder in the presence
of computation noise, e.g. due to faulty logic in the processing units, which
represents a new source of errors that may occur during the decoding process.
To this end, we first introduce probabilistic models for the arithmetic and
logic units of the the finite-precision Min-Sum decoder, and then carry out the
density evolution analysis of the noisy Min-Sum decoder. We show that in some
particular cases, the noise introduced by the device can help the Min-Sum
decoder to escape from fixed points attractors, and may actually result in an
increased correction capacity with respect to the noiseless decoder. We also
reveal the existence of a specific threshold phenomenon, referred to as
functional threshold. The behavior of the noisy decoder is demonstrated in the
asymptotic limit of the code-length -- by using "noisy" density evolution
equations -- and it is also verified in the finite-length case by Monte-Carlo
simulation.Comment: 46 pages (draft version); extended version of the paper with same
title, submitted to IEEE Transactions on Communication
Analysis and Design of Finite Alphabet Iterative Decoders Robust to Faulty Hardware
This paper addresses the problem of designing LDPC decoders robust to
transient errors introduced by a faulty hardware. We assume that the faulty
hardware introduces errors during the message passing updates and we propose a
general framework for the definition of the message update faulty functions.
Within this framework, we define symmetry conditions for the faulty functions,
and derive two simple error models used in the analysis. With this analysis, we
propose a new interpretation of the functional Density Evolution threshold
previously introduced, and show its limitations in case of highly unreliable
hardware. However, we show that under restricted decoder noise conditions, the
functional threshold can be used to predict the convergence behavior of FAIDs
under faulty hardware. In particular, we reveal the existence of robust and
non-robust FAIDs and propose a framework for the design of robust decoders. We
finally illustrate robust and non-robust decoders behaviors of finite length
codes using Monte Carlo simulations.Comment: 30 pages, submitted to IEEE Transactions on Communication
Decoding LDPC Codes with Probabilistic Local Maximum Likelihood Bit Flipping
Communication channels are inherently noisy making error correction coding a major topic of research for modern communication systems. Error correction coding is the addition of redundancy to information transmitted over communication channels to enable detection and recovery of erroneous information. Low-density parity-check (LDPC) codes are a class of error correcting codes that have been effective in maintaining reliability of information transmitted over communication channels. Multiple algorithms have been developed to benefit from the LDPC coding scheme to improve recovery of erroneous information. This work develops a matrix construction that stores the information error probability statistics for a communication channel. This combined with the error correcting capability of LDPC codes enabled the development of the Probabilistic Local Maximum Likelihood Bit Flipping (PLMLBF) algorithm, which is the focus of this research work
Gurafu hyogen o riyoshita ayamari teisei hoshiki no kosei ni kansuru kenkyu
制度:新 ; 報告番号:乙2221号 ; 学位の種類:博士(工学) ; 授与年月日:2009/3/24 ; 早大学位記番号:新508
Characterization and Efficient Search of Non-Elementary Trapping Sets of LDPC Codes with Applications to Stopping Sets
In this paper, we propose a characterization for non-elementary trapping sets
(NETSs) of low-density parity-check (LDPC) codes. The characterization is based
on viewing a NETS as a hierarchy of embedded graphs starting from an ETS. The
characterization corresponds to an efficient search algorithm that under
certain conditions is exhaustive. As an application of the proposed
characterization/search, we obtain lower and upper bounds on the stopping
distance of LDPC codes.
We examine a large number of regular and irregular LDPC codes, and
demonstrate the efficiency and versatility of our technique in finding lower
and upper bounds on, and in many cases the exact value of, . Finding
, or establishing search-based lower or upper bounds, for many of the
examined codes are out of the reach of any existing algorithm
- …