1,328 research outputs found
Density Evolution and Functional Threshold for the Noisy Min-Sum Decoder
This paper investigates the behavior of the Min-Sum decoder running on noisy
devices. The aim is to evaluate the robustness of the decoder in the presence
of computation noise, e.g. due to faulty logic in the processing units, which
represents a new source of errors that may occur during the decoding process.
To this end, we first introduce probabilistic models for the arithmetic and
logic units of the the finite-precision Min-Sum decoder, and then carry out the
density evolution analysis of the noisy Min-Sum decoder. We show that in some
particular cases, the noise introduced by the device can help the Min-Sum
decoder to escape from fixed points attractors, and may actually result in an
increased correction capacity with respect to the noiseless decoder. We also
reveal the existence of a specific threshold phenomenon, referred to as
functional threshold. The behavior of the noisy decoder is demonstrated in the
asymptotic limit of the code-length -- by using "noisy" density evolution
equations -- and it is also verified in the finite-length case by Monte-Carlo
simulation.Comment: 46 pages (draft version); extended version of the paper with same
title, submitted to IEEE Transactions on Communication
Decoding of Non-Binary LDPC Codes Using the Information Bottleneck Method
Recently, a novel lookup table based decoding method for binary low-density
parity-check codes has attracted considerable attention. In this approach,
mutual-information maximizing lookup tables replace the conventional operations
of the variable nodes and the check nodes in message passing decoding.
Moreover, the exchanged messages are represented by integers with very small
bit width. A machine learning framework termed the information bottleneck
method is used to design the corresponding lookup tables. In this paper, we
extend this decoding principle from binary to non-binary codes. This is not a
straightforward extension, but requires a more sophisticated lookup table
design to cope with the arithmetic in higher order Galois fields. Provided bit
error rate simulations show that our proposed scheme outperforms the log-max
decoding algorithm and operates close to sum-product decoding.Comment: This paper has been presented at IEEE International Conference on
Communications (ICC'19) in Shangha
Rewriting Flash Memories by Message Passing
This paper constructs WOM codes that combine rewriting and error correction
for mitigating the reliability and the endurance problems in flash memory. We
consider a rewriting model that is of practical interest to flash applications
where only the second write uses WOM codes. Our WOM code construction is based
on binary erasure quantization with LDGM codes, where the rewriting uses
message passing and has potential to share the efficient hardware
implementations with LDPC codes in practice. We show that the coding scheme
achieves the capacity of the rewriting model. Extensive simulations show that
the rewriting performance of our scheme compares favorably with that of polar
WOM code in the rate region where high rewriting success probability is
desired. We further augment our coding schemes with error correction
capability. By drawing a connection to the conjugate code pairs studied in the
context of quantum error correction, we develop a general framework for
constructing error-correction WOM codes. Under this framework, we give an
explicit construction of WOM codes whose codewords are contained in BCH codes.Comment: Submitted to ISIT 201
- …