965 research outputs found
Efficient Linear Programming Decoding of HDPC Codes
We propose several improvements for Linear Programming (LP) decoding
algorithms for High Density Parity Check (HDPC) codes. First, we use the
automorphism groups of a code to create parity check matrix diversity and to
generate valid cuts from redundant parity checks. Second, we propose an
efficient mixed integer decoder utilizing the branch and bound method. We
further enhance the proposed decoders by removing inactive constraints and by
adapting the parity check matrix prior to decoding according to the channel
observations. Based on simulation results the proposed decoders achieve near-ML
performance with reasonable complexity.Comment: Submitted to the IEEE Transactions on Communications, November 200
Pruning Neural Belief Propagation Decoders
We consider near maximum-likelihood (ML) decoding of short linear block codes
based on neural belief propagation (BP) decoding recently introduced by
Nachmani et al.. While this method significantly outperforms conventional BP
decoding, the underlying parity-check matrix may still limit the overall
performance. In this paper, we introduce a method to tailor an overcomplete
parity-check matrix to (neural) BP decoding using machine learning. We consider
the weights in the Tanner graph as an indication of the importance of the
connected check nodes (CNs) to decoding and use them to prune unimportant CNs.
As the pruning is not tied over iterations, the final decoder uses a different
parity-check matrix in each iteration. For Reed-Muller and short low-density
parity-check codes, we achieve performance within 0.27 dB and 1.5 dB of the ML
performance while reducing the complexity of the decoder
Spherical and Hyperbolic Toric Topology-Based Codes On Graph Embedding for Ising MRF Models: Classical and Quantum Topology Machine Learning
The paper introduces the application of information geometry to describe the
ground states of Ising models by utilizing parity-check matrices of cyclic and
quasi-cyclic codes on toric and spherical topologies. The approach establishes
a connection between machine learning and error-correcting coding. This
proposed approach has implications for the development of new embedding methods
based on trapping sets. Statistical physics and number geometry applied for
optimize error-correcting codes, leading to these embedding and sparse
factorization methods. The paper establishes a direct connection between DNN
architecture and error-correcting coding by demonstrating how state-of-the-art
architectures (ChordMixer, Mega, Mega-chunk, CDIL, ...) from the long-range
arena can be equivalent to of block and convolutional LDPC codes (Cage-graph,
Repeat Accumulate). QC codes correspond to certain types of chemical elements,
with the carbon element being represented by the mixed automorphism
Shu-Lin-Fossorier QC-LDPC code. The connections between Belief Propagation and
the Permanent, Bethe-Permanent, Nishimori Temperature, and Bethe-Hessian Matrix
are elaborated upon in detail. The Quantum Approximate Optimization Algorithm
(QAOA) used in the Sherrington-Kirkpatrick Ising model can be seen as analogous
to the back-propagation loss function landscape in training DNNs. This
similarity creates a comparable problem with TS pseudo-codeword, resembling the
belief propagation method. Additionally, the layer depth in QAOA correlates to
the number of decoding belief propagation iterations in the Wiberg decoding
tree. Overall, this work has the potential to advance multiple fields, from
Information Theory, DNN architecture design (sparse and structured prior graph
topology), efficient hardware design for Quantum and Classical DPU/TPU (graph,
quantize and shift register architect.) to Materials Science and beyond.Comment: 71 pages, 42 Figures, 1 Table, 1 Appendix. arXiv admin note: text
overlap with arXiv:2109.08184 by other author
- …