12,663 research outputs found

    Trellises and Trellis-Based Decoding Algorithms for Linear Block Codes

    Get PDF
    A code trellis is a graphical representation of a code, block or convolutional, in which every path represents a codeword (or a code sequence for a convolutional code). This representation makes it possible to implement Maximum Likelihood Decoding (MLD) of a code with reduced decoding complexity. The most well known trellis-based MLD algorithm is the Viterbi algorithm. The trellis representation was first introduced and used for convolutional codes [23]. This representation, together with the Viterbi decoding algorithm, has resulted in a wide range of applications of convolutional codes for error control in digital communications over the last two decades. There are two major reasons for this inactive period of research in this area. First, most coding theorists at that time believed that block codes did not have simple trellis structure like convolutional codes and maximum likelihood decoding of linear block codes using the Viterbi algorithm was practically impossible, except for very short block codes. Second, since almost all of the linear block codes are constructed algebraically or based on finite geometries, it was the belief of many coding theorists that algebraic decoding was the only way to decode these codes. These two reasons seriously hindered the development of efficient soft-decision decoding methods for linear block codes and their applications to error control in digital communications. This led to a general belief that block codes are inferior to convolutional codes and hence, that they were not useful. Chapter 2 gives a brief review of linear block codes. The goal is to provide the essential background material for the development of trellis structure and trellis-based decoding algorithms for linear block codes in the later chapters. Chapters 3 through 6 present the fundamental concepts, finite-state machine model, state space formulation, basic structural properties, state labeling, construction procedures, complexity, minimality, and sectionalization of trellises. Chapter 7 discusses trellis decomposition and subtrellises for low-weight codewords. Chapter 8 first presents well known methods for constructing long powerful codes from short component codes or component codes of smaller dimensions, and then provides methods for constructing their trellises which include Shannon and Cartesian product techniques. Chapter 9 deals with convolutional codes, puncturing, zero-tail termination and tail-biting.Chapters 10 through 13 present various trellis-based decoding algorithms, old and new. Chapter 10 first discusses the application of the well known Viterbi decoding algorithm to linear block codes, optimum sectionalization of a code trellis to minimize computation complexity, and design issues for IC (integrated circuit) implementation of a Viterbi decoder. Then it presents a new decoding algorithm for convolutional codes, named Differential Trellis Decoding (DTD) algorithm. Chapter 12 presents a suboptimum reliability-based iterative decoding algorithm with a low-weight trellis search for the most likely codeword. This decoding algorithm provides a good trade-off between error performance and decoding complexity. All the decoding algorithms presented in Chapters 10 through 12 are devised to minimize word error probability. Chapter 13 presents decoding algorithms that minimize bit error probability and provide the corresponding soft (reliability) information at the output of the decoder. Decoding algorithms presented are the MAP (maximum a posteriori probability) decoding algorithm and the Soft-Output Viterbi Algorithm (SOVA) algorithm. Finally, the minimization of bit error probability in trellis-based MLD is discussed

    Universal and Robust Distributed Network Codes

    Full text link
    Random linear network codes can be designed and implemented in a distributed manner, with low computational complexity. However, these codes are classically implemented over finite fields whose size depends on some global network parameters (size of the network, the number of sinks) that may not be known prior to code design. Also, if new nodes join the entire network code may have to be redesigned. In this work, we present the first universal and robust distributed linear network coding schemes. Our schemes are universal since they are independent of all network parameters. They are robust since if nodes join or leave, the remaining nodes do not need to change their coding operations and the receivers can still decode. They are distributed since nodes need only have topological information about the part of the network upstream of them, which can be naturally streamed as part of the communication protocol. We present both probabilistic and deterministic schemes that are all asymptotically rate-optimal in the coding block-length, and have guarantees of correctness. Our probabilistic designs are computationally efficient, with order-optimal complexity. Our deterministic designs guarantee zero error decoding, albeit via codes with high computational complexity in general. Our coding schemes are based on network codes over ``scalable fields". Instead of choosing coding coefficients from one field at every node, each node uses linear coding operations over an ``effective field-size" that depends on the node's distance from the source node. The analysis of our schemes requires technical tools that may be of independent interest. In particular, we generalize the Schwartz-Zippel lemma by proving a non-uniform version, wherein variables are chosen from sets of possibly different sizes. We also provide a novel robust distributed algorithm to assign unique IDs to network nodes.Comment: 12 pages, 7 figures, 1 table, under submission to INFOCOM 201

    Iterative decoding for MIMO channels via modified sphere decoding

    Get PDF
    In recent years, soft iterative decoding techniques have been shown to greatly improve the bit error rate performance of various communication systems. For multiantenna systems employing space-time codes, however, it is not clear what is the best way to obtain the soft information required of the iterative scheme with low complexity. In this paper, we propose a modification of the Fincke-Pohst (sphere decoding) algorithm to estimate the maximum a posteriori probability of the received symbol sequence. The new algorithm solves a nonlinear integer least squares problem and, over a wide range of rates and signal-to-noise ratios, has polynomial-time complexity. Performance of the algorithm, combined with convolutional, turbo, and low-density parity check codes, is demonstrated on several multiantenna channels. The results for systems that employ space-time modulation schemes seem to indicate that the best performing schemes are those that support the highest mutual information between the transmitted and received signals, rather than the best diversity gain

    Spatially Coupled Turbo Codes

    Get PDF
    In this paper, we introduce the concept of spatially coupled turbo codes (SC-TCs), as the turbo codes counterpart of spatially coupled low-density parity-check codes. We describe spatial coupling for both Berrou et al. and Benedetto et al. parallel and serially concatenated codes. For the binary erasure channel, we derive the exact density evolution (DE) equations of SC-TCs by using the method proposed by Kurkoski et al. to compute the decoding erasure probability of convolutional encoders. Using DE, we then analyze the asymptotic behavior of SC-TCs. We observe that the belief propagation (BP) threshold of SC-TCs improves with respect to that of the uncoupled ensemble and approaches its maximum a posteriori threshold. This phenomenon is especially significant for serially concatenated codes, whose uncoupled ensemble suffers from a poor BP threshold.Comment: in Proc. 8th International Symposium on Turbo Codes & Iterative Information Processing 2014, Bremen, Germany, August 2014. To appear. (The PCC ensemble is changed with respect to the one in the previous version of the paper. However, it gives identical thresholds

    Structured Random Linear Codes (SRLC): Bridging the Gap between Block and Convolutional Codes

    Get PDF
    Several types of AL-FEC (Application-Level FEC) codes for the Packet Erasure Channel exist. Random Linear Codes (RLC), where redundancy packets consist of random linear combinations of source packets over a certain finite field, are a simple yet efficient coding technique, for instance massively used for Network Coding applications. However the price to pay is a high encoding and decoding complexity, especially when working on GF(28)GF(2^8), which seriously limits the number of packets in the encoding window. On the opposite, structured block codes have been designed for situations where the set of source packets is known in advance, for instance with file transfer applications. Here the encoding and decoding complexity is controlled, even for huge block sizes, thanks to the sparse nature of the code and advanced decoding techniques that exploit this sparseness (e.g., Structured Gaussian Elimination). But their design also prevents their use in convolutional use-cases featuring an encoding window that slides over a continuous set of incoming packets. In this work we try to bridge the gap between these two code classes, bringing some structure to RLC codes in order to enlarge the use-cases where they can be efficiently used: in convolutional mode (as any RLC code), but also in block mode with either tiny, medium or large block sizes. We also demonstrate how to design compact signaling for these codes (for encoder/decoder synchronization), which is an essential practical aspect.Comment: 7 pages, 12 figure
    • …
    corecore