5,115 research outputs found

    p-Adic valuation of weights in Abelian codes over /spl Zopf/(p/sup d/)

    Get PDF
    Counting polynomial techniques introduced by Wilson are used to provide analogs of a theorem of McEliece. McEliece's original theorem relates the greatest power of p dividing the Hamming weights of words in cyclic codes over GF (p) to the length of the smallest unity-product sequence of nonzeroes of the code. Calderbank, Li, and Poonen presented analogs for cyclic codes over /spl Zopf/(2/sup d/) using various weight functions (Hamming, Lee, and Euclidean weight as well as count of occurrences of a particular symbol). Some of these results were strengthened by Wilson, who also considered the alphabet /spl Zopf/(p/sup d/) for p an arbitrary prime. These previous results, new strengthened versions, and generalizations are proved here in a unified and comprehensive fashion for the larger class of Abelian codes over /spl Zopf/(p/sup d/) with p any prime. For Abelian codes over /spl Zopf//sub 4/, combinatorial methods for use with counting polynomials are developed. These show that the analogs of McEliece's theorem obtained by Wilson (for Hamming weight, Lee weight, and symbol counts) and the analog obtained here for Euclidean weight are sharp in the sense that they give the maximum power of 2 that divides the weights of all the codewords whose Fourier transforms have a specified support

    List Decoding Tensor Products and Interleaved Codes

    Full text link
    We design the first efficient algorithms and prove new combinatorial bounds for list decoding tensor products of codes and interleaved codes. We show that for {\em every} code, the ratio of its list decoding radius to its minimum distance stays unchanged under the tensor product operation (rather than squaring, as one might expect). This gives the first efficient list decoders and new combinatorial bounds for some natural codes including multivariate polynomials where the degree in each variable is bounded. We show that for {\em every} code, its list decoding radius remains unchanged under mm-wise interleaving for an integer mm. This generalizes a recent result of Dinur et al \cite{DGKS}, who proved such a result for interleaved Hadamard codes (equivalently, linear transformations). Using the notion of generalized Hamming weights, we give better list size bounds for {\em both} tensoring and interleaving of binary linear codes. By analyzing the weight distribution of these codes, we reduce the task of bounding the list size to bounding the number of close-by low-rank codewords. For decoding linear transformations, using rank-reduction together with other ideas, we obtain list size bounds that are tight over small fields.Comment: 32 page

    Approximating solution structure of the Weighted Sentence Alignment problem

    Full text link
    We study the complexity of approximating solution structure of the bijective weighted sentence alignment problem of DeNero and Klein (2008). In particular, we consider the complexity of finding an alignment that has a significant overlap with an optimal alignment. We discuss ways of representing the solution for the general weighted sentence alignment as well as phrases-to-words alignment problem, and show that computing a string which agrees with the optimal sentence partition on more than half (plus an arbitrarily small polynomial fraction) positions for the phrases-to-words alignment is NP-hard. For the general weighted sentence alignment we obtain such bound from the agreement on a little over 2/3 of the bits. Additionally, we generalize the Hamming distance approximation of a solution structure to approximating it with respect to the edit distance metric, obtaining similar lower bounds

    Neural networks, error-correcting codes, and polynomials over the binary n-cube

    Get PDF
    Several ways of relating the concept of error-correcting codes to the concept of neural networks are presented. Performing maximum-likelihood decoding in a linear block error-correcting code is shown to be equivalent to finding a global maximum of the energy function of a certain neural network. Given a linear block code, a neural network can be constructed in such a way that every codeword corresponds to a local maximum. The connection between maximization of polynomials over the n-cube and error-correcting codes is also investigated; the results suggest that decoding techniques can be a useful tool for solving such maximization problems. The results are generalized to both nonbinary and nonlinear codes

    On a Class of Optimal Nonbinary Linear Unequal-Error-Protection Codes for Two Sets of Messages

    Get PDF
    Several authors have addressed the problem of designing good linear unequal error protection (LUEP) codes. However, very little is known about good nonbinary LUEP codes. We present a class of optimal nonbinary LUEP codes for two different sets of messages. By combining t-error-correcting ReedSolomon (RS) codes and shortened nonbinary Hamming codes, we obtain nonbinary LUEP codes that protect one set of messages against any t or fewer symbol errors and the remaining set of messages against any single symbol error. For t ≥ 2, we show that these codes are optimal in the sense of achieving the Hamming lower bound on the number of redundant symbols of a nonbinary LUEP code with the same parameters

    A General Upper Bound on the Size of Constant-Weight Conflict-Avoiding Codes

    Full text link
    Conflict-avoiding codes are used in the multiple-access collision channel without feedback. The number of codewords in a conflict-avoiding code is the number of potential users that can be supported in the system. In this paper, a new upper bound on the size of conflict-avoiding codes is proved. This upper bound is general in the sense that it is applicable to all code lengths and all Hamming weights. Several existing constructions for conflict-avoiding codes, which are known to be optimal for Hamming weights equal to four and five, are shown to be optimal for all Hamming weights in general.Comment: 10 pages, 1 figur
    • …
    corecore