1,211 research outputs found

    List decoding of a class of affine variety codes

    Full text link
    Consider a polynomial FF in mm variables and a finite point ensemble S=S1×...×SmS=S_1 \times ... \times S_m. When given the leading monomial of FF with respect to a lexicographic ordering we derive improved information on the possible number of zeros of FF of multiplicity at least rr from SS. We then use this information to design a list decoding algorithm for a large class of affine variety codes.Comment: 11 pages, 5 table

    Reed-Muller codes for random erasures and errors

    Full text link
    This paper studies the parameters for which Reed-Muller (RM) codes over GF(2)GF(2) can correct random erasures and random errors with high probability, and in particular when can they achieve capacity for these two classical channels. Necessarily, the paper also studies properties of evaluations of multi-variate GF(2)GF(2) polynomials on random sets of inputs. For erasures, we prove that RM codes achieve capacity both for very high rate and very low rate regimes. For errors, we prove that RM codes achieve capacity for very low rate regimes, and for very high rates, we show that they can uniquely decode at about square root of the number of errors at capacity. The proofs of these four results are based on different techniques, which we find interesting in their own right. In particular, we study the following questions about E(m,r)E(m,r), the matrix whose rows are truth tables of all monomials of degree r\leq r in mm variables. What is the most (resp. least) number of random columns in E(m,r)E(m,r) that define a submatrix having full column rank (resp. full row rank) with high probability? We obtain tight bounds for very small (resp. very large) degrees rr, which we use to show that RM codes achieve capacity for erasures in these regimes. Our decoding from random errors follows from the following novel reduction. For every linear code CC of sufficiently high rate we construct a new code CC', also of very high rate, such that for every subset SS of coordinates, if CC can recover from erasures in SS, then CC' can recover from errors in SS. Specializing this to RM codes and using our results for erasures imply our result on unique decoding of RM codes at high rate. Finally, two of our capacity achieving results require tight bounds on the weight distribution of RM codes. We obtain such bounds extending the recent \cite{KLP} bounds from constant degree to linear degree polynomials

    Efficiently decoding Reed-Muller codes from random errors

    Full text link
    Reed-Muller codes encode an mm-variate polynomial of degree rr by evaluating it on all points in {0,1}m\{0,1\}^m. We denote this code by RM(m,r)RM(m,r). The minimal distance of RM(m,r)RM(m,r) is 2mr2^{m-r} and so it cannot correct more than half that number of errors in the worst case. For random errors one may hope for a better result. In this work we give an efficient algorithm (in the block length n=2mn=2^m) for decoding random errors in Reed-Muller codes far beyond the minimal distance. Specifically, for low rate codes (of degree r=o(m)r=o(\sqrt{m})) we can correct a random set of (1/2o(1))n(1/2-o(1))n errors with high probability. For high rate codes (of degree mrm-r for r=o(m/logm)r=o(\sqrt{m/\log m})), we can correct roughly mr/2m^{r/2} errors. More generally, for any integer rr, our algorithm can correct any error pattern in RM(m,m(2r+2))RM(m,m-(2r+2)) for which the same erasure pattern can be corrected in RM(m,m(r+1))RM(m,m-(r+1)). The results above are obtained by applying recent results of Abbe, Shpilka and Wigderson (STOC, 2015), Kumar and Pfister (2015) and Kudekar et al. (2015) regarding the ability of Reed-Muller codes to correct random erasures. The algorithm is based on solving a carefully defined set of linear equations and thus it is significantly different than other algorithms for decoding Reed-Muller codes that are based on the recursive structure of the code. It can be seen as a more explicit proof of a result of Abbe et al. that shows a reduction from correcting erasures to correcting errors, and it also bares some similarities with the famous Berlekamp-Welch algorithm for decoding Reed-Solomon codes.Comment: 18 pages, 2 figure

    Neural networks, error-correcting codes, and polynomials over the binary n-cube

    Get PDF
    Several ways of relating the concept of error-correcting codes to the concept of neural networks are presented. Performing maximum-likelihood decoding in a linear block error-correcting code is shown to be equivalent to finding a global maximum of the energy function of a certain neural network. Given a linear block code, a neural network can be constructed in such a way that every codeword corresponds to a local maximum. The connection between maximization of polynomials over the n-cube and error-correcting codes is also investigated; the results suggest that decoding techniques can be a useful tool for solving such maximization problems. The results are generalized to both nonbinary and nonlinear codes

    Decoding Reed-Muller codes over product sets

    Get PDF
    We give a polynomial time algorithm to decode multivariate polynomial codes of degree dd up to half their minimum distance, when the evaluation points are an arbitrary product set SmS^m, for every d<Sd < |S|. Previously known algorithms can achieve this only if the set SS has some very special algebraic structure, or if the degree dd is significantly smaller than S|S|. We also give a near-linear time randomized algorithm, which is based on tools from list-decoding, to decode these codes from nearly half their minimum distance, provided d0d 0. Our result gives an mm-dimensional generalization of the well known decoding algorithms for Reed-Solomon codes, and can be viewed as giving an algorithmic version of the Schwartz-Zippel lemma.Comment: 25 pages, 0 figure

    Information Sets of Multiplicity Codes

    Get PDF
    We here provide a method for systematic encoding of the Multiplicity codes introduced by Kopparty, Saraf and Yekhanin in 2011. The construction is built on an idea of Kop-party. We properly define information sets for these codes and give detailed proofs of the validity of Kopparty's construction, that use generating functions. We also give a complexity estimate of the associated encoding algorithm.Comment: International Symposium on Information Theory, Jun 2015, Hong-Kong, China. IEE
    corecore