184 research outputs found

    Upper Bounds on the Rate of Low Density Stabilizer Codes for the Quantum Erasure Channel

    Full text link
    Using combinatorial arguments, we determine an upper bound on achievable rates of stabilizer codes used over the quantum erasure channel. This allows us to recover the no-cloning bound on the capacity of the quantum erasure channel, R is below 1-2p, for stabilizer codes: we also derive an improved upper bound of the form : R is below 1-2p-D(p) with a function D(p) that stays positive for 0 < p < 1/2 and for any family of stabilizer codes whose generators have weights bounded from above by a constant - low density stabilizer codes. We obtain an application to percolation theory for a family of self-dual tilings of the hyperbolic plane. We associate a family of low density stabilizer codes with appropriate finite quotients of these tilings. We then relate the probability of percolation to the probability of a decoding error for these codes on the quantum erasure channel. The application of our upper bound on achievable rates of low density stabilizer codes gives rise to an upper bound on the critical probability for these tilings.Comment: 32 page

    Fault-Tolerance of "Bad" Quantum Low-Density Parity Check Codes

    Full text link
    We discuss error-correction properties for families of quantum low-density parity check (LDPC) codes with relative distance that tends to zero in the limit of large blocklength. In particular, we show that any family of LDPC codes, quantum or classical, where distance scales as a positive power of the block length, d∝nαd \propto n^\alpha, α>0\alpha>0, can correct all errors with certainty if the error rate per (qu)bit is sufficiently small. We specifically analyze the case of LDPC version of the quantum hypergraph-product codes recently suggested by Tillich and Z\'emor. These codes are a finite-rate generalization of the toric codes, and, for sufficiently large quantum computers, offer an advantage over the toric codes.Comment: 4.5 pages, 1 figur

    Magic state distillation with punctured polar codes

    Get PDF
    We present a scheme for magic state distillation using punctured polar codes. Our results build on some recent work by Bardet et al. (ISIT, 2016) who discovered that polar codes can be described algebraically as decreasing monomial codes. Using this powerful framework, we construct tri-orthogonal quantum codes (Bravyi et al., PRA, 2012) that can be used to distill magic states for the TT gate. An advantage of these codes is that they permit the use of the successive cancellation decoder whose time complexity scales as O(Nlog⁡(N))O(N\log(N)). We supplement this with numerical simulations for the erasure channel and dephasing channel. We obtain estimates for the dimensions and error rates for the resulting codes for block sizes up to 2202^{20} for the erasure channel and 2162^{16} for the dephasing channel. The dimension of the triply-even codes we obtain is shown to scale like O(N0.8)O(N^{0.8}) for the binary erasure channel at noise rate 0.010.01 and O(N0.84)O(N^{0.84}) for the dephasing channel at noise rate 0.0010.001. The corresponding bit error rates drop to roughly 8×10−288\times10^{-28} for the erasure channel and 7×10−157 \times 10^{-15} for the dephasing channel respectively.Comment: 18 pages, 4 figure

    Distance Verification for Classical and Quantum LDPC Codes

    Get PDF
    The techniques of distance verification known for general linear codes are first applied to the quantum stabilizer codes. Then, these techniques are considered for classical and quantum (stabilizer) low-density-parity-check (LDPC) codes. New complexity bounds for distance verification with provable performance are derived using the average weight spectra of the ensembles of LDPC codes. These bounds are expressed in terms of the erasure-correcting capacity of the corresponding ensemble. We also present a new irreducible-cluster technique that can be applied to any LDPC code and takes advantage of parity-checks’ sparsity for both the classical and quantum LDPC codes. This technique reduces complexity exponents of all existing deterministic techniques designed for generic stabilizer codes with small relative distances, which also include all known families of the quantum stabilizer LDPC codes

    Distance Verification for Classical and Quantum LDPC Codes

    Get PDF
    The techniques of distance verification known for general linear codes are first applied to the quantum stabilizer codes. Then, these techniques are considered for classical and quantum (stabilizer) low-density-parity-check (LDPC) codes. New complexity bounds for distance verification with provable performance are derived using the average weight spectra of the ensembles of LDPC codes. These bounds are expressed in terms of the erasure-correcting capacity of the corresponding ensemble. We also present a new irreducible-cluster technique that can be applied to any LDPC code and takes advantage of parity-checks’ sparsity for both the classical and quantum LDPC codes. This technique reduces complexity exponents of all existing deterministic techniques designed for generic stabilizer codes with small relative distances, which also include all known families of the quantum stabilizer LDPC codes

    Thresholds for correcting errors, erasures, and faulty syndrome measurements in degenerate quantum codes

    Get PDF
    We suggest a technique for constructing lower (existence) bounds for the fault-tolerant threshold to scalable quantum computation applicable to degenerate quantum codes with sublinear distance scaling. We give explicit analytic expressions combining probabilities of erasures, depolarizing errors, and phenomenological syndrome measurement errors for quantum LDPC codes with logarithmic or larger distances. These threshold estimates are parametrically better than the existing analytical bound based on percolation.Comment: 10 pages, 2 figure

    Scrambling speed of random quantum circuits

    Full text link
    Random transformations are typically good at "scrambling" information. Specifically, in the quantum setting, scrambling usually refers to the process of mapping most initial pure product states under a unitary transformation to states which are macroscopically entangled, in the sense of being close to completely mixed on most subsystems containing a fraction fn of all n particles for some constant f. While the term scrambling is used in the context of the black hole information paradox, scrambling is related to problems involving decoupling in general, and to the question of how large isolated many-body systems reach local thermal equilibrium under their own unitary dynamics. Here, we study the speed at which various notions of scrambling/decoupling occur in a simplified but natural model of random two-particle interactions: random quantum circuits. For a circuit representing the dynamics generated by a local Hamiltonian, the depth of the circuit corresponds to time. Thus, we consider the depth of these circuits and we are typically interested in what can be done in a depth that is sublinear or even logarithmic in the size of the system. We resolve an outstanding conjecture raised in the context of the black hole information paradox with respect to the depth at which a typical quantum circuit generates an entanglement assisted encoding against the erasure channel. In addition, we prove that typical quantum circuits of poly(log n) depth satisfy a stronger notion of scrambling and can be used to encode alpha n qubits into n qubits so that up to beta n errors can be corrected, for some constants alpha, beta > 0.Comment: 24 pages, 2 figures. Superseded by http://arxiv.org/abs/1307.063
    • 

    corecore