11,193 research outputs found
A MILP approach for designing robust variable-length codes based on exact free distance computation
International audienceThis paper addresses the design of joint source-channel variable-length codes with maximal free distance for given codeword lengths. While previous design methods are mainly based on bounds on the free distance of the code, the proposed algorithm exploits an exact characterization of the free distance. The code optimization is cast in the framework of mixed-integer linear programming and allows to tackle practical alphabet sizes in reasonable computing time
Convergence Speed of the Consensus Algorithm with Interference and Sparse Long-Range Connectivity
We analyze the effect of interference on the convergence rate of average
consensus algorithms, which iteratively compute the measurement average by
message passing among nodes. It is usually assumed that these algorithms
converge faster with a greater exchange of information (i.e., by increased
network connectivity) in every iteration. However, when interference is taken
into account, it is no longer clear if the rate of convergence increases with
network connectivity. We study this problem for randomly-placed
consensus-seeking nodes connected through an interference-limited network. We
investigate the following questions: (a) How does the rate of convergence vary
with increasing communication range of each node? and (b) How does this result
change when each node is allowed to communicate with a few selected far-off
nodes? When nodes schedule their transmissions to avoid interference, we show
that the convergence speed scales with , where is the
communication range and is the number of dimensions. This scaling is the
result of two competing effects when increasing : Increased schedule length
for interference-free transmission vs. the speed gain due to improved
connectivity. Hence, although one-dimensional networks can converge faster from
a greater communication range despite increased interference, the two effects
exactly offset one another in two-dimensions. In higher dimensions, increasing
the communication range can actually degrade the rate of convergence. Our
results thus underline the importance of factoring in the effect of
interference in the design of distributed estimation algorithms.Comment: 27 pages, 4 figure
Interactive Channel Capacity Revisited
We provide the first capacity approaching coding schemes that robustly
simulate any interactive protocol over an adversarial channel that corrupts any
fraction of the transmitted symbols. Our coding schemes achieve a
communication rate of over any
adversarial channel. This can be improved to for
random, oblivious, and computationally bounded channels, or if parties have
shared randomness unknown to the channel.
Surprisingly, these rates exceed the interactive channel capacity bound
which [Kol and Raz; STOC'13] recently proved for random errors. We conjecture
and to be the optimal rates for their respective settings
and therefore to capture the interactive channel capacity for random and
adversarial errors.
In addition to being very communication efficient, our randomized coding
schemes have multiple other advantages. They are computationally efficient,
extremely natural, and significantly simpler than prior (non-capacity
approaching) schemes. In particular, our protocols do not employ any coding but
allow the original protocol to be performed as-is, interspersed only by short
exchanges of hash values. When hash values do not match, the parties backtrack.
Our approach is, as we feel, by far the simplest and most natural explanation
for why and how robust interactive communication in a noisy environment is
possible
Topological quantum memory
We analyze surface codes, the topological quantum error-correcting codes
introduced by Kitaev. In these codes, qubits are arranged in a two-dimensional
array on a surface of nontrivial topology, and encoded quantum operations are
associated with nontrivial homology cycles of the surface. We formulate
protocols for error recovery, and study the efficacy of these protocols. An
order-disorder phase transition occurs in this system at a nonzero critical
value of the error rate; if the error rate is below the critical value (the
accuracy threshold), encoded information can be protected arbitrarily well in
the limit of a large code block. This phase transition can be accurately
modeled by a three-dimensional Z_2 lattice gauge theory with quenched disorder.
We estimate the accuracy threshold, assuming that all quantum gates are local,
that qubits can be measured rapidly, and that polynomial-size classical
computations can be executed instantaneously. We also devise a robust recovery
procedure that does not require measurement or fast classical processing;
however for this procedure the quantum gates are local only if the qubits are
arranged in four or more spatial dimensions. We discuss procedures for
encoding, measurement, and performing fault-tolerant universal quantum
computation with surface codes, and argue that these codes provide a promising
framework for quantum computing architectures.Comment: 39 pages, 21 figures, REVTe
Irregular Variable Length Coding
In this thesis, we introduce Irregular Variable Length Coding (IrVLC) and investigate its applications, characteristics and performance in the context of digital multimedia broadcast telecommunications. During IrVLC encoding, the multimedia signal is represented using a sequence of concatenated binary codewords. These are selected from a codebook, comprising a number of codewords, which, in turn, comprise various numbers of bits. However, during IrVLC encoding, the multimedia signal is decomposed into particular fractions, each of which is represented using a different codebook. This is in contrast to regular Variable Length Coding (VLC), in which the entire multimedia signal is encoded using the same codebook. The application of IrVLCs to joint source and channel coding is investigated in the context of a video transmission scheme. Our novel video codec represents the video signal using tessellations of Variable-Dimension Vector Quantisation (VDVQ) tiles. These are selected from a codebook, comprising a number of tiles having various dimensions. The selected tessellation of VDVQ tiles is signalled using a corresponding sequence of concatenated codewords from a Variable Length Error Correction (VLEC) codebook. This VLEC codebook represents a specific joint source and channel coding case of VLCs, which facilitates both compression and error correction. However, during video encoding, only particular combinations of the VDVQ tiles will perfectly tessellate, owing to their various dimensions. As a result, only particular sub-sets of the VDVQ codebook and, hence, of the VLEC codebook may be employed to convey particular fractions of the video signal. Therefore, our novel video codec can be said to employ IrVLCs. The employment of IrVLCs to facilitate Unequal Error Protection (UEP) is also demonstrated. This may be applied when various fractions of the source signal have different error sensitivities, as is typical in audio, speech, image and video signals, for example. Here, different VLEC codebooks having appropriately selected error correction capabilities may be employed to encode the particular fractions of the source signal. This approach may be expected to yield a higher reconstruction quality than equal protection in cases where the various fractions of the source signal have different error sensitivities. Finally, this thesis investigates the application of IrVLCs to near-capacity operation using EXtrinsic Information Transfer (EXIT) chart analysis. Here, a number of component VLEC codebooks having different inverted EXIT functions are employed to encode particular fractions of the source symbol frame. We show that the composite inverted IrVLC EXIT function may be obtained as a weighted average of the inverted component VLC EXIT functions. Additionally, EXIT chart matching is employed to shape the inverted IrVLC EXIT function to match the EXIT function of a serially concatenated inner channel code, creating a narrow but still open EXIT chart tunnel. In this way, iterative decoding convergence to an infinitesimally low probability of error is facilitated at near-capacity channel SNRs
- …