8,263 research outputs found
High-rate self-synchronizing codes
Self-synchronization under the presence of additive noise can be achieved by
allocating a certain number of bits of each codeword as markers for
synchronization. Difference systems of sets are combinatorial designs which
specify the positions of synchronization markers in codewords in such a way
that the resulting error-tolerant self-synchronizing codes may be realized as
cosets of linear codes. Ideally, difference systems of sets should sacrifice as
few bits as possible for a given code length, alphabet size, and
error-tolerance capability. However, it seems difficult to attain optimality
with respect to known bounds when the noise level is relatively low. In fact,
the majority of known optimal difference systems of sets are for exceptionally
noisy channels, requiring a substantial amount of bits for synchronization. To
address this problem, we present constructions for difference systems of sets
that allow for higher information rates while sacrificing optimality to only a
small extent. Our constructions utilize optimal difference systems of sets as
ingredients and, when applied carefully, generate asymptotically optimal ones
with higher information rates. We also give direct constructions for optimal
difference systems of sets with high information rates and error-tolerance that
generate binary and ternary self-synchronizing codes.Comment: 9 pages, no figure, 2 tables. Final accepted version for publication
in the IEEE Transactions on Information Theory. Material presented in part at
the International Symposium on Information Theory and its Applications,
Honolulu, HI USA, October 201
Parsing a sequence of qubits
We develop a theoretical framework for frame synchronization, also known as
block synchronization, in the quantum domain which makes it possible to attach
classical and quantum metadata to quantum information over a noisy channel even
when the information source and sink are frame-wise asynchronous. This
eliminates the need of frame synchronization at the hardware level and allows
for parsing qubit sequences during quantum information processing. Our
framework exploits binary constant-weight codes that are self-synchronizing.
Possible applications may include asynchronous quantum communication such as a
self-synchronizing quantum network where one can hop into the channel at any
time, catch the next coming quantum information with a label indicating the
sender, and reply by routing her quantum information with control qubits for
quantum switches all without assuming prior frame synchronization between
users.Comment: 11 pages, 2 figures, 1 table. Final accepted version for publication
in the IEEE Transactions on Information Theor
A study of high density bit transition requirements versus the effects on BCH error correcting coding
Several methods for increasing bit transition densities in a data stream are summarized, discussed in detail, and compared against constraints imposed by the 2 MHz data link of the space shuttle high rate multiplexer unit. These methods include use of alternate pulse code modulation waveforms, data stream modification by insertion, alternate bit inversion, differential encoding, error encoding, and use of bit scramblers. The psuedo-random cover sequence generator was chosen for application to the 2 MHz data link of the space shuttle high rate multiplexer unit. This method is fully analyzed and a design implementation proposed
A study of high density bit transition requirements versus the effects on BCH error correcting codes
The use of PN sequence generators to create a minimum number of transitions in an NRZ bit stream is described. The CSG encoder/decoder design was constructed and demonstrated
Synchronization recovery and state model reduction for soft decoding of variable length codes
Variable length codes exhibit de-synchronization problems when transmitted
over noisy channels. Trellis decoding techniques based on Maximum A Posteriori
(MAP) estimators are often used to minimize the error rate on the estimated
sequence. If the number of symbols and/or bits transmitted are known by the
decoder, termination constraints can be incorporated in the decoding process.
All the paths in the trellis which do not lead to a valid sequence length are
suppressed. This paper presents an analytic method to assess the expected error
resilience of a VLC when trellis decoding with a sequence length constraint is
used. The approach is based on the computation, for a given code, of the amount
of information brought by the constraint. It is then shown that this quantity
as well as the probability that the VLC decoder does not re-synchronize in a
strict sense, are not significantly altered by appropriate trellis states
aggregation. This proves that the performance obtained by running a
length-constrained Viterbi decoder on aggregated state models approaches the
one obtained with the bit/symbol trellis, with a significantly reduced
complexity. It is then shown that the complexity can be further decreased by
projecting the state model on two state models of reduced size
Error resilient image transmission using T-codes and edge-embedding
Current image communication applications involve image transmission over noisy channels, where the image gets damaged. The loss of synchronization at the decoder due to these errors increases the damage in the reconstructed image. Our main goal in this research is to develop an algorithm that has the capability to detect errors, achieve synchronization and conceal errors.;In this thesis we studied the performance of T-codes in comparison with Huffman codes. We develop an algorithm for the selection of best T-code set. We have shown that T-codes exhibit better synchronization properties when compared to Huffman Codes. In this work we developed an algorithm that extracts edge patterns from each 8x8 block, classifies edge patterns into different classes. In this research we also propose a novel scrambling algorithm to hide edge pattern of a block into neighboring 8x8 blocks of the image. This scrambled hidden data is used in the detection of errors and concealment of errors. We also develop an algorithm to protect the hidden data from getting damaged in the course of transmission
Carrier recovery techniques on satellite mobile channels
An analytical method and a stored channel model were used to evaluate error performance of uncoded quadrature phase shift keying (QPSK) and M-ary phase shift keying (MPSK) trellis coded modulation (TCM) over shadowed satellite mobile channels in the presence of phase jitter for various carrier recovery techniques
Vulnerability of LTE to Hostile Interference
LTE is well on its way to becoming the primary cellular standard, due to its
performance and low cost. Over the next decade we will become dependent on LTE,
which is why we must ensure it is secure and available when we need it.
Unfortunately, like any wireless technology, disruption through radio jamming
is possible. This paper investigates the extent to which LTE is vulnerable to
intentional jamming, by analyzing the components of the LTE downlink and uplink
signals. The LTE physical layer consists of several physical channels and
signals, most of which are vital to the operation of the link. By taking into
account the density of these physical channels and signals with respect to the
entire frame, as well as the modulation and coding schemes involved, we come up
with a series of vulnerability metrics in the form of jammer to signal ratios.
The ``weakest links'' of the LTE signals are then identified, and used to
establish the overall vulnerability of LTE to hostile interference.Comment: 4 pages, see below for citation. M. Lichtman, J. Reed, M. Norton, T.
Clancy, "Vulnerability of LTE to Hostile Interference'', IEEE Global
Conference on Signal and Information Processing (GlobalSIP), Dec 201
Occam's Quantum Strop: Synchronizing and Compressing Classical Cryptic Processes via a Quantum Channel
A stochastic process's statistical complexity stands out as a fundamental
property: the minimum information required to synchronize one process generator
to another. How much information is required, though, when synchronizing over a
quantum channel? Recent work demonstrated that representing causal similarity
as quantum state-indistinguishability provides a quantum advantage. We
generalize this to synchronization and offer a sequence of constructions that
exploit extended causal structures, finding substantial increase of the quantum
advantage. We demonstrate that maximum compression is determined by the
process's cryptic order---a classical, topological property closely allied to
Markov order, itself a measure of historical dependence. We introduce an
efficient algorithm that computes the quantum advantage and close noting that
the advantage comes at a cost---one trades off prediction for generation
complexity.Comment: 10 pages, 6 figures;
http://csc.ucdavis.edu/~cmg/compmech/pubs/oqs.ht
- …