331 research outputs found
An Iteratively Decodable Tensor Product Code with Application to Data Storage
The error pattern correcting code (EPCC) can be constructed to provide a
syndrome decoding table targeting the dominant error events of an inter-symbol
interference channel at the output of the Viterbi detector. For the size of the
syndrome table to be manageable and the list of possible error events to be
reasonable in size, the codeword length of EPCC needs to be short enough.
However, the rate of such a short length code will be too low for hard drive
applications. To accommodate the required large redundancy, it is possible to
record only a highly compressed function of the parity bits of EPCC's tensor
product with a symbol correcting code. In this paper, we show that the proposed
tensor error-pattern correcting code (T-EPCC) is linear time encodable and also
devise a low-complexity soft iterative decoding algorithm for EPCC's tensor
product with q-ary LDPC (T-EPCC-qLDPC). Simulation results show that
T-EPCC-qLDPC achieves almost similar performance to single-level qLDPC with a
1/2 KB sector at 50% reduction in decoding complexity. Moreover, 1 KB
T-EPCC-qLDPC surpasses the performance of 1/2 KB single-level qLDPC at the same
decoder complexity.Comment: Hakim Alhussien, Jaekyun Moon, "An Iteratively Decodable Tensor
Product Code with Application to Data Storage
Error control for reliable digital data transmission and storage systems
A problem in designing semiconductor memories is to provide some measure of error control without requiring excessive coding overhead or decoding time. In LSI and VLSI technology, memories are often organized on a multiple bit (or byte) per chip basis. For example, some 256K-bit DRAM's are organized in 32Kx8 bit-bytes. Byte oriented codes such as Reed Solomon (RS) codes can provide efficient low overhead error control for such memories. However, the standard iterative algorithm for decoding RS codes is too slow for these applications. In this paper we present some special decoding techniques for extended single-and-double-error-correcting RS codes which are capable of high speed operation. These techniques are designed to find the error locations and the error values directly from the syndrome without having to use the iterative alorithm to find the error locator polynomial. Two codes are considered: (1) a d sub min = 4 single-byte-error-correcting (SBEC), double-byte-error-detecting (DBED) RS code; and (2) a d sub min = 6 double-byte-error-correcting (DBEC), triple-byte-error-detecting (TBED) RS code
An Introduction to Quantum Error Correction and Fault-Tolerant Quantum Computation
Quantum states are very delicate, so it is likely some sort of quantum error
correction will be necessary to build reliable quantum computers. The theory of
quantum error-correcting codes has some close ties to and some striking
differences from the theory of classical error-correcting codes. Many quantum
codes can be described in terms of the stabilizer of the codewords. The
stabilizer is a finite Abelian group, and allows a straightforward
characterization of the error-correcting properties of the code. The stabilizer
formalism for quantum codes also illustrates the relationships to classical
coding theory, particularly classical codes over GF(4), the finite field with
four elements. To build a quantum computer which behaves correctly in the
presence of errors, we also need a theory of fault-tolerant quantum
computation, instructing us how to perform quantum gates on qubits which are
encoded in a quantum error-correcting code. The threshold theorem states that
it is possible to create a quantum computer to perform an arbitrary quantum
computation provided the error rate per physical gate or time step is below
some constant threshold value.Comment: 46 pages, with large margins. Includes quant-ph/0004072 plus 30 pages
of new material, mostly on fault-toleranc
ADVANCED SIGNAL PROCESSING FOR MAGNETIC RECORDING ON PERPENDICULARLY MAGNETIZED MEDIA
In magnetic recording channels (MRCs) the readback signal is corrupted by many kinds of impairments, such as electronic noise, media noise, intersymbol interference (ISI), inter-track interference (ITI) and different types of erasures. The growth in demand for the information storage, leads to the continuing pursuit of higher recording density, which enhances the impact of the noise contamination and makes the recovery of the user data from magnetic media more challenging. In this dissertation, we develop advanced signal processing techniques to mitigate these impairments in MRCs.We focus on magnetic recording on perpendicularly magnetized media, from the state-of-the art continuous media to bit-patterned media, which is a possible choice for the next generation of products. We propose novel techniques for soft-input soft-output channel detection, soft iterative decoding of low-density parity-check (LDPC) codes as well as LDPC code designs for MRCs.First we apply the optimal subblock-by-subblock detector (OBBD) to nonbinary LDPC coded perpendicular magnetic recording channels (PMRCs) and derive a symbol-based detector to do the turbo equalization exactly. Second, we propose improved belief-propagation (BP) decoders for both binary and nonbinary LDPC coded PMRCs, which provide significant gains over the standard BP decoder. Third, we introduce novel LDPC code design techniques to construct LDPC codes with fewer short cycles. Performance improvement is achieved by applying the new LDPC codes to PMRCs. Fourth, we do a substantial investigation on Reed-Solomon (RS) plus LDPC coded PMRCs. Finally, we continue our research on bit-patterned magnetic recording (BPMR) channels at extremely high recording densities. A multi-track detection technique is proposed to mitigate the severe ITI in BPMR channels. The multi-track detection with both joint-track and two-dimensional (2D) equalization provide significant performance improvement compared to conventional equalization and detection methods
Study and simulation of low rate video coding schemes
The semiannual report is included. Topics covered include communication, information science, data compression, remote sensing, color mapped images, robust coding scheme for packet video, recursively indexed differential pulse code modulation, image compression technique for use on token ring networks, and joint source/channel coder design
Algebraic Codes For Error Correction In Digital Communication Systems
Access to the full-text thesis is no longer available at the author's request, due to 3rd party copyright restrictions. Access removed on 29.11.2016 by CS (TIS).Metadata merged with duplicate record (http://hdl.handle.net/10026.1/899) on 20.12.2016 by CS (TIS).C. Shannon presented theoretical conditions under which communication was possible
error-free in the presence of noise. Subsequently the notion of using error
correcting codes to mitigate the effects of noise in digital transmission was introduced
by R. Hamming. Algebraic codes, codes described using powerful tools from
algebra took to the fore early on in the search for good error correcting codes. Many
classes of algebraic codes now exist and are known to have the best properties of
any known classes of codes. An error correcting code can be described by three of its
most important properties length, dimension and minimum distance. Given codes
with the same length and dimension, one with the largest minimum distance will
provide better error correction. As a result the research focuses on finding improved
codes with better minimum distances than any known codes.
Algebraic geometry codes are obtained from curves. They are a culmination of years
of research into algebraic codes and generalise most known algebraic codes. Additionally
they have exceptional distance properties as their lengths become arbitrarily
large. Algebraic geometry codes are studied in great detail with special attention
given to their construction and decoding. The practical performance of these codes
is evaluated and compared with previously known codes in different communication
channels. Furthermore many new codes that have better minimum distance
to the best known codes with the same length and dimension are presented from
a generalised construction of algebraic geometry codes. Goppa codes are also an
important class of algebraic codes. A construction of binary extended Goppa codes
is generalised to codes with nonbinary alphabets and as a result many new codes
are found. This construction is shown as an efficient way to extend another well
known class of algebraic codes, BCH codes. A generic method of shortening codes
whilst increasing the minimum distance is generalised. An analysis of this method
reveals a close relationship with methods of extending codes. Some new codes from
Goppa codes are found by exploiting this relationship. Finally an extension method
for BCH codes is presented and this method is shown be as good as a well known
method of extension in certain cases
- …