13 research outputs found
The Road From Classical to Quantum Codes: A Hashing Bound Approaching Design Procedure
Powerful Quantum Error Correction Codes (QECCs) are required for stabilizing
and protecting fragile qubits against the undesirable effects of quantum
decoherence. Similar to classical codes, hashing bound approaching QECCs may be
designed by exploiting a concatenated code structure, which invokes iterative
decoding. Therefore, in this paper we provide an extensive step-by-step
tutorial for designing EXtrinsic Information Transfer (EXIT) chart aided
concatenated quantum codes based on the underlying quantum-to-classical
isomorphism. These design lessons are then exemplified in the context of our
proposed Quantum Irregular Convolutional Code (QIRCC), which constitutes the
outer component of a concatenated quantum code. The proposed QIRCC can be
dynamically adapted to match any given inner code using EXIT charts, hence
achieving a performance close to the hashing bound. It is demonstrated that our
QIRCC-based optimized design is capable of operating within 0.4 dB of the noise
limit
Improved Spectral Bound for Quasi-Cyclic Codes
Spectral bounds form a powerful tool to estimate the minimum distances of
quasi-cyclic codes. They generalize the defining set bounds of cyclic codes to
those of quasi-cyclic codes. Based on the eigenvalues of quasi-cyclic codes and
the corresponding eigenspaces, we provide an improved spectral bound for
quasi-cyclic codes. Numerical results verify that the improved bound
outperforms the Jensen bound in almost all cases. Based on the improved bound,
we propose a general construction of quasi-cyclic codes with excellent designed
minimum distances. For the quasi-cyclic codes produced by this general
construction, the improved spectral bound is always sharper than the Jensen
bound
Codificación para corrección de errores con aplicación en sistemas de transmisión y almacenamiento de información
Tesis (DCI)--FCEFN-UNC, 2013Trata de una técnica de diseño de códigos de chequeo de paridad de baja densidad ( más conocidas por sigla en ingles como LDPC) y un nuevo algoritmo de post- procesamiento para la reducción del piso de erro
Architectures for soft-decision decoding of non-binary codes
En esta tesis se estudia el dise¿no de decodificadores no-binarios para la correcci'on
de errores en sistemas de comunicaci'on modernos de alta velocidad. El objetivo
es proponer soluciones de baja complejidad para los algoritmos de decodificaci'on
basados en los c'odigos de comprobaci'on de paridad de baja densidad no-binarios
(NB-LDPC) y en los c'odigos Reed-Solomon, con la finalidad de implementar arquitecturas
hardware eficientes.
En la primera parte de la tesis se analizan los cuellos de botella existentes en los
algoritmos y en las arquitecturas de decodificadores NB-LDPC y se proponen soluciones
de baja complejidad y de alta velocidad basadas en el volteo de s'¿mbolos.
En primer lugar, se estudian las soluciones basadas en actualizaci'on por inundaci
'on con el objetivo de obtener la mayor velocidad posible sin tener en cuenta la
ganancia de codificaci'on. Se proponen dos decodificadores diferentes basados en
clipping y t'ecnicas de bloqueo, sin embargo, la frecuencia m'axima est'a limitada
debido a un exceso de cableado. Por este motivo, se exploran algunos m'etodos
para reducir los problemas de rutado en c'odigos NB-LDPC. Como soluci'on se
propone una arquitectura basada en difusi'on parcial para algoritmos de volteo
de s'¿mbolos que mitiga la congesti'on por rutado. Como las soluciones de actualizaci
'on por inundaci'on de mayor velocidad son sub-'optimas desde el punto de
vista de capacidad de correci'on, decidimos dise¿nar soluciones para la actualizaci'on
serie, con el objetivo de alcanzar una mayor velocidad manteniendo la ganancia
de codificaci'on de los algoritmos originales de volteo de s'¿mbolo. Se presentan dos
algoritmos y arquitecturas de actualizaci'on serie, reduciendo el 'area y aumentando
de la velocidad m'axima alcanzable. Por 'ultimo, se generalizan los algoritmos de
volteo de s'¿mbolo y se muestra como algunos casos particulares puede lograr una
ganancia de codificaci'on cercana a los algoritmos Min-sum y Min-max con una
menor complejidad. Tambi'en se propone una arquitectura eficiente, que muestra
que el 'area se reduce a la mitad en comparaci'on con una soluci'on de mapeo directo.
En la segunda parte de la tesis, se comparan algoritmos de decodificaci'on Reed-
Solomon basados en decisi'on blanda, concluyendo que el algoritmo de baja complejidad
Chase (LCC) es la soluci'on m'as eficiente si la alta velocidad es el objetivo principal. Sin embargo, los esquemas LCC se basan en la interpolaci'on, que introduce
algunas limitaciones hardware debido a su complejidad. Con el fin de reducir
la complejidad sin modificar la capacidad de correcci'on, se propone un esquema
de decisi'on blanda para LCC basado en algoritmos de decisi'on dura. Por 'ultimo
se dise¿na una arquitectura eficiente para este nuevo esquemaGarcía Herrero, FM. (2013). Architectures for soft-decision decoding of non-binary codes [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/33753TESISPremiad
High Performance Decoder Architectures for Error Correction Codes
Due to the rapid development of the information industry, modern communication and storage systems require much higher data rates and reliability to server various demanding applications. However, these systems suffer from noises from the practical channels. Various error correction codes (ECCs), such as Reed-Solomon (RS) codes, convolutional codes, turbo codes, Low-Density Parity-Check (LDPC) codes and so on, have been adopted in lots of current standards. With the increasing data rate, the research of more advanced ECCs and the corresponding efficient decoders will never stop.Binary LDPC codes have been adopted in lots of modern communication and storage applications due their superior error performance and efficient hardware decoder implementations. Non-binary LDPC (NB-LDPC) codes are an important extension of traditional binary LDPC codes. Compared with its binary counterpart, NB-LDPC codes show better error performance under short to moderate block lengths and higher order modulations. Moreover, NB-LDPC codes have lower error floor than binary LDPC codes. In spite of the excellent error performance, it is hard for current communication and storage systems to adopt NB-LDPC codes due to complex decoding algorithms and decoder architectures. In terms of hardware implementation, current NB-LDPC decoders need much larger area and achieve much lower data throughput.Besides the recently proposed NB-LDPC codes, polar codes, discovered by Ar{\i}kan, appear as a very promising candidate for future communication and storage systems. Polar codes are considered as a major breakthrough in recent coding theory society. Polar codes are proved to be capacity achieving codes over binary input symmetric memoryless channels. Besides, polar codes can be decoded by the successive cancelation (SC) algorithm with of complexity of , where is the block length. The main sticking point of polar codes to date is that their error performance under short to moderate block lengths is inferior compared with LDPC codes or turbo codes. The list decoding technique can be used to improve the error performance of SC algorithms at the cost higher computational and memory complexities. Besides, the hardware implementation of current SC based decoders suffer from long decoding latency which is unsuitable for modern high speed communications.ECCs also find their applications in improving the reliability of network coding. Random linear network coding is an efficient technique for disseminating information in networks, but it is highly susceptible to errors. K\ {o}tter-Kschischang (KK) codes and Mahdavifar-Vardy (MV) codes are two important families of subspace codes that provide error control in noncoherent random linear network coding. List decoding has been used to decode MV codes beyond half distance. Existing hardware implementations of the rank metric decoder for KK codes suffer from limited throughput, long latency and high area complexity. The interpolation-based list decoding algorithm for MV codes still has high computational complexity, and its feasibility for hardware implementations has not been investigated.In this exam, we present efficient decoding algorithms and hardware decoder architectures for NB-LDPC codes, polar codes, KK and MV codes. For NB-LDPC codes, an efficient shuffled decoder architecture is presented to reduce the number of average iterations and improve the throughput. Besides, a fully parallel decoder architecture for NB-LDPC codes with short or moderate block lengths is also presented. Our fully parallel decoder architecture achieves much higher throughput and area efficiency compared with the state-of-art NB-LDPC decoders. For polar codes, a memory efficient list decoder architecture is first presented. Based on our reduced latency list decoding algorithm for polar codes, a high throughput list decoder architecture is also presented. At last, we present efficient decoder architectures for both KK and MV codes