105 research outputs found
Orthogonal Multiple Access with Correlated Sources: Feasible Region and Pragmatic Schemes
In this paper, we consider orthogonal multiple access coding schemes, where
correlated sources are encoded in a distributed fashion and transmitted,
through additive white Gaussian noise (AWGN) channels, to an access point (AP).
At the AP, component decoders, associated with the source encoders, iteratively
exchange soft information by taking into account the source correlation. The
first goal of this paper is to investigate the ultimate achievable performance
limits in terms of a multi-dimensional feasible region in the space of channel
parameters, deriving insights on the impact of the number of sources. The
second goal is the design of pragmatic schemes, where the sources use
"off-the-shelf" channel codes. In order to analyze the performance of given
coding schemes, we propose an extrinsic information transfer (EXIT)-based
approach, which allows to determine the corresponding multi-dimensional
feasible regions. On the basis of the proposed analytical framework, the
performance of pragmatic coded schemes, based on serially concatenated
convolutional codes (SCCCs), is discussed
Myths and Realities of Rateless Coding
Fixed-rate and rateless channel codes are generally treated separately in the related research literature and so, a novice in the field inevitably gets the impression that these channel codes are unrelated. By contrast, in this treatise, we endeavor to further develop a link between the traditional fixed-rate codes and the recently developed rateless codes by delving into their underlying attributes. This joint treatment is beneficial for two principal reasons. First, it facilitates the task of researchers and practitioners, who might be familiar with fixed-rate codes and would like to jump-start their understanding of the recently developed concepts in the rateless reality. Second, it provides grounds for extending the use of the well-understood code design tools — originally contrived for fixed-rate codes — to the realm of rateless codes. Indeed, these versatile tools proved to be vital in the design of diverse fixed-rate-coded communications systems, and thus our hope is that they will further elucidate the associated performance ramifications of the rateless coded schemes
Network flow algorithms for wireless networks and design and analysis of rate compatible LDPC codes
While Shannon already characterized the capacity of point-to-point channels back in 1948, characterizing the capacity of wireless networks has been a challenging problem. The deterministic channel model proposed by Avestimehr, etc. (2007 - 1) has been a promising approach
for approximating the Gaussian channel capacity and has been widely studied recently. Motivated by this model, an improved combinatorial algorithm is considered for finding the unicast capacity for wireless information flow on such deterministic networks in the first part of this
thesis. Our algorithm fully explores the useful combinatorial features intrinsic in the problem. Our improvement applies generally with any size of finite fields associated with the channel model. Comparing with other related algorithms, our improved algorithm has very competitive
performance in complexity.
In the second part of our work, we consider the design and analysis of rate-compatible LDPC codes. Rate-compatible LDPC codes are basically a family of nested codes, operating at different code rates and all of them can be encoded and decoded using a single encoder and
decoder pair. Those properties make rate-compatible LDPC codes a good choice for changing channel conditions, like in wireless communications. The previous work on the design and analysis of LDPC codes are all targeting at a specific code rate and no work is known on
the design and analysis of rate-compatible LDPC codes so that the code performance at all code rates in the family is manageable and predictable. In our work, we proposed algorithms for the design and analysis of rate-compatible LDPC codes with good performance and make the code performance at all code rates manageable and predictable. Our work is based on E2RC codes, while our approaches in the design and analysis can be applied more generally not only to E2RC codes, but to other suitable scenarios, like the design of IRA codes. Most encouragingly, we obtain families of rate-compatible codes whose gaps to capacity are at most 0.3 dB across the range of rates when the maximum variable node degree is twenty, which is very promising compared with other existing results
Self-concatenated code design and its application in power-efficient cooperative communications
In this tutorial, we have focused on the design of binary self-concatenated coding schemes with the help of EXtrinsic Information Transfer (EXIT) charts and Union bound analysis. The design methodology of future iteratively decoded self-concatenated aided cooperative communication schemes is presented. In doing so, we will identify the most important milestones in the area of channel coding, concatenated coding schemes and cooperative communication systems till date and suggest future research directions
Novel LDPC coding and decoding strategies: design, analysis, and algorithms
In this digital era, modern communication systems play an essential part in nearly every aspect of life, with examples ranging from mobile networks and satellite communications to Internet and data transfer. Unfortunately, all communication systems in a practical setting are noisy, which indicates that we can either improve the physical characteristics of the channel or find a possible systematical solution, i.e. error control coding. The history of error control coding dates back to 1948 when Claude Shannon published his celebrated work “A Mathematical Theory of Communication”, which built a framework for channel coding, source coding and information theory. For the first time, we saw evidence for the existence of channel codes, which enable reliable communication as long as the information rate of the code does not surpass the so-called channel capacity. Nevertheless, in the following 60 years none of the codes have been proven closely to approach the theoretical bound until the arrival of turbo codes and the renaissance of LDPC codes. As a strong contender of turbo codes, the advantages of LDPC codes include parallel implementation of decoding algorithms and, more crucially, graphical construction of codes. However, there are also some drawbacks to LDPC codes, e.g. significant performance degradation due to the presence of short cycles or very high decoding latency. In this thesis, we will focus on the practical realisation of finite-length LDPC codes and devise algorithms to tackle those issues.
Firstly, rate-compatible (RC) LDPC codes with short/moderate block lengths are investigated on the basis of optimising the graphical structure of the tanner graph (TG), in order to achieve a variety of code rates (0.1 < R < 0.9) by only using a single encoder-decoder pair. As is widely recognised in the literature, the presence of short cycles considerably reduces the overall performance of LDPC codes which significantly limits their application in communication systems. To reduce the impact of short cycles effectively for different code rates, algorithms for counting short cycles and a graph-related metric called Extrinsic Message Degree (EMD) are applied with the development of the proposed puncturing and extension techniques. A complete set of simulations are carried out to demonstrate that the proposed RC designs can largely minimise the performance loss caused by puncturing or extension.
Secondly, at the decoding end, we study novel decoding strategies which compensate for the negative effect of short cycles by reweighting part of the extrinsic messages exchanged between the nodes of a TG. The proposed reweighted belief propagation (BP) algorithms aim to implement efficient decoding, i.e. accurate signal reconstruction and low decoding latency, for LDPC codes via various design methods. A variable factor appearance probability belief propagation (VFAP-BP) algorithm is proposed along with an improved version called a locally-optimized reweighted (LOW)-BP algorithm, both of which can be employed to enhance decoding performance significantly for regular and irregular LDPC codes. More importantly, the optimisation of reweighting parameters only takes place in an offline stage so that no additional computational complexity is required during the real-time decoding process.
Lastly, two iterative detection and decoding (IDD) receivers are presented for multiple-input multiple-output (MIMO) systems operating in a spatial multiplexing configuration. QR decomposition (QRD)-type IDD receivers utilise the proposed multiple-feedback (MF)-QRD or variable-M (VM)-QRD detection algorithm with a standard BP decoding algorithm, while knowledge-aided (KA)-type receivers are equipped with a simple soft parallel interference cancellation (PIC) detector and the proposed reweighted BP decoders. In the uncoded scenario, the proposed MF-QRD and VM-QRD algorithms are shown to approach optimal performance, yet require a reduced computational complexity. In the LDPC-coded scenario, simulation results have illustrated that the proposed QRD-type IDD receivers can offer near-optimal performance after a small number of detection/decoding iterations and the proposed KA-type IDD receivers significantly outperform receivers using alternative decoding algorithms, while requiring similar decoding complexity
Constellation Shaping for WDM systems using 256QAM/1024QAM with Probabilistic Optimization
In this paper, probabilistic shaping is numerically and experimentally
investigated for increasing the transmission reach of wavelength division
multiplexed (WDM) optical communication system employing quadrature amplitude
modulation (QAM). An optimized probability mass function (PMF) of the QAM
symbols is first found from a modified Blahut-Arimoto algorithm for the optical
channel. A turbo coded bit interleaved coded modulation system is then applied,
which relies on many-to-one labeling to achieve the desired PMF, thereby
achieving shaping gain. Pilot symbols at rate at most 2% are used for
synchronization and equalization, making it possible to receive input
constellations as large as 1024QAM. The system is evaluated experimentally on a
10 GBaud, 5 channels WDM setup. The maximum system reach is increased w.r.t.
standard 1024QAM by 20% at input data rate of 4.65 bits/symbol and up to 75% at
5.46 bits/symbol. It is shown that rate adaptation does not require changing of
the modulation format. The performance of the proposed 1024QAM shaped system is
validated on all 5 channels of the WDM signal for selected distances and rates.
Finally, it was shown via EXIT charts and BER analysis that iterative
demapping, while generally beneficial to the system, is not a requirement for
achieving the shaping gain.Comment: 10 pages, 12 figures, Journal of Lightwave Technology, 201
Spatially Coupled Turbo-Like Codes
The focus of this thesis is on proposing and analyzing a powerful class of codes on graphs---with trellis constraints---that can simultaneously approach capacity and achieve very low error floor. In particular, we propose the concept of spatial coupling for turbo-like code (SC-TC) ensembles and investigate the impact of coupling on the performance of these codes. The main elements of this study can be summarized by the following four major topics. First, we considered the spatial coupling of parallel concatenated codes (PCCs), serially concatenated codes (SCCs), and hybrid concatenated codes (HCCs).We also proposed two extensions of braided convolutional codes (BCCs) to higher coupling memories. Second, we investigated the impact of coupling on the asymptotic behavior of the proposed ensembles in term of the decoding thresholds. For that, we derived the exact density evolution (DE) equations of the proposed SC-TC ensembles over the binary erasure channel. Using the DE equations, we found the thresholds of the coupled and uncoupled ensembles under belief propagation (BP) decoding for a wide range of rates. We also computed the maximum a-posteriori (MAP) thresholds of the underlying uncoupled ensembles. Our numerical results confirm that TCs have excellent MAP thresholds, and for a large enough coupling memory, the BP threshold of an SC-TC ensemble improves to the MAP threshold of the underlying TC ensemble. This phenomenon is called threshold saturation and we proved its occurrence for SC-TCs by use of a proof technique based on the potential function of the ensembles.Third, we investigated and discussed the performance of SC-TCs in the finite length regime. We proved that under certain conditions the minimum distance of an SC-TCs is either larger or equal to that of its underlying uncoupled ensemble. Based on this fact, we performed a weight enumerator (WE) analysis for the underlying uncoupled ensembles to investigate the error floor performance of the SC-TC ensembles. We computed bounds on the error rate performance and minimum distance of the TC ensembles. These bounds indicate very low error floor for SCC, HCC, and BCC ensembles, and show that for HCC, and BCC ensembles, the minimum distance grows linearly with the input block length.The results from the DE and WE analysis demonstrate that the performance of TCs benefits from spatial coupling in both waterfall and error floor regions. While uncoupled TC ensembles with close-to-capacity performance exhibit a high error floor, our results show that SC-TCs can simultaneously approach capacity and achieve very low error floor.Fourth, we proposed a unified ensemble of TCs that includes all the considered TC classes. We showed that for each of the original classes of TCs, it is possible to find an equivalent ensemble by proper selection of the design parameters in the unified ensemble. This unified ensemble not only helps us to understand the connections and trade-offs between the TC ensembles but also can be considered as a bridge between TCs and generalized low-density parity check codes
On the Convergence Speed of Turbo Demodulation with Turbo Decoding
Iterative processing is widely adopted nowadays in modern wireless receivers
for advanced channel codes like turbo and LDPC codes. Extension of this
principle with an additional iterative feedback loop to the demapping function
has proven to provide substantial error performance gain. However, the adoption
of iterative demodulation with turbo decoding is constrained by the additional
implied implementation complexity, heavily impacting latency and power
consumption. In this paper, we analyze the convergence speed of these combined
two iterative processes in order to determine the exact required number of
iterations at each level. Extrinsic information transfer (EXIT) charts are used
for a thorough analysis at different modulation orders and code rates. An
original iteration scheduling is proposed reducing two demapping iterations
with reasonable performance loss of less than 0.15 dB. Analyzing and
normalizing the computational and memory access complexity, which directly
impact latency and power consumption, demonstrates the considerable gains of
the proposed scheduling and the promising contributions of the proposed
analysis.Comment: Submitted to IEEE Transactions on Signal Processing on April 27, 201
- …