2,269 research outputs found
Accessible Capacity of Secondary Users
A new problem formulation is presented for the Gaussian interference channels
(GIFC) with two pairs of users, which are distinguished as primary users and
secondary users, respectively. The primary users employ a pair of encoder and
decoder that were originally designed to satisfy a given error performance
requirement under the assumption that no interference exists from other users.
In the scenario when the secondary users attempt to access the same medium, we
are interested in the maximum transmission rate (defined as {\em accessible
capacity}) at which secondary users can communicate reliably without affecting
the error performance requirement by the primary users under the constraint
that the primary encoder (not the decoder) is kept unchanged. By modeling the
primary encoder as a generalized trellis code (GTC), we are then able to treat
the secondary link and the cross link from the secondary transmitter to the
primary receiver as finite state channels (FSCs). Based on this, upper and
lower bounds on the accessible capacity are derived. The impact of the error
performance requirement by the primary users on the accessible capacity is
analyzed by using the concept of interference margin. In the case of
non-trivial interference margin, the secondary message is split into common and
private parts and then encoded by superposition coding, which delivers a lower
bound on the accessible capacity. For some special cases, these bounds can be
computed numerically by using the BCJR algorithm. Numerical results are also
provided to gain insight into the impacts of the GTC and the error performance
requirement on the accessible capacity.Comment: 42 pages, 12 figures, 2 tables; Submitted to IEEE Transactions on
Information Theory on December, 2010, Revised on November, 201
Iterative source and channel decoding relying on correlation modelling for wireless video transmission
Since joint source-channel decoding (JSCD) is capable of exploiting the residual redundancy in the source signals for improving the attainable error resilience, it has attracted substantial attention. Motivated by the principle of exploiting the source redundancy at the receiver, in this treatise we study the application of iterative source channel decoding (ISCD) aided video communications, where the video signal is modelled by a first-order Markov process. Firstly, we derive reduced-complexity formulas for the first-order Markov modelling (FOMM) aided source decoding. Then we propose a bit-based iterative horizontal vertical scanline model (IHVSM) aided source decoding algorithm, where a horizontal and a vertical source decoder are employed for exchanging their extrinsic information using the iterative decoding philosophy. The iterative IHVSM aided decoder is then employed in a forward error correction (FEC) encoded uncompressed video transmission scenario, where the IHVSM and the FEC decoder exchange softbit-information for performing turbo-like ISCD for the sake of improving the reconstructed video quality. Finally, we benchmark the attainable system performance against a near-lossless H.264/AVC video communication system and the existing FOMM based softbit source decoding scheme, where The financial support of the RC-UK under the auspices of the India-UK Advanced Technology Centre (IU-ATC) and that of the EU under the CONCERTO project as well as that of the European Research Council’s Advanced Fellow Grant is gratefully acknowledged. The softbit decoding is performed by a one-dimensional Markov model aided decoder. Our simulation results show that Eb=N0 improvements in excess of 2.8 dB are attainable by the proposed technique in uncompressed video applications
Burst-by-Burst Adaptive Decision Feedback Equalised TCM, TTCM and BICM for H.263-Assisted Wireless Video Telephony
Decision Feedback Equaliser (DFE) aided wideband Burst-by-Burst (BbB) Adaptive Trellis Coded Modulation (TCM), Turbo Trellis Coded Modulation (TTCM) and Bit-Interleaved Coded Modulation (BICM) assisted H.263-based video transceivers are proposed and characterised in performance terms when communicating over the COST 207 Typical Urban wideband fading channel. Specifically, four different modulation modes, namely 4QAM, 8PSK, 16QAM and 64QAM are invoked and protected by the above-mentioned coded modulation schemes. The TTCM assisted scheme was found to provide the best video performance, although at the cost of the highest complexity. A range of lower-complexity arrangements will also be characterised. Finally, in order to confirm these findings in an important practical environment, we have also investigated the adaptive TTCM scheme in the CDMA-based Universal Mobile Telecommunications System's (UMTS) Terrestrial Radio Access (UTRA) scenario and the good performance of adaptive TTCM scheme recorded when communicating over the COST 207 channels was retained in the UTRA environment
Probabilistic Shaping for Finite Blocklengths: Distribution Matching and Sphere Shaping
In this paper, we provide for the first time a systematic comparison of
distribution matching (DM) and sphere shaping (SpSh) algorithms for short
blocklength probabilistic amplitude shaping. For asymptotically large
blocklengths, constant composition distribution matching (CCDM) is known to
generate the target capacity-achieving distribution. As the blocklength
decreases, however, the resulting rate loss diminishes the efficiency of CCDM.
We claim that for such short blocklengths and over the additive white Gaussian
channel (AWGN), the objective of shaping should be reformulated as obtaining
the most energy-efficient signal space for a given rate (rather than matching
distributions). In light of this interpretation, multiset-partition DM (MPDM),
enumerative sphere shaping (ESS) and shell mapping (SM), are reviewed as
energy-efficient shaping techniques. Numerical results show that MPDM and SpSh
have smaller rate losses than CCDM. SpSh--whose sole objective is to maximize
the energy efficiency--is shown to have the minimum rate loss amongst all. We
provide simulation results of the end-to-end decoding performance showing that
up to 1 dB improvement in power efficiency over uniform signaling can be
obtained with MPDM and SpSh at blocklengths around 200. Finally, we present a
discussion on the complexity of these algorithms from the perspective of
latency, storage and computations.Comment: 18 pages, 10 figure
Self-concatenated code design and its application in power-efficient cooperative communications
In this tutorial, we have focused on the design of binary self-concatenated coding schemes with the help of EXtrinsic Information Transfer (EXIT) charts and Union bound analysis. The design methodology of future iteratively decoded self-concatenated aided cooperative communication schemes is presented. In doing so, we will identify the most important milestones in the area of channel coding, concatenated coding schemes and cooperative communication systems till date and suggest future research directions
Vector quantization
During the past ten years Vector Quantization (VQ) has developed from a theoretical possibility promised by Shannon's source coding theorems into a powerful and competitive technique for speech and image coding and compression at medium to low bit rates. In this survey, the basic ideas behind the design of vector quantizers are sketched and some comments made on the state-of-the-art and current research efforts
- …