26 research outputs found
Performance of RS and BCH Codes over Correlated Rayleigh Fading Channel using QAM Modulation Technique
The quest for a reliable reception of the replica of a transmitted signal over wireless medium has become very important in the transmission and reception of information. In order to mitigate the effect of the degradation of the transmitted signal, error correction techniques are employed to reliably recover the erroneous bits inherent especially when the channel is in deep fading. In this paper, the performance of Reed Solomon (RS) and Bose-Chaudhuri-Hochquenghem (BCH) codes over correlated Rayleigh fading is presented. This work also investigated the effect of error correction code in the variation of the modulation order of the QAM. The simulation was carried out by encoding randomly generated data using both RS and BCH codes and then modulated by varying the modulation order of the quadrature amplitude modulation before transmitted over a correlated Rayleigh distributed channel. At the receiver, the received signal was demodulated and decoded. The bits in error received were detected and corrected to retrieve back the original transmitted signal. The system was evaluated in bit error rate and the result showed that BCH code performs better for higher modulation order. Keywords: Correlated channel, Rayleigh fading, M-QAM, RS code, BCH code
Book Review: A Conceptual Review of “Digital Communication Systems”
(Author: Simon Haykin, 2014)Haykin, S. 2014. Digital Communication Systems.John Wiley & Sons, Inc., Hoboken, NJ, USA.Available: <http://www.wiley.com/WileyCDA/WileyTitle/productCd-EHEP001809.html>
Improved quality block-based low bit rate video coding.
The aim of this research is to develop algorithms for enhancing the subjective quality and coding efficiency of standard block-based video coders. In the past few years, numerous video coding standards based on motion-compensated block-transform structure have been established where block-based motion estimation is used for reducing the correlation between consecutive images and block transform is used for coding the resulting motion-compensated residual images. Due to the use of predictive differential coding and variable length coding techniques, the output data rate exhibits extreme fluctuations. A rate control algorithm is devised for achieving a stable output data rate. This rate control algorithm, which is essentially a bit-rate estimation algorithm, is then employed in a bit-allocation algorithm for improving the visual quality of the coded images, based on some prior knowledge of the images. Block-based hybrid coders achieve high compression ratio mainly due to the employment of a motion estimation and compensation stage in the coding process. The conventional bit-allocation strategy for these coders simply assigns the bits required by the motion vectors and the rest to the residual image. However, at very low bit-rates, this bit-allocation strategy is inadequate as the motion vector bits takes up a considerable portion of the total bit-rate. A rate-constrained selection algorithm is presented where an analysis-by-synthesis approach is used for choosing the best motion vectors in term of resulting bit rate and image quality. This selection algorithm is then implemented for mode selection. A simple algorithm based on the above-mentioned bit-rate estimation algorithm is developed for the latter to reduce the computational complexity. For very low bit-rate applications, it is well-known that block-based coders suffer from blocking artifacts. A coding mode is presented for reducing these annoying artifacts by coding a down-sampled version of the residual image with a smaller quantisation step size. Its applications for adaptive source/channel coding and for coding fast changing sequences are examined
Near far resistant detection for CDMA personal communication systems.
The growth of Personal Communications, the keyword of the 90s, has already the signs of a technological revolution. The foundations of this revolution are currently set through the standardization of the Universal Mobile Telecommunication System (UMTS), a communication system with synergistic terrestrial and satellite segments. The main characteristic of the UMTS radio interface, is the provision of ISDN services. Services with higher than voice data rates require more spectrum, thus techniques that utilize spectrum as efficiently as possible are currently at the forefront of the research community interests. Two of the most spectrally efficient multiple access technologies, namely. Code Division Multiple Access (CDMA) and Time Division Multiple Access (TDMA) concentrate the efforts of the European telecommunity.This thesis addresses problems and. proposes solutions for CDMA systems that must comply with the UMTS requirements. Prompted by Viterbi's call for further extending the potential of CDMA through signal processing at the receiving end, we propose new Minimum Mean Square Error receiver architectures. MMSE detection schemes offer significant advantages compared to the conventional correlation based receivers as they are NEar FAr Resistant (NEFAR) over a wide range of interfering power levels. The NEFAR characteristic of these detectors reduces considerably the requirements of the power control loops currently found in commercial CDMA systems. MMSE detectors are also found, to have significant performance gains over other well established interference cancellation techniques like the decorrelating detector, especially in heavily loaded system conditions. The implementation architecture of MMSE receivers can be either Multiple-Input Multiple Output (MIMO) or Single-Input Single-Output. The later offers not only complexity that is comparable to the conventional detector, but also has the inherent advantage of employing adaptive algorithms which can be used to provide both the dispreading and the interference cancellation function, without the knowledge of the codes of interfering users. Furthermore, in multipath fading channels, adaptive MMSE detectors can exploit the multipath diversity acting as RAKE combiners. The later ability is distinctive to MMSE based receivers, and it is achieved in an autonomous fashion, without the knowledge of the multipath intensity profile. The communicator achieves its performance objectives by the synergy of the signal processor and the channel decoder. According to the propositions of this thesis, the form of the signal processor needs to be changed, in order to exploit the horizons of spread spectrum signaling. However, maximum likelihood channel decoding algorithms need not change. It is the way that these algorithms are utilized that needs to be revis ed. In this respect, we identify three major utilization scenarios and an attempt is made to quantify which of the three best matches the requirements of a UMTS oriented CDMA radio interface. Based on our findings, channel coding can be used as a mapping technique from the information bit to a more ''intelligent" chip, matching the ''intelligence" of the signal processor
Introduction to Forward-Error-Correcting Coding
This reference publication introduces forward error correcting (FEC) and stresses definitions and basic calculations for use by engineers. The seven chapters include 41 example problems, worked in detail to illustrate points. A glossary of terms is included, as well as an appendix on the Q function. Block and convolutional codes are covered
Recommended from our members
Detection and reconstruction of error control codes for engineered and biological regulatory systems.
A fundamental challenge for all communication systems, engineered or living, is the problem of achieving efficient, secure, and error-free communication over noisy channels. Information theoretic principals have been used to develop effective coding theory algorithms to successfully transmit information in engineering systems. Living systems also successfully transmit biological information through genetic processes such as replication, transcription, and translation, where the genome of an organism is the contents of the transmission. Decoding of received bit streams is fairly straightforward when the channel encoding algorithms are efficient and known. If the encoding scheme is unknown or part of the data is missing or intercepted, how would one design a viable decoder for the received transmission? For such systems blind reconstruction of the encoding/decoding system would be a vital step in recovering the original message. Communication engineers may not frequently encounter this situation, but for computational biologists and biotechnologist this is an immediate challenge. The goal of this work is to develop methods for detecting and reconstructing the encoder/decoder system for engineered and biological data. Building on Sandia's strengths in discrete mathematics, algorithms, and communication theory, we use linear programming and will use evolutionary computing techniques to construct efficient algorithms for modeling the coding system for minimally errored engineered data stream and genomic regulatory DNA and RNA sequences. The objective for the initial phase of this project is to construct solid parallels between biological literature and fundamental elements of communication theory. In this light, the milestones for FY2003 were focused on defining genetic channel characteristics and providing an initial approximation for key parameters, including coding rate, memory length, and minimum distance values. A secondary objective addressed the question of determining similar parameters for a received, noisy, error-control encoded data set. In addition to these goals, we initiated exploration of algorithmic approaches to determine if a data set could be approximated with an error-control code and performed initial investigations into optimization based methodologies for extracting the encoding algorithm given the coding rate of an encoded noise-free and noisy data stream