888 research outputs found

    Optimized Scalable Image and Video Transmission for MIMO Wireless Channels

    Get PDF
    In this chapter, we focus on proposing new strategies to efficiently transfer a compressed image/video content through wireless links using a multiple antenna technology. The proposed solutions can be considered as application layer physical layer (APP-PHY) cross layer design methods as they involve optimizing both application and physical layers. After a wide state-of-the-art study, we present two main solutions. The first focuses on using a new precoding algorithm that takes into account the image/video content structure when assigning transmission powers. We showed that its results are better than the existing conventional precoders. Second, a link adaptation process is integrated to efficiently assign coding parameters as a function of the channel state. Simulations over a realistic channel environment show that the link adaptation activates a dynamic process that results in a good image/video reconstruction quality even if the channel is varying. Finally, we incorporated soft decoding algorithms at the receiver side, and we showed that they could induce further improvements. In fact, almost 5 dB peak signal-to-noise ratio (PSNR) improvements are demonstrated in the case of transmission over a Rayleigh channel

    Self-concatenated code design and its application in power-efficient cooperative communications

    No full text
    In this tutorial, we have focused on the design of binary self-concatenated coding schemes with the help of EXtrinsic Information Transfer (EXIT) charts and Union bound analysis. The design methodology of future iteratively decoded self-concatenated aided cooperative communication schemes is presented. In doing so, we will identify the most important milestones in the area of channel coding, concatenated coding schemes and cooperative communication systems till date and suggest future research directions

    Advanced Modulation and Coding Technology Conference

    Get PDF
    The objectives, approach, and status of all current LeRC-sponsored industry contracts and university grants are presented. The following topics are covered: (1) the LeRC Space Communications Program, and Advanced Modulation and Coding Projects; (2) the status of four contracts for development of proof-of-concept modems; (3) modulation and coding work done under three university grants, two small business innovation research contracts, and two demonstration model hardware development contracts; and (4) technology needs and opportunities for future missions

    VLSI architectures design for encoders of High Efficiency Video Coding (HEVC) standard

    Get PDF
    The growing popularity of high resolution video and the continuously increasing demands for high quality video on mobile devices are producing stronger needs for more efficient video encoder. Concerning these desires, HEVC, a newest video coding standard, has been developed by a joint team formed by ISO/IEO MPEG and ITU/T VCEG. Its design goal is to achieve a 50% compression gain over its predecessor H.264 with an equal or even higher perceptual video quality. Motion Estimation (ME) being as one of the most critical module in video coding contributes almost 50%-70% of computational complexity in the video encoder. This high consumption of the computational resources puts a limit on the performance of encoders, especially for full HD or ultra HD videos, in terms of coding speed, bit-rate and video quality. Thus the major part of this work concentrates on the computational complexity reduction and improvement of timing performance of motion estimation algorithms for HEVC standard. First, a new strategy to calculate the SAD (Sum of Absolute Difference) for motion estimation is designed based on the statistics on property of pixel data of video sequences. This statistics demonstrates the size relationship between the sum of two sets of pixels has a determined connection with the distribution of the size relationship between individual pixels from the two sets. Taking the advantage of this observation, only a small proportion of pixels is necessary to be involved in the SAD calculation. Simulations show that the amount of computations required in the full search algorithm is reduced by about 58% on average and up to 70% in the best case. Secondly, from the scope of parallelization an enhanced TZ search for HEVC is proposed using novel schemes of multiple MVPs (motion vector predictor) and shared MVP. Specifically, resorting to multiple MVPs the initial search process is performed in parallel at multiple search centers, and the ME processing engine for PUs within one CU are parallelized based on the MVP sharing scheme on CU (coding unit) level. Moreover, the SAD module for ME engine is also parallelly implemented for PU size of 32×32. Experiments indicate it achieves an appreciable improvement on the throughput and coding efficiency of the HEVC video encoder. In addition, the other part of this thesis is contributed to the VLSI architecture design for finding the first W maximum/minimum values targeting towards high speed and low hardware cost. The architecture based on the novel bit-wise AND scheme has only half of the area of the best reference solution and its critical path delay is comparable with other implementations. While the FPCG (full parallel comparison grid) architecture, which utilizes the optimized comparator-based structure, achieves 3.6 times faster on average on the speed and even 5.2 times faster at best comparing with the reference architectures. Finally the architecture using the partial sorting strategy reaches a good balance on the timing performance and area, which has a slightly lower or comparable speed with FPCG architecture and a acceptable hardware cost

    On hard-decision forward error correction with application to high-throughput fiber-optic communications

    Get PDF
    The advent of the Internet not only changed the communication methods significantly, but also the life-style of the human beings. The number of Internet users has grown exponentially in the last decade, and the number of users exceeded 3.4 billion in 2016. Fiber links serve as the Internet backbone, hence, the fast grow of the Internet network and the sheer of new applications is highly driven by advances in optical communications. The emergence of coherent optical systems has led to a more efficient use of the available spectrum compared to traditional on-off keying transmission, and has made it possible to increase the supported data rates. To achieve high spectral efficiencies and improve the transmission reach, coding in combination with a higher order modulation, a scheme known as coded modulation (CM), has become indispensable in fiber-optic communications. In the recent years, graph-based codes such as low-density parity-check codes and soft decision decoding (SDD) have been adopted for long-haul coherent optical systems. SDD yields very high net coding gains but at the expense of a relatively high decoding complexity, which brings implementation challenges at very high data rates. Hard decision decoding (HDD) is an appealing alternative that reduces the decoding complexity. This motivates the focus of this thesis on forward error correction (FEC) with HDD for high-throughput, low power fiber-optic communications.In this thesis, we start by studying the performance bounds of HDD. In particular, we derive achievable information rates (AIRs) for CM with HDD for both bit-wise and symbol-wise decoding, and show that bit-wise HDD yields significantly higher AIRs. We also design nonbinary staircase codes using density evolution. Finite length simulation results of binary and nonbinary staircase codes corroborate the conclusions arising from the AIR analysis, i.e., for HDD binary codes are preferable. Then, we consider probabilistic shaping. In particular, we extend the probabilistic amplitude shaping (PAS) scheme recently introduced by B\uf6cherer et al. to HDD based on staircase codes. Finally, we focus on new decoding algorithms for product-like codes to close the gap between HDD and SDD, while keeping the decoding complexity low. In particular, we propose three novel decoding algorithms for product-like codes based on assisting the HDD with some level of soft information. The proposed algorithms provide a clear performance-complexity tradeoff. In particular, we show that up to roughly half of the gap between SDD and HDD can be closed with limited complexity increase with respect to HDD

    Publications of the Jet Propulsion Laboratory 1976

    Get PDF
    The formalized technical reporting, released January through December 1975, that resulted from scientific and engineering work performed, or managed, by the Jet Propulsion Laboratory is described and indexed. The following classes of publications are included: (1) technical reports; (2) technical memorandums; (3) articles from bi-monthly Deep Space Network (DSN) progress report; (4) special publications; and (5) articles published in the open literature. The publications are indexed by: (1) author, (2) subject, and (3) publication type and number. A descriptive entry appears under the name of each author of each publication; an abstract is included with the entry for the primary (first-listed) author. Unless designated otherwise, all publications listed are unclassified

    Design of large polyphase filters in the Quadratic Residue Number System

    Full text link

    Trade-off analysis of modes of data handling for earth resources (ERS), volume 1

    Get PDF
    Data handling requirements are reviewed for earth observation missions along with likely technology advances. Parametric techniques for synthesizing potential systems are developed. Major tasks include: (1) review of the sensors under development and extensions of or improvements in these sensors; (2) development of mission models for missions spanning land, ocean, and atmosphere observations; (3) summary of data handling requirements including the frequency of coverage, timeliness of dissemination, and geographic relationships between points of collection and points of dissemination; (4) review of data routing to establish ways of getting data from the collection point to the user; (5) on-board data processing; (6) communications link; and (7) ground data processing. A detailed synthesis of three specific missions is included

    Progressive Source-Channel Coding for Multimedia Transmission over Noisy and Lossy Channels with and without Feedback

    Get PDF
    Rate-scalable or layered lossy source-coding is useful for progressive transmission of multimedia sources, where the receiver can reconstruct the source incrementally. This thesis considers ``joint source-channel'' schemes for such a progressive transmission, in the presence of noise or loss, with and without the use of a feedback link. First we design image communication schemes for memoryless and finite state channels using limited and explicitly constrained use of the feedback channel in the form of a variable incremental redundancy Hybrid ARQ protocol. Constraining feedback allows a direct comparison with schemes without feedback. Optimized feedback based systems are shown to have useful gains. Second, we develop a controlled Markov chain approach for constrained feedback Hybrid ARQ protocol design. The proposed methodology allows the protocol to be chosen from a collection of signal flow graphs, and also allows explicit control over the tradeoffs in throughput, reliability and complexity. Next we consider progressive image transmission in the absence of feedback. We assign unequal error protection to the bits of a rate-scalable source-coder using rate compatible channel codes. We show that, under the framework, the source and channel bits can be ``scheduled'' in a single bitstream in such a way that operational optimality is retained for different transmission budgets, creating a rate-scalable joint source-channel coder. Next we undertake the design of a joint source-channel decoder that uses ``distortion aware'' ACK/NACK feedback generation. For memoryless channels, and Type-I HARQ, the design of optimal ACK/NACK generation and decoding by packet combining is cast and solved as a sequential decision problem. We obtain dynamic programming based optimal solutions and also propose suboptimal, lower complexity distortion-aware decoders and feedback generation rules which outperform conventional BER based rules such as CRC-check. Finally we design operational rate-distortion optimal ACK/NACK feedback generation rules for transmitting a tree structured quantizer over a memoryless channel. We show that the optimal feedback generation rules are embedded, that is, they allow incremental switching to higher rates during the transmission. Also, we obtain the structure of the feedback generation rules in terms of a feedback threshold function that simplifies the implementation
    corecore