717 research outputs found

    Scalable wavelet-based coding of irregular meshes with interactive region-of-interest support

    Get PDF
    This paper proposes a novel functionality in wavelet-based irregular mesh coding, which is interactive region-of-interest (ROI) support. The proposed approach enables the user to define the arbitrary ROIs at the decoder side and to prioritize and decode these regions at arbitrarily high-granularity levels. In this context, a novel adaptive wavelet transform for irregular meshes is proposed, which enables: 1) varying the resolution across the surface at arbitrarily fine-granularity levels and 2) dynamic tiling, which adapts the tile sizes to the local sampling densities at each resolution level. The proposed tiling approach enables a rate-distortion-optimal distribution of rate across spatial regions. When limiting the highest resolution ROI to the visible regions, the fine granularity of the proposed adaptive wavelet transform reduces the required amount of graphics memory by up to 50%. Furthermore, the required graphics memory for an arbitrary small ROI becomes negligible compared to rendering without ROI support, independent of any tiling decisions. Random access is provided by a novel dynamic tiling approach, which proves to be particularly beneficial for large models of over 10(6) similar to 10(7) vertices. The experiments show that the dynamic tiling introduces a limited lossless rate penalty compared to an equivalent codec without ROI support. Additionally, rate savings up to 85% are observed while decoding ROIs of tens of thousands of vertices

    An overview of JPEG 2000

    Get PDF
    JPEG-2000 is an emerging standard for still image compression. This paper provides a brief history of the JPEG-2000 standardization process, an overview of the standard, and some description of the capabilities provided by the standard. Part I of the JPEG-2000 standard specifies the minimum compliant decoder, while Part II describes optional, value-added extensions. Although the standard specifies only the decoder and bitstream syntax, in this paper we describe JPEG-2000 from the point of view of encoding. We take this approach, as we believe it is more amenable to a compact description more easily understood by most readers.

    Adaptation of Zerotrees Using Signed Binary Digit Representations for 3D Image Coding

    Get PDF
    Zerotrees of wavelet coefficients have shown a good adaptability for the compression of three-dimensional images. EZW, the original algorithm using zerotree, shows good performance and was successfully adapted to 3D image compression. This paper focuses on the adaptation of EZW for the compression of hyperspectral images. The subordinate pass is suppressed to remove the necessity to keep the significant pixels in memory. To compensate the loss due to this removal, signed binary digit representations are used to increase the efficiency of zerotrees. Contextual arithmetic coding with very limited contexts is also used. Finally, we show that this simplified version of 3D-EZW performs almost as well as the original one

    Distributed video coding for wireless video sensor networks: a review of the state-of-the-art architectures

    Get PDF
    Distributed video coding (DVC) is a relatively new video coding architecture originated from two fundamental theorems namely, Slepian–Wolf and Wyner–Ziv. Recent research developments have made DVC attractive for applications in the emerging domain of wireless video sensor networks (WVSNs). This paper reviews the state-of-the-art DVC architectures with a focus on understanding their opportunities and gaps in addressing the operational requirements and application needs of WVSNs

    Wavelet-Based Embedded Rate Scalable Still Image Coders: A review

    Get PDF
    Embedded scalable image coding algorithms based on the wavelet transform have received considerable attention lately in academia and in industry in terms of both coding algorithms and standards activity. In addition to providing a very good coding performance, the embedded coder has the property that the bit stream can be truncated at any point and still decodes a reasonably good image. In this paper we present some state-of-the-art wavelet-based embedded rate scalable still image coders. In addition, the JPEG2000 still image compression standard is presented.

    Setting priorities: a new SPIHT-compatible algorithm for image compression

    Get PDF
    We introduce a new algorithm for progressive or multiresolution image compression. The algorithm improves on the Set Partitioning in Hierarchical Trees (SPIHT) algorithm by replacing the SPIHT encoder. The new encoder optimizes the multiresolution code performance relative to a user- defined probability distribution over the code's rates or resolutions. The new algorithm's decoder is identical to the SPIHT decoder. The resulting code achieves the optimal expected performance across resolutions subject to the constraints imposed by the use of the SPIHT decoder and the distribution over resolutions set by the user. The encoder optimization yields performance improvements at the rates or resolutions of greatest importance at the expense of performance degradation at low priority rates or resolutions. The algorithm is fully compatible at the decoder with the original SPIHT algorithm. In particular, the decoder requires no knowledge of the priority function employed at the encoder. Experimental results on an image containing both text and photographic material yield up to 0.86 dB performance improvement over SPIHT at the resolution of highest priority

    An Enhanced Wavelet based Image Compression Technique

    Get PDF
    With the fast expansion of multimedia technologies, the compression of multimedia data has become an important aspect. Image compression is important for efficient storage and transmission of images. The limitation in bandwidth of wireless channels has made data compression a necessity. Wireless channels are bandwidth limited and due to this constraint of wireless channels, progressive image transmission has gained much popularity and acceptance. The Embedded Zerotree Wavelet algorithm (EZW) is based on progressive encoding, in which bits in the bit stream are generated in order of importance. The EZW algorithm, code all the frequency band of wavelet coefficients as the same importance without considering the amount of information in each frequency band. This paper presents an enhanced wavelet based approach to overcome the limitation of the Embedded Zerotree Wavelet (EZW) algorithm. This method divides the image into some sub-blocks

    Efficient compression of motion compensated residuals

    Get PDF
    EThOS - Electronic Theses Online ServiceGBUnited Kingdo

    Embed[d]ed Zerotree Codec

    Get PDF
    This thesis discusses the findings of the final year project involving the VHDL (V= Very High Speed Integrated Circuit, Hardware Description Language) design and simulation of an EZT (Embedded Zero Tree) codec. The basis of image compression and the various image compression techniques that are available today have been explored. This provided a clear understanding of image compression as a whole. An in depth understanding of wavelet transform theory was vital to the understanding of the edge that this transform provides over other transforms for image compression. Both the mathematics of it and how it is implemented using sets of high pass and low pass filters have been studied and presented. At the heart of the EZT codec is the EZW (Embedded Zerotree Wavelet) algorithm, as this is the algorithm that has been implemented in the codec. This required a thorough study and understanding of the algorithm and the various terms used in it. A generic single processor codec capable of handling any size of zerotree coefficients of images was designed. Once the coding and decoding strategy of this single processor had been figured out, it was easily extended to a codec with three parallel processors. This parallel architecture uses the same coding and decoding methods as in the single processor except that each processor in the parallel processing now handles only a third of the coefficients, thus promising a much speedier codec as compared to the first one. Both designs were then translated into VHDL behavioral level codes. The codes were then simulated and the results were verified. Once the simulations were completed the next aim for the project, namely synthesizing the design, was embarked upon. Of the two logical parts of the encoder, only the significance map generator has been synthesized
    corecore