3 research outputs found
2-step scalar deadzone quantization for bitplane image coding
Modern lossy image coding systems generate a quality progressive codestream that, truncated at increasing rates, produces an image with decreasing distortion. Quality progressivity is commonly provided by an embedded quantizer that employs uniform scalar deadzone quantization (USDQ) together with a bitplane coding strategy. This paper introduces a 2-step scalar deadzone quantization (2SDQ) scheme that achieves same coding performance as that of USDQ while reducing the coding passes and the emitted symbols of the bitplane coding engine. This serves to reduce the computational costs of the codec and/or to code high dynamic range images. The main insights behind 2SDQ are the use of two quantization step sizes that approximate wavelet coefficients with more or less precision depending on their density, and a rate-distortion optimization technique that adjusts the distortion decreases produced when coding 2SDQ indexes. The integration of 2SDQ in current codecs is straightforward. The applicability and efficiency of 2SDQ are demonstrated within the framework of JPEG2000
On the design of fast and efficient wavelet image coders with reduced memory usage
Image compression is of great importance in multimedia systems and
applications because it drastically reduces bandwidth requirements for
transmission and memory requirements for storage. Although earlier
standards for image compression were based on the Discrete Cosine
Transform (DCT), a recently developed mathematical technique, called
Discrete Wavelet Transform (DWT), has been found to be more efficient
for image coding.
Despite improvements in compression efficiency, wavelet image coders
significantly increase memory usage and complexity when compared with
DCT-based coders. A major reason for the high memory requirements is
that the usual algorithm to compute the wavelet transform requires the
entire image to be in memory. Although some proposals reduce the memory
usage, they present problems that hinder their implementation. In
addition, some wavelet image coders, like SPIHT (which has become a
benchmark for wavelet coding), always need to hold the entire image in
memory.
Regarding the complexity of the coders, SPIHT can be considered quite
complex because it performs bit-plane coding with multiple image scans.
The wavelet-based JPEG 2000 standard is still more complex because it
improves coding efficiency through time-consuming methods, such as an
iterative optimization algorithm based on the Lagrange multiplier
method, and high-order context modeling.
In this thesis, we aim to reduce memory usage and complexity in
wavelet-based image coding, while preserving compression efficiency. To
this end, a run-length encoder and a tree-based wavelet encoder are
proposed. In addition, a new algorithm to efficiently compute the
wavelet transform is presented. This algorithm achieves low memory
consumption using line-by-line processing, and it employs recursion to
automatically place the order in which the wavelet transform is
computed, solving some synchronization problems that have not been
tackled by previous proposals. The proposed encodeOliver Gil, JS. (2006). On the design of fast and efficient wavelet image coders with reduced memory usage [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/1826Palanci