3 research outputs found
On the design of fast and efficient wavelet image coders with reduced memory usage
Image compression is of great importance in multimedia systems and
applications because it drastically reduces bandwidth requirements for
transmission and memory requirements for storage. Although earlier
standards for image compression were based on the Discrete Cosine
Transform (DCT), a recently developed mathematical technique, called
Discrete Wavelet Transform (DWT), has been found to be more efficient
for image coding.
Despite improvements in compression efficiency, wavelet image coders
significantly increase memory usage and complexity when compared with
DCT-based coders. A major reason for the high memory requirements is
that the usual algorithm to compute the wavelet transform requires the
entire image to be in memory. Although some proposals reduce the memory
usage, they present problems that hinder their implementation. In
addition, some wavelet image coders, like SPIHT (which has become a
benchmark for wavelet coding), always need to hold the entire image in
memory.
Regarding the complexity of the coders, SPIHT can be considered quite
complex because it performs bit-plane coding with multiple image scans.
The wavelet-based JPEG 2000 standard is still more complex because it
improves coding efficiency through time-consuming methods, such as an
iterative optimization algorithm based on the Lagrange multiplier
method, and high-order context modeling.
In this thesis, we aim to reduce memory usage and complexity in
wavelet-based image coding, while preserving compression efficiency. To
this end, a run-length encoder and a tree-based wavelet encoder are
proposed. In addition, a new algorithm to efficiently compute the
wavelet transform is presented. This algorithm achieves low memory
consumption using line-by-line processing, and it employs recursion to
automatically place the order in which the wavelet transform is
computed, solving some synchronization problems that have not been
tackled by previous proposals. The proposed encodeOliver Gil, JS. (2006). On the design of fast and efficient wavelet image coders with reduced memory usage [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/1826Palanci
Scalable video compression with optimized visual performance and random accessibility
This thesis is concerned with maximizing the coding efficiency, random accessibility and visual performance of scalable compressed video. The unifying theme behind this work is the use of finely embedded localized coding structures, which govern the extent to which these goals may be jointly achieved.
The first part focuses on scalable volumetric image compression. We investigate 3D transform and coding techniques which exploit inter-slice statistical redundancies without compromising slice accessibility. Our study shows that the motion-compensated temporal discrete wavelet transform (MC-TDWT) practically achieves an upper bound to the compression efficiency of slice transforms. From a video coding perspective, we find that most of the coding gain is attributed to offsetting the learning penalty in adaptive arithmetic coding through 3D code-block extension, rather than inter-frame context modelling.
The second aspect of this thesis examines random accessibility. Accessibility refers to the ease with which a region of interest is accessed (subband samples needed for reconstruction are retrieved) from a compressed video bitstream, subject to spatiotemporal code-block constraints. We investigate the fundamental implications of motion compensation for random access efficiency and the compression performance of scalable interactive video. We demonstrate that inclusion of motion compensation operators within the lifting steps of a temporal subband transform incurs a random access penalty which depends on the characteristics of the motion field.
The final aspect of this thesis aims to minimize the perceptual impact of visible distortion in scalable reconstructed video. We present a visual optimization strategy based on distortion scaling which raises the distortion-length slope of perceptually significant samples. This alters the codestream embedding order during post-compression rate-distortion optimization, thus allowing visually sensitive sites to be encoded with higher fidelity at a given bit-rate.
For visual sensitivity analysis, we propose a contrast perception model that incorporates an adaptive masking slope. This versatile feature provides a context which models perceptual significance. It enables scene structures that otherwise suffer significant degradation to be preserved at lower bit-rates. The novelty in our approach derives from a set of "perceptual mappings" which account for quantization noise shaping effects induced by motion-compensated temporal synthesis. The proposed technique reduces wavelet compression artefacts and improves the perceptual quality of video