16 research outputs found
An optimal design procedure for intraband vector quantized subband coding
Journal ArticleAbstTact- Subband coding with vector quantization is addressed in this paper. Forming the data vectors from both between and within the subbands is considered. The former of these two schemes is referred to as interband coding and the latter as intraband coding. Interband coder design is relatively straightforward since the design of the single codebook involved follows readily from a representative set of interband data vectors. Intraband coder design is more complicated since it entails the selection of a vector dimension and a bit-rate for each subband. The main contribution of this work is an optimal methodology for intraband subband vector quantizer design. The problem formulation includes constraints on the bit-rate and the encoding complexity and is solved with nonlinear programming methods. Subband vector quantization image coding in conjunction with a human visual system model is thoroughly investigated. Results of a large number of experiments indicate that the optimal intraband coder yields superior results from quantitative as well as subjective points of view than the interband coder for comparable bit-rates. This improvement becomes more pronounced as the computational complexity of the intraband encoder is allowed to increase
Applications of wavelet-based compression to multidimensional Earth science data
A data compression algorithm involving vector quantization (VQ) and the discrete wavelet transform (DWT) is applied to two different types of multidimensional digital earth-science data. The algorithms (WVQ) is optimized for each particular application through an optimization procedure that assigns VQ parameters to the wavelet transform subbands subject to constraints on compression ratio and encoding complexity. Preliminary results of compressing global ocean model data generated on a Thinking Machines CM-200 supercomputer are presented. The WVQ scheme is used in both a predictive and nonpredictive mode. Parameters generated by the optimization algorithm are reported, as are signal-to-noise (SNR) measurements of actual quantized data. The problem of extrapolating hydrodynamic variables across the continental landmasses in order to compute the DWT on a rectangular grid is discussed. Results are also presented for compressing Landsat TM 7-band data using the WVQ scheme. The formulation of the optimization problem is presented along with SNR measurements of actual quantized data. Postprocessing applications are considered in which the seven spectral bands are clustered into 256 clusters using a k-means algorithm and analyzed using the Los Alamos multispectral data analysis program, SPECTRUM, both before and after being compressed using the WVQ program
Discrete multitone modulation with principal component filter banks
Discrete multitone (DMT) modulation is an attractive method for communication over a nonflat channel with possibly colored noise. The uniform discrete Fourier transform (DFT) filter bank and cosine modulated filter bank have in the past been used in this system because of low complexity. We show in this paper that principal component filter banks (PCFB) which are known to be optimal for data compression and denoising applications, are also optimal for a number of criteria in DMT modulation communication. For example, the PCFB of the effective channel noise power spectrum (noise psd weighted by the inverse of the channel gain) is optimal for DMT modulation in the sense of maximizing bit rate for fixed power and error probabilities. We also establish an optimality property of the PCFB when scalar prefilters and postfilters are used around the channel. The difference between the PCFB and a traditional filter bank such as the brickwall filter bank or DFT filter bank is significant for effective power spectra which depart considerably from monotonicity. The twisted pair channel with its bridged taps, next and fext noises, and AM interference, therefore appears to be a good candidate for the application of a PCFB. This is demonstrated with the help of numerical results for the case of the ADSL channel
Parental finite state vector quantizer and vector wavelet transform-linear predictive coding.
by Lam Chi Wah.Thesis submitted in: December 1997.Thesis (M.Phil.)--Chinese University of Hong Kong, 1998.Includes bibliographical references (leaves 89-91).Abstract also in Chinese.Chapter Chapter 1 --- Introduction to Data Compression and Image Coding --- p.1Chapter 1.1 --- Introduction --- p.1Chapter 1.2 --- Fundamental Principle of Data Compression --- p.2Chapter 1.3 --- Some Data Compression Algorithms --- p.3Chapter 1.4 --- Image Coding Overview --- p.4Chapter 1.5 --- Image Transformation --- p.5Chapter 1.6 --- Quantization --- p.7Chapter 1.7 --- Lossless Coding --- p.8Chapter Chapter 2 --- Subband Coding and Wavelet Transform --- p.9Chapter 2.1 --- Subband Coding Principle --- p.9Chapter 2.2 --- Perfect Reconstruction --- p.11Chapter 2.3 --- Multi-Channel System --- p.13Chapter 2.4 --- Discrete Wavelet Transform --- p.13Chapter Chapter 3 --- Vector Quantization (VQ) --- p.16Chapter 3.1 --- Introduction --- p.16Chapter 3.2 --- Basic Vector Quantization Procedure --- p.17Chapter 3.3 --- Codebook Searching and the LBG Algorithm --- p.18Chapter 3.3.1 --- Codebook --- p.18Chapter 3.3.2 --- LBG Algorithm --- p.19Chapter 3.4 --- Problem of VQ and Variations of VQ --- p.21Chapter 3.4.1 --- Classified VQ (CVQ) --- p.22Chapter 3.4.2 --- Finite State VQ (FSVQ) --- p.23Chapter 3.5 --- Vector Quantization on Wavelet Coefficients --- p.24Chapter Chapter 4 --- Vector Wavelet Transform-Linear Predictor Coding --- p.26Chapter 4.1 --- Image Coding Using Wavelet Transform with Vector Quantization --- p.26Chapter 4.1.1 --- Future Standard --- p.26Chapter 4.1.2 --- Drawback of DCT --- p.27Chapter 4.1.3 --- "Wavelet Coding and VQ, the Future Trend" --- p.28Chapter 4.2 --- Mismatch between Scalar Transformation and VQ --- p.29Chapter 4.3 --- Vector Wavelet Transform (VWT) --- p.30Chapter 4.4 --- Example of Vector Wavelet Transform --- p.34Chapter 4.5 --- Vector Wavelet Transform - Linear Predictive Coding (VWT-LPC) --- p.36Chapter 4.6 --- An Example of VWT-LPC --- p.38Chapter Chapter 5 --- Vector Quantizaton with Inter-band Bit Allocation (IBBA) --- p.40Chapter 5.1 --- Bit Allocation Problem --- p.40Chapter 5.2 --- Bit Allocation for Wavelet Subband Vector Quantizer --- p.42Chapter 5.2.1 --- Multiple Codebooks --- p.42Chapter 5.2.2 --- Inter-band Bit Allocation (IBBA) --- p.42Chapter Chapter 6 --- Parental Finite State Vector Quantizers (PFSVQ) --- p.45Chapter 6.1 --- Introduction --- p.45Chapter 6.2 --- Parent-Child Relationship Between Subbands --- p.46Chapter 6.3 --- Wavelet Subband Vector Structures for VQ --- p.48Chapter 6.3.1 --- VQ on Separate Bands --- p.48Chapter 6.3.2 --- InterBand Information for Intraband Vectors --- p.49Chapter 6.3.3 --- Cross band Vector Methods --- p.50Chapter 6.4 --- Parental Finite State Vector Quantization Algorithms --- p.52Chapter 6.4.1 --- Scheme I: Parental Finite State VQ with Parent Index Equals Child Class Number --- p.52Chapter 6.4.2 --- Scheme II: Parental Finite State VQ with Parent Index Larger than Child Class Number --- p.55Chapter Chapter 7 --- Simulation Result --- p.58Chapter 7.1 --- Introduction --- p.58Chapter 7.2 --- Simulation Result of Vector Wavelet Transform (VWT) --- p.59Chapter 7.3 --- Simulation Result of Vector Wavelet Transform - Linear Predictive Coding (VWT-LPC) --- p.61Chapter 7.3.1 --- First Test --- p.61Chapter 7.3.2 --- Second Test --- p.61Chapter 7.3.3 --- Third Test --- p.61Chapter 7.4 --- Simulation Result of Vector Quantization Using Inter-band Bit Allocation (IBBA) --- p.62Chapter 7.5 --- Simulation Result of Parental Finite State Vector Quantizers (PFSVQ) --- p.63Chapter Chapter 8 --- Conclusion --- p.86REFERENCE --- p.8
Sparse representation based hyperspectral image compression and classification
Abstract
This thesis presents a research work on applying sparse representation to lossy hyperspectral image
compression and hyperspectral image classification. The proposed lossy hyperspectral image
compression framework introduces two types of dictionaries distinguished by the terms sparse
representation spectral dictionary (SRSD) and multi-scale spectral dictionary (MSSD), respectively.
The former is learnt in the spectral domain to exploit the spectral correlations, and the
latter in wavelet multi-scale spectral domain to exploit both spatial and spectral correlations in
hyperspectral images. To alleviate the computational demand of dictionary learning, either a
base dictionary trained offline or an update of the base dictionary is employed in the compression
framework. The proposed compression method is evaluated in terms of different objective
metrics, and compared to selected state-of-the-art hyperspectral image compression schemes, including
JPEG 2000. The numerical results demonstrate the effectiveness and competitiveness of
both SRSD and MSSD approaches.
For the proposed hyperspectral image classification method, we utilize the sparse coefficients
for training support vector machine (SVM) and k-nearest neighbour (kNN) classifiers. In particular,
the discriminative character of the sparse coefficients is enhanced by incorporating contextual
information using local mean filters. The classification performance is evaluated and compared
to a number of similar or representative methods. The results show that our approach could outperform
other approaches based on SVM or sparse representation.
This thesis makes the following contributions. It provides a relatively thorough investigation
of applying sparse representation to lossy hyperspectral image compression. Specifically,
it reveals the effectiveness of sparse representation for the exploitation of spectral correlations
in hyperspectral images. In addition, we have shown that the discriminative character of sparse
coefficients can lead to superior performance in hyperspectral image classification.EM201
The 1993 Space and Earth Science Data Compression Workshop
The Earth Observing System Data and Information System (EOSDIS) is described in terms of its data volume, data rate, and data distribution requirements. Opportunities for data compression in EOSDIS are discussed
Scalable video compression with optimized visual performance and random accessibility
This thesis is concerned with maximizing the coding efficiency, random accessibility and visual performance of scalable compressed video. The unifying theme behind this work is the use of finely embedded localized coding structures, which govern the extent to which these goals may be jointly achieved.
The first part focuses on scalable volumetric image compression. We investigate 3D transform and coding techniques which exploit inter-slice statistical redundancies without compromising slice accessibility. Our study shows that the motion-compensated temporal discrete wavelet transform (MC-TDWT) practically achieves an upper bound to the compression efficiency of slice transforms. From a video coding perspective, we find that most of the coding gain is attributed to offsetting the learning penalty in adaptive arithmetic coding through 3D code-block extension, rather than inter-frame context modelling.
The second aspect of this thesis examines random accessibility. Accessibility refers to the ease with which a region of interest is accessed (subband samples needed for reconstruction are retrieved) from a compressed video bitstream, subject to spatiotemporal code-block constraints. We investigate the fundamental implications of motion compensation for random access efficiency and the compression performance of scalable interactive video. We demonstrate that inclusion of motion compensation operators within the lifting steps of a temporal subband transform incurs a random access penalty which depends on the characteristics of the motion field.
The final aspect of this thesis aims to minimize the perceptual impact of visible distortion in scalable reconstructed video. We present a visual optimization strategy based on distortion scaling which raises the distortion-length slope of perceptually significant samples. This alters the codestream embedding order during post-compression rate-distortion optimization, thus allowing visually sensitive sites to be encoded with higher fidelity at a given bit-rate.
For visual sensitivity analysis, we propose a contrast perception model that incorporates an adaptive masking slope. This versatile feature provides a context which models perceptual significance. It enables scene structures that otherwise suffer significant degradation to be preserved at lower bit-rates. The novelty in our approach derives from a set of "perceptual mappings" which account for quantization noise shaping effects induced by motion-compensated temporal synthesis. The proposed technique reduces wavelet compression artefacts and improves the perceptual quality of video
Novel methods in image halftoning
Ankara : Department of Electrical and Electronics Engineering and Institute of Engineering and Science, Bilkent Univ., 1998.Thesis (Master's) -- Bilkent University, 1998.Includes bibliographical references leaves 97-101Halftoning refers to the problem of rendering continuous-tone (contone) images on display and printing devices which are capable of reproducing only a limited number of colors. A new adaptive halftoning method using the adaptive QR- RLS algorithm is developed for error diffusion which is one of the halftoning techniques. Also, a diagonal scanning strategy to exploit the human visual system properties in processing the image is proposed. Simulation results on color images demonstrate the superior quality of the new method compared to the existing methods. Another problem studied in this thesis is inverse halftoning which is the problem of recovering a contone image from a given halftoned image. A novel inverse halftoning method is developed for restoring a contone image from the halftoned image. A set theoretic formulation is used where sets are defined using the prior information about the problem. A new space domain projection is introduced assuming the halftoning is performed ,with error diffusion, and the error diffusion filter kernel is known. The space domain, frequency domain, and space-scale domain projections are used alternately to obtain a feasible solution for the inverse halftoning problem which does not have a unique solution. Simulation results for both grayscale and color images give good results, and demonstrate the effectiveness of the proposed inverse halftoning method.Bozkurt, GözdeM.S
Distortion-constraint compression of three-dimensional CLSM images using image pyramid and vector quantization
The confocal microscopy imaging techniques, which allow optical sectioning, have
been successfully exploited in biomedical studies. Biomedical scientists can benefit
from more realistic visualization and much more accurate diagnosis by processing and
analysing on a three-dimensional image data. The lack of efficient image compression
standards makes such large volumetric image data slow to transfer over limited
bandwidth networks. It also imposes large storage space requirements and high cost in
archiving and maintenance.
Conventional two-dimensional image coders do not take into account inter-frame
correlations in three-dimensional image data. The standard multi-frame coders, like
video coders, although they have good performance in capturing motion information,
are not efficiently designed for coding multiple frames representing a stack of optical
planes of a real object. Therefore a real three-dimensional image compression
approach should be investigated.
Moreover the reconstructed image quality is a very important concern in compressing
medical images, because it could be directly related to the diagnosis accuracy. Most of
the state-of-the-arts methods are based on transform coding, for instance JPEG is based on discrete-cosine-transform CDCT) and JPEG2000 is based on discrete-
wavelet-transform (DWT). However in DCT and DWT methods, the control
of the reconstructed image quality is inconvenient, involving considerable costs in
computation, since they are fundamentally rate-parameterized methods rather than
distortion-parameterized methods. Therefore it is very desirable to develop a
transform-based distortion-parameterized compression method, which is expected to
have high coding performance and also able to conveniently and accurately control
the final distortion according to the user specified quality requirement.
This thesis describes our work in developing a distortion-constraint three-dimensional
image compression approach, using vector quantization techniques combined with
image pyramid structures. We are expecting our method to have:
1. High coding performance in compressing three-dimensional microscopic
image data, compared to the state-of-the-art three-dimensional image coders
and other standardized two-dimensional image coders and video coders.
2. Distortion-control capability, which is a very desirable feature in medical 2. Distortion-control capability, which is a very desirable feature in medical
image compression applications, is superior to the rate-parameterized methods
in achieving a user specified quality requirement.
The result is a three-dimensional image compression method, which has outstanding
compression performance, measured objectively, for volumetric microscopic images.
The distortion-constraint feature, by which users can expect to achieve a target image
quality rather than the compressed file size, offers more flexible control of the
reconstructed image quality than its rate-constraint counterparts in medical image
applications. Additionally, it effectively reduces the artifacts presented in other
approaches at low bit rates and also attenuates noise in the pre-compressed images.
Furthermore, its advantages in progressive transmission and fast decoding make it
suitable for bandwidth limited tele-communications and web-based image browsing
applications