129 research outputs found

    An optimal design procedure for intraband vector quantized subband coding

    Get PDF
    Journal ArticleAbstTact- Subband coding with vector quantization is addressed in this paper. Forming the data vectors from both between and within the subbands is considered. The former of these two schemes is referred to as interband coding and the latter as intraband coding. Interband coder design is relatively straightforward since the design of the single codebook involved follows readily from a representative set of interband data vectors. Intraband coder design is more complicated since it entails the selection of a vector dimension and a bit-rate for each subband. The main contribution of this work is an optimal methodology for intraband subband vector quantizer design. The problem formulation includes constraints on the bit-rate and the encoding complexity and is solved with nonlinear programming methods. Subband vector quantization image coding in conjunction with a human visual system model is thoroughly investigated. Results of a large number of experiments indicate that the optimal intraband coder yields superior results from quantitative as well as subjective points of view than the interband coder for comparable bit-rates. This improvement becomes more pronounced as the computational complexity of the intraband encoder is allowed to increase

    Image compression techniques using vector quantization

    Get PDF

    Wavelet-based adaptive video coding for packet-switching networks

    Get PDF
    Thesis (M.S.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 1995.Includes bibliographical references (leaves [92]-97).by Ye Gu.M.S

    Optimization of Coding of AR Sources for Transmission Across Channels with Loss

    Get PDF

    Multi-image classification and compression using vector quantization

    Get PDF
    Vector Quantization (VQ) is an image processing technique based on statistical clustering, and designed originally for image compression. In this dissertation, several methods for multi-image classification and compression based on a VQ design are presented. It is demonstrated that VQ can perform joint multi-image classification and compression by associating a class identifier with each multi-spectral signature codevector. We extend the Weighted Bayes Risk VQ (WBRVQ) method, previously used for single-component images, that explicitly incorporates a Bayes risk component into the distortion measure used in the VQ quantizer design and thereby permits a flexible trade-off between classification and compression priorities. In the specific case of multi-spectral images, we investigate the application of the Multi-scale Retinex algorithm as a preprocessing stage, before classification and compression, that performs dynamic range compression, reduces the dependence on lighting conditions, and generally enhances apparent spatial resolution. The goals of this research are four-fold: (1) to study the interrelationship between statistical clustering, classification and compression in a multi-image VQ context; (2) to study mixed-pixel classification and combined classification and compression for simulated and actual, multispectral and hyperspectral multi-images; (3) to study the effects of multi-image enhancement on class spectral signatures; and (4) to study the preservation of scientific data integrity as a function of compression. In this research, a key issue is not just the subjective quality of the resulting images after classification and compression but also the effect of multi-image dimensionality on the complexity of the optimal coder design

    Learning to compress and search visual data in large-scale systems

    Full text link
    The problem of high-dimensional and large-scale representation of visual data is addressed from an unsupervised learning perspective. The emphasis is put on discrete representations, where the description length can be measured in bits and hence the model capacity can be controlled. The algorithmic infrastructure is developed based on the synthesis and analysis prior models whose rate-distortion properties, as well as capacity vs. sample complexity trade-offs are carefully optimized. These models are then extended to multi-layers, namely the RRQ and the ML-STC frameworks, where the latter is further evolved as a powerful deep neural network architecture with fast and sample-efficient training and discrete representations. For the developed algorithms, three important applications are developed. First, the problem of large-scale similarity search in retrieval systems is addressed, where a double-stage solution is proposed leading to faster query times and shorter database storage. Second, the problem of learned image compression is targeted, where the proposed models can capture more redundancies from the training images than the conventional compression codecs. Finally, the proposed algorithms are used to solve ill-posed inverse problems. In particular, the problems of image denoising and compressive sensing are addressed with promising results.Comment: PhD thesis dissertatio

    Signal processing for high-definition television

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Mathematics, 1995.Includes bibliographical references (p. 60-62).by Peter Monta.Ph.D
    corecore