523 research outputs found

    Quantization Error Minimization by Reducing Median Difference at Quantization Interval Class

    Get PDF
    In this paper, a new technique to define the size of quantization interval is defined. In general, high quantization error will occur if large interval is used at a large difference value class whereas low quantization error will occur if a small interval is used at a large difference value class. However, the existence of too many class intervals will lead to a higher system complexity. Thus, this research is mainly about designing a quantization algorithm that can provide an efficient interval as possible to reduce the quantization error. The novelty of the proposed algorithm is to utilize the high occurrence of zero coefficient by re-allocating the non-zero coefficient in a group for quantization. From the experimental results provided, this new algorithm is able to produce a high compressed image without compromising with the image quality

    Design and Optimization of Graph Transform for Image and Video Compression

    Get PDF
    The main contribution of this thesis is the introduction of new methods for designing adaptive transforms for image and video compression. Exploiting graph signal processing techniques, we develop new graph construction methods targeted for image and video compression applications. In this way, we obtain a graph that is, at the same time, a good representation of the image and easy to transmit to the decoder. To do so, we investigate different research directions. First, we propose a new method for graph construction that employs innovative edge metrics, quantization and edge prediction techniques. Then, we propose to use a graph learning approach and we introduce a new graph learning algorithm targeted for image compression that defines the connectivities between pixels by taking into consideration the coding of the image signal and the graph topology in rate-distortion term. Moreover, we also present a new superpixel-driven graph transform that uses clusters of superpixel as coding blocks and then computes the graph transform inside each region. In the second part of this work, we exploit graphs to design directional transforms. In fact, an efficient representation of the image directional information is extremely important in order to obtain high performance image and video coding. In this thesis, we present a new directional transform, called Steerable Discrete Cosine Transform (SDCT). This new transform can be obtained by steering the 2D-DCT basis in any chosen direction. Moreover, we can also use more complex steering patterns than a single pure rotation. In order to show the advantages of the SDCT, we present a few image and video compression methods based on this new directional transform. The obtained results show that the SDCT can be efficiently applied to image and video compression and it outperforms the classical DCT and other directional transforms. Along the same lines, we present also a new generalization of the DFT, called Steerable DFT (SDFT). Differently from the SDCT, the SDFT can be defined in one or two dimensions. The 1D-SDFT represents a rotation in the complex plane, instead the 2D-SDFT performs a rotation in the 2D Euclidean space

    Combined Industry, Space and Earth Science Data Compression Workshop

    Get PDF
    The sixth annual Space and Earth Science Data Compression Workshop and the third annual Data Compression Industry Workshop were held as a single combined workshop. The workshop was held April 4, 1996 in Snowbird, Utah in conjunction with the 1996 IEEE Data Compression Conference, which was held at the same location March 31 - April 3, 1996. The Space and Earth Science Data Compression sessions seek to explore opportunities for data compression to enhance the collection, analysis, and retrieval of space and earth science data. Of particular interest is data compression research that is integrated into, or has the potential to be integrated into, a particular space or earth science data information system. Preference is given to data compression research that takes into account the scien- tist's data requirements, and the constraints imposed by the data collection, transmission, distribution and archival systems

    Consolidating Literature for Images Compression and Its Techniques

    Get PDF
    With the proliferation of readily available image content, image compression has become a topic of considerable importance. As, rapidly increase of digital imaging demand, storage capability aspect should be considered. Therefore, image compression refers to reducing the size of image for minimizing storage without harming the image quality. Thus, an appropriate technique is needed for image compression for saving capacity as well as not losing valuable information. This paper consolidates literature whose characteristics have focused on image compression, thresholding algorithms, quantization algorithms. Later, related research on these areas are presented

    DESIGN AND IMPLEMENTATION OF NON-UNIFORM QUANTIZERS FOR DISCRETE INPUT SAMPLES

    Get PDF
    This paper describes an algorithm for grayscale image compression based on non-uniform quantizers designed for discrete input samples. Non-uniform quantization is performed in two steps for unit variance, whereas design is done by introducing a discrete variance. The best theoretical and experimental results are obtained for those discrete values of variance which provide the operating range of quantizer located in the vicinity of maximal signal value that can appear on the entrance. The experiment is performed by applying proposed quantizers for compression of standard test grayscale images as a classic example of discrete input source. The proposed fixed non-uniform quantizers, designed for discrete input samples, provide up to 4.93 [dB] higher PSQNR compared to the fixed piecewise uniform quantizers designed for discrete input samples

    Image Compression Techniques: A Survey in Lossless and Lossy algorithms

    Get PDF
    The bandwidth of the communication networks has been increased continuously as results of technological advances. However, the introduction of new services and the expansion of the existing ones have resulted in even higher demand for the bandwidth. This explains the many efforts currently being invested in the area of data compression. The primary goal of these works is to develop techniques of coding information sources such as speech, image and video to reduce the number of bits required to represent a source without significantly degrading its quality. With the large increase in the generation of digital image data, there has been a correspondingly large increase in research activity in the field of image compression. The goal is to represent an image in the fewest number of bits without losing the essential information content within. Images carry three main type of information: redundant, irrelevant, and useful. Redundant information is the deterministic part of the information, which can be reproduced without loss from other information contained in the image. Irrelevant information is the part of information that has enormous details, which are beyond the limit of perceptual significance (i.e., psychovisual redundancy). Useful information, on the other hand, is the part of information, which is neither redundant nor irrelevant. Human usually observes decompressed images. Therefore, their fidelities are subject to the capabilities and limitations of the Human Visual System. This paper provides a survey on various image compression techniques, their limitations, compression rates and highlights current research in medical image compression
    corecore