13,755 research outputs found

    Image coding using wavelet transform and adaptive block truncation coding

    Get PDF
    This thesis presents a new image coding using wavelet transform and adaptive block truncation coding. Images are first pre-processed by the wavelet transform and then coded by the adaptive block truncation coding. Algorithms for both monochrome and color images are proposed and experimentally studied. The adaptive block truncation coding is also modified to achieve better performance. For coding monochrome images at the bit-rate region between 0.8 to 1.2 bits/pixel, the performance of the new coding is comparable to the ones of subband codings and other image codings using the wavelet transform; however, the new coding offers less computational load. The new coding also gives a good reconstruction of a color image at the bit-rate of 1.0 bit/pixel. The comparison between the new coding and the original adaptive block truncation coding is also given. The discussion on effects of a filter and a number of decomposition levels used for an implementation of the wavelet transform is included in this thesis, as well

    Regularity scalable image coding based on wavelet singularity detection

    Get PDF
    In this paper, we propose an adaptive algorithm for scalable wavelet image coding, which is based on the general feature, the regularity, of images. In pattern recognition or computer vision, regularity of images is estimated from the oriented wavelet coefficients and quantified by the Lipschitz exponents. To estimate the Lipschitz exponents, evaluating the interscale evolution of the wavelet transform modulus sum (WTMS) over the directional cone of influence was proven to be a better approach than tracing the wavelet transform modulus maxima (WTMM). This is because the irregular sampling nature of the WTMM complicates the reconstruction process. Moreover, examples were found to show that the WTMM representation cannot uniquely characterize a signal. It implies that the reconstruction of signal from its WTMM may not be consistently stable. Furthermore, the WTMM approach requires much more computational effort. Therefore, we use the WTMS approach to estimate the regularity of images from the separable wavelet transformed coefficients. Since we do not concern about the localization issue, we allow the decimation to occur when we evaluate the interscale evolution. After the regularity is estimated, this information is utilized in our proposed adaptive regularity scalable wavelet image coding algorithm. This algorithm can be simply embedded into any wavelet image coders, so it is compatible with the existing scalable coding techniques, such as the resolution scalable and signal-to-noise ratio (SNR) scalable coding techniques, without changing the bitstream format, but provides more scalable levels with higher peak signal-to-noise ratios (PSNRs) and lower bit rates. In comparison to the other feature-based wavelet scalable coding algorithms, the proposed algorithm outperforms them in terms of visual perception, computational complexity and coding efficienc

    Wavelet Based Image Coding Schemes : A Recent Survey

    Full text link
    A variety of new and powerful algorithms have been developed for image compression over the years. Among them the wavelet-based image compression schemes have gained much popularity due to their overlapping nature which reduces the blocking artifacts that are common phenomena in JPEG compression and multiresolution character which leads to superior energy compaction with high quality reconstructed images. This paper provides a detailed survey on some of the popular wavelet coding techniques such as the Embedded Zerotree Wavelet (EZW) coding, Set Partitioning in Hierarchical Tree (SPIHT) coding, the Set Partitioned Embedded Block (SPECK) Coder, and the Embedded Block Coding with Optimized Truncation (EBCOT) algorithm. Other wavelet-based coding techniques like the Wavelet Difference Reduction (WDR) and the Adaptive Scanned Wavelet Difference Reduction (ASWDR) algorithms, the Space Frequency Quantization (SFQ) algorithm, the Embedded Predictive Wavelet Image Coder (EPWIC), Compression with Reversible Embedded Wavelet (CREW), the Stack-Run (SR) coding and the recent Geometric Wavelet (GW) coding are also discussed. Based on the review, recommendations and discussions are presented for algorithm development and implementation.Comment: 18 pages, 7 figures, journa

    Stack-run adaptive wavelet image compression

    Get PDF
    We report on the development of an adaptive wavelet image coder based on stack-run representation of the quantized coefficients. The coder works by selecting an optimal wavelet packet basis for the given image and encoding the quantization indices for significant coefficients and zero runs between coefficients using a 4-ary arithmetic coder. Due to the fact that our coder exploits the redundancies present within individual subbands, its addressing complexity is much lower than that of the wavelet zerotree coding algorithms. Experimental results show coding gains of up to 1:4dB over the benchmark wavelet coding algorithm

    In-Band Disparity Compensation for Multiview Image Compression and View Synthesis

    Get PDF

    Perceptual Copyright Protection Using Multiresolution Wavelet-Based Watermarking And Fuzzy Logic

    Full text link
    In this paper, an efficiently DWT-based watermarking technique is proposed to embed signatures in images to attest the owner identification and discourage the unauthorized copying. This paper deals with a fuzzy inference filter to choose the larger entropy of coefficients to embed watermarks. Unlike most previous watermarking frameworks which embedded watermarks in the larger coefficients of inner coarser subbands, the proposed technique is based on utilizing a context model and fuzzy inference filter by embedding watermarks in the larger-entropy coefficients of coarser DWT subbands. The proposed approaches allow us to embed adaptive casting degree of watermarks for transparency and robustness to the general image-processing attacks such as smoothing, sharpening, and JPEG compression. The approach has no need the original host image to extract watermarks. Our schemes have been shown to provide very good results in both image transparency and robustness.Comment: 13 pages, 7 figure
    corecore