7,159 research outputs found

    Simulated Annealing for JPEG Quantization

    Full text link
    JPEG is one of the most widely used image formats, but in some ways remains surprisingly unoptimized, perhaps because some natural optimizations would go outside the standard that defines JPEG. We show how to improve JPEG compression in a standard-compliant, backward-compatible manner, by finding improved default quantization tables. We describe a simulated annealing technique that has allowed us to find several quantization tables that perform better than the industry standard, in terms of both compressed size and image fidelity. Specifically, we derive tables that reduce the FSIM error by over 10% while improving compression by over 20% at quality level 95 in our tests; we also provide similar results for other quality levels. While we acknowledge our approach can in some images lead to visible artifacts under large magnification, we believe use of these quantization tables, or additional tables that could be found using our methodology, would significantly reduce JPEG file sizes with improved overall image quality.Comment: Appendix not included in arXiv version due to size restrictions. For full paper go to: http://www.eecs.harvard.edu/~michaelm/SimAnneal/PAPER/simulated-annealing-jpeg.pd

    Logarithmical hopping encoding: a low computational complexity algorithm for image compression

    Get PDF
    LHE (logarithmical hopping encoding) is a computationally efficient image compression algorithm that exploits the Weber–Fechner law to encode the error between colour component predictions and the actual value of such components. More concretely, for each pixel, luminance and chrominance predictions are calculated as a function of the surrounding pixels and then the error between the predictions and the actual values are logarithmically quantised. The main advantage of LHE is that although it is capable of achieving a low-bit rate encoding with high quality results in terms of peak signal-to-noise ratio (PSNR) and image quality metrics with full-reference (FSIM) and non-reference (blind/referenceless image spatial quality evaluator), its time complexity is O( n) and its memory complexity is O(1). Furthermore, an enhanced version of the algorithm is proposed, where the output codes provided by the logarithmical quantiser are used in a pre-processing stage to estimate the perceptual relevance of the image blocks. This allows the algorithm to downsample the blocks with low perceptual relevance, thus improving the compression rate. The performance of LHE is especially remarkable when the bit per pixel rate is low, showing much better quality, in terms of PSNR and FSIM, than JPEG and slightly lower quality than JPEG-2000 but being more computationally efficient

    Semantic Perceptual Image Compression using Deep Convolution Networks

    Full text link
    It has long been considered a significant problem to improve the visual quality of lossy image and video compression. Recent advances in computing power together with the availability of large training data sets has increased interest in the application of deep learning cnns to address image recognition and image processing tasks. Here, we present a powerful cnn tailored to the specific task of semantic image understanding to achieve higher visual quality in lossy compression. A modest increase in complexity is incorporated to the encoder which allows a standard, off-the-shelf jpeg decoder to be used. While jpeg encoding may be optimized for generic images, the process is ultimately unaware of the specific content of the image to be compressed. Our technique makes jpeg content-aware by designing and training a model to identify multiple semantic regions in a given image. Unlike object detection techniques, our model does not require labeling of object positions and is able to identify objects in a single pass. We present a new cnn architecture directed specifically to image compression, which generates a map that highlights semantically-salient regions so that they can be encoded at higher quality as compared to background regions. By adding a complete set of features for every class, and then taking a threshold over the sum of all feature activations, we generate a map that highlights semantically-salient regions so that they can be encoded at a better quality compared to background regions. Experiments are presented on the Kodak PhotoCD dataset and the MIT Saliency Benchmark dataset, in which our algorithm achieves higher visual quality for the same compressed size.Comment: Accepted to Data Compression Conference, 11 pages, 5 figure

    Compression artifact suppression for color images with dual-domain SE-ARResNet

    Get PDF
    JPEG compression has been a popular lossy image compression technique and is widely used in digital imaging. Restoring high-quality images from their compressed JPEG counterparts, however, is an ill-posed inverse problem but could be of great use in improving the visual quality of images. With the representational power that convolutional neural networks (CNNs) demon- strate, we show that it is possible to suppress JPEG compression artifacts and recover visually pleasing images. To recover original high-quality and high-resolution images from JPEG compressed images, we leverage prior knowledge of JPEG compression into consideration by exploiting frequency redundancies with the CNN in discrete cosine domain and constrain the quantization loss, in addition to exploiting spatial redundancies in the pixel domain. This data-driven approach tar- gets removing compression artifacts, including blocking, blurring, ringing and banding artifacts, and recovering high-frequency information for reconstruction. We design a deep CNN in each domain and fuse the outputs with an aggregation network to produce the output image. To improve the model performance, we leverage the robustness and ability to tackle vanishing gradient problems of ResNet to build a deep network, and utilize squeeze-and- excitation block, a technique typically found beneficial in classification tasks, to this regression problem to exploit global information in a larger scale. We refer to the module proposed in this work as squeeze-and-excitation artifact removal ResNet (SE-ARResNet). Prior work in this field mainly focuses on reconstructing a grayscale image or the luminance channel of the image. We demonstrate that we can reconstruct color images effectively and robustly with the dual-domain CNN approach
    corecore