1,452 research outputs found
A Generative Model of Natural Texture Surrogates
Natural images can be viewed as patchworks of different textures, where the
local image statistics is roughly stationary within a small neighborhood but
otherwise varies from region to region. In order to model this variability, we
first applied the parametric texture algorithm of Portilla and Simoncelli to
image patches of 64X64 pixels in a large database of natural images such that
each image patch is then described by 655 texture parameters which specify
certain statistics, such as variances and covariances of wavelet coefficients
or coefficient magnitudes within that patch.
To model the statistics of these texture parameters, we then developed
suitable nonlinear transformations of the parameters that allowed us to fit
their joint statistics with a multivariate Gaussian distribution. We find that
the first 200 principal components contain more than 99% of the variance and
are sufficient to generate textures that are perceptually extremely close to
those generated with all 655 components. We demonstrate the usefulness of the
model in several ways: (1) We sample ensembles of texture patches that can be
directly compared to samples of patches from the natural image database and can
to a high degree reproduce their perceptual appearance. (2) We further
developed an image compression algorithm which generates surprisingly accurate
images at bit rates as low as 0.14 bits/pixel. Finally, (3) We demonstrate how
our approach can be used for an efficient and objective evaluation of samples
generated with probabilistic models of natural images.Comment: 34 pages, 9 figure
Combining nonlinear multiresolution system and vector quantization for still image compression
It is popular to use multiresolution systems for image coding and compression. However, general-purpose techniques such as filter banks and wavelets are linear. While these systems are rigorous, nonlinear features in the signals cannot be utilized in a single entity for compression. Linear filters are known to blur the edges. Thus, the low-resolution images are typically blurred, carrying little information. We propose and demonstrate that edge- preserving filters such as median filters can be used in generating a multiresolution system using the Laplacian pyramid. The signals in the detail images are small and localized in the edge areas. Principal component vector quantization (PCVQ) is used to encode the detail images. PCVQ is a tree-structured VQ which allows fast codebook design and encoding/decoding. In encoding, the quantization error at each level is fed back through the pyramid to the previous level so that ultimately all the error is confined to the first level. With simple coding methods, we demonstrate that images with PSNR 33 dB can be obtained at 0.66 bpp without the use of entropy coding. When the rate is decreased to 0.25 bpp, the PSNR of 30 dB can still be achieved. Combined with an earlier result, our work demonstrate that nonlinear filters can be used for multiresolution systems and image coding
Flat zones filtering, connected operators, and filters by reconstruction
This correspondence deals with the notion of connected operators. Starting from the definition for operator acting on sets, it is shown how to extend it to operators acting on function. Typically, a connected operator acting on a function is a transformation that enlarges the partition of the space created by the flat zones of the functions. It is shown that from any connected operator acting on sets, one can construct a connected operator for functions (however, it is not the unique way of generating connected operators for functions). Moreover, the concept of pyramid is introduced in a formal way. It is shown that, if a pyramid is based on connected operators, the flat zones of the functions increase with the level of the pyramid. In other words, the flat zones are nested. Filters by reconstruction are defined and their main properties are presented. Finally, some examples of application of connected operators and use of flat zones are described.Peer ReviewedPostprint (published version
A statistical reduced-reference method for color image quality assessment
Although color is a fundamental feature of human visual perception, it has
been largely unexplored in the reduced-reference (RR) image quality assessment
(IQA) schemes. In this paper, we propose a natural scene statistic (NSS)
method, which efficiently uses this information. It is based on the statistical
deviation between the steerable pyramid coefficients of the reference color
image and the degraded one. We propose and analyze the multivariate generalized
Gaussian distribution (MGGD) to model the underlying statistics. In order to
quantify the degradation, we develop and evaluate two measures based
respectively on the Geodesic distance between two MGGDs and on the closed-form
of the Kullback Leibler divergence. We performed an extensive evaluation of
both metrics in various color spaces (RGB, HSV, CIELAB and YCrCb) using the TID
2008 benchmark and the FRTV Phase I validation process. Experimental results
demonstrate the effectiveness of the proposed framework to achieve a good
consistency with human visual perception. Furthermore, the best configuration
is obtained with CIELAB color space associated to KLD deviation measure
On color image quality assessment using natural image statistics
Color distortion can introduce a significant damage in visual quality
perception, however, most of existing reduced-reference quality measures are
designed for grayscale images. In this paper, we consider a basic extension of
well-known image-statistics based quality assessment measures to color images.
In order to evaluate the impact of color information on the measures
efficiency, two color spaces are investigated: RGB and CIELAB. Results of an
extensive evaluation using TID 2013 benchmark demonstrates that significant
improvement can be achieved for a great number of distortion type when the
CIELAB color representation is used
Perceptual Copyright Protection Using Multiresolution Wavelet-Based Watermarking And Fuzzy Logic
In this paper, an efficiently DWT-based watermarking technique is proposed to
embed signatures in images to attest the owner identification and discourage
the unauthorized copying. This paper deals with a fuzzy inference filter to
choose the larger entropy of coefficients to embed watermarks. Unlike most
previous watermarking frameworks which embedded watermarks in the larger
coefficients of inner coarser subbands, the proposed technique is based on
utilizing a context model and fuzzy inference filter by embedding watermarks in
the larger-entropy coefficients of coarser DWT subbands. The proposed
approaches allow us to embed adaptive casting degree of watermarks for
transparency and robustness to the general image-processing attacks such as
smoothing, sharpening, and JPEG compression. The approach has no need the
original host image to extract watermarks. Our schemes have been shown to
provide very good results in both image transparency and robustness.Comment: 13 pages, 7 figure
BLADE: Filter Learning for General Purpose Computational Photography
The Rapid and Accurate Image Super Resolution (RAISR) method of Romano,
Isidoro, and Milanfar is a computationally efficient image upscaling method
using a trained set of filters. We describe a generalization of RAISR, which we
name Best Linear Adaptive Enhancement (BLADE). This approach is a trainable
edge-adaptive filtering framework that is general, simple, computationally
efficient, and useful for a wide range of problems in computational
photography. We show applications to operations which may appear in a camera
pipeline including denoising, demosaicing, and stylization
- …