167,636 research outputs found
Generalized residual vector quantization for large scale data
Vector quantization is an essential tool for tasks involving large scale
data, for example, large scale similarity search, which is crucial for
content-based information retrieval and analysis. In this paper, we propose a
novel vector quantization framework that iteratively minimizes quantization
error. First, we provide a detailed review on a relevant vector quantization
method named \textit{residual vector quantization} (RVQ). Next, we propose
\textit{generalized residual vector quantization} (GRVQ) to further improve
over RVQ. Many vector quantization methods can be viewed as the special cases
of our proposed framework. We evaluate GRVQ on several large scale benchmark
datasets for large scale search, classification and object retrieval. We
compared GRVQ with existing methods in detail. Extensive experiments
demonstrate our GRVQ framework substantially outperforms existing methods in
term of quantization accuracy and computation efficiency.Comment: published on International Conference on Multimedia and Expo 201
Greedy vector quantization
We investigate the greedy version of the -optimal vector quantization
problem for an -valued random vector . We show the
existence of a sequence such that minimizes
(-mean quantization error at level induced by
). We show that this sequence produces -rate
optimal -tuples ( the -mean
quantization error at level induced by goes to at rate
). Greedy optimal sequences also satisfy, under natural
additional assumptions, the distortion mismatch property: the -tuples
remain rate optimal with respect to the -norms, .
Finally, we propose optimization methods to compute greedy sequences, adapted
from usual Lloyd's I and Competitive Learning Vector Quantization procedures,
either in their deterministic (implementable when ) or stochastic
versions.Comment: 31 pages, 4 figures, few typos corrected (now an extended version of
an eponym paper to appear in Journal of Approximation
Vector quantization
During the past ten years Vector Quantization (VQ) has developed from a theoretical possibility promised by Shannon's source coding theorems into a powerful and competitive technique for speech and image coding and compression at medium to low bit rates. In this survey, the basic ideas behind the design of vector quantizers are sketched and some comments made on the state-of-the-art and current research efforts
Semilogarithmic Nonuniform Vector Quantization of Two-Dimensional Laplacean Source for Small Variance Dynamics
In this paper high dynamic range nonuniform two-dimensional vector quantization model for Laplacean source was provided. Semilogarithmic A-law compression characteristic was used as radial scalar compression characteristic of two-dimensional vector quantization. Optimal number value of concentric quantization domains (amplitude levels) is expressed in the function of parameter A. Exact distortion analysis with obtained closed form expressions is provided. It has been shown that proposed model provides high SQNR values in wide range of variances, and overachieves quality obtained by scalar A-law quantization at same bit rate, so it can be used in various switching and adaptation implementations for realization of high quality signal compression
- …
