2,912 research outputs found
IMAX: incremental maintenance of schema-based XML statistics
Journal ArticleCurrent approaches for estimating the cardinality of XML queries are applicable to a static scenario wherein the underlying XML data does not change subsequent to the collection of statistics on the repository. However, in practice, many XML-based applications are dynamic and involve frequent updates to the data. In this paper, we investigate efficient strategies for incrementally maintaining statistical summaries as and when updates are applied to the data. Specifically, we propose algorithms that handle both the addition of new documents as well as random insertions in the existing document trees. We also show, through a detailed performance evaluation, that our incremental techniques are significantly faster than the naive recomputation approach; and that estimation accuracy can be maintained even with a fixed memory budget
06472 Abstracts Collection - XQuery Implementation Paradigms
From 19.11.2006 to 22.11.2006, the Dagstuhl Seminar 06472 ``XQuery Implementation Paradigms'' was held in the International Conference and Research Center (IBFI), Schloss Dagstuhl. During the seminar, several participants presented their current research, and ongoing work and open problems were discussed. Abstracts of the presentations given during the seminar as well as abstracts of seminar results and ideas are put together in this paper. The first section describes the seminar topics and goals in general. Links to extended abstracts or full papers are provided, if available
Distributed estimation from relative measurements of heterogeneous and uncertain quality
This paper studies the problem of estimation from relative measurements in a
graph, in which a vector indexed over the nodes has to be reconstructed from
pairwise measurements of differences between its components associated to nodes
connected by an edge. In order to model heterogeneity and uncertainty of the
measurements, we assume them to be affected by additive noise distributed
according to a Gaussian mixture. In this original setup, we formulate the
problem of computing the Maximum-Likelihood (ML) estimates and we design two
novel algorithms, based on Least Squares regression and
Expectation-Maximization (EM). The first algorithm (LS- EM) is centralized and
performs the estimation from relative measurements, the soft classification of
the measurements, and the estimation of the noise parameters. The second
algorithm (Distributed LS-EM) is distributed and performs estimation and soft
classification of the measurements, but requires the knowledge of the noise
parameters. We provide rigorous proofs of convergence of both algorithms and we
present numerical experiments to evaluate and compare their performance with
classical solutions. The experiments show the robustness of the proposed
methods against different kinds of noise and, for the Distributed LS-EM,
against errors in the knowledge of noise parameters.Comment: Submitted to IEEE transaction
Local Behavior of Sparse Analysis Regularization: Applications to Risk Estimation
In this paper, we aim at recovering an unknown signal x0 from noisy
L1measurements y=Phi*x0+w, where Phi is an ill-conditioned or singular linear
operator and w accounts for some noise. To regularize such an ill-posed inverse
problem, we impose an analysis sparsity prior. More precisely, the recovery is
cast as a convex optimization program where the objective is the sum of a
quadratic data fidelity term and a regularization term formed of the L1-norm of
the correlations between the sought after signal and atoms in a given
(generally overcomplete) dictionary. The L1-sparsity analysis prior is weighted
by a regularization parameter lambda>0. In this paper, we prove that any
minimizers of this problem is a piecewise-affine function of the observations y
and the regularization parameter lambda. As a byproduct, we exploit these
properties to get an objectively guided choice of lambda. In particular, we
develop an extension of the Generalized Stein Unbiased Risk Estimator (GSURE)
and show that it is an unbiased and reliable estimator of an appropriately
defined risk. The latter encompasses special cases such as the prediction risk,
the projection risk and the estimation risk. We apply these risk estimators to
the special case of L1-sparsity analysis regularization. We also discuss
implementation issues and propose fast algorithms to solve the L1 analysis
minimization problem and to compute the associated GSURE. We finally illustrate
the applicability of our framework to parameter(s) selection on several imaging
problems
- …