5,696 research outputs found
The combinatorics of Borel covers
In this paper we extend previous studies of selection principles for families
of open covers of sets of real numbers to also include families of countable
Borel covers.
The main results of the paper could be summarized as follows:
1. Some of the classes which were different for open covers are equal for
Borel covers -- Section 1;
2. Some Borel classes coincide with classes that have been studied under a
different guise by other authors -- Section 4.Comment: Regular updat
An inverse of the evaluation functional for typed Lambda-calculus
In any model of typed λ-calculus conianing some basic
arithmetic, a functional p - * (procedure—* expression)
will be defined which inverts the evaluation functional
for typed X-terms, Combined with the evaluation
functional, p-e yields an efficient normalization algorithm.
The method is extended to X-calculi with constants
and is used to normalize (the X-representations
of) natural deduction proofs of (higher order) arithmetic.
A consequence of theoretical interest is a strong
completeness theorem for βη-reduction, generalizing
results of Friedman [1] and Statman [31: If two Xterms
have the same value in some model containing
representations of the primitive recursive functions
(of level 1) then they are provably equal in the βη-
calculus
A predicative variant of a realizability tripos for the Minimalist Foundation.
open2noHere we present a predicative variant of a realizability tripos validating
the intensional level of the Minimalist Foundation extended with Formal Church
thesis.the file attached contains the whole number of the journal including the mentioned pubblicationopenMaietti, Maria Emilia; Maschio, SamueleMaietti, MARIA EMILIA; Maschio, Samuel
Robust Subspace Learning: Robust PCA, Robust Subspace Tracking, and Robust Subspace Recovery
PCA is one of the most widely used dimension reduction techniques. A related
easier problem is "subspace learning" or "subspace estimation". Given
relatively clean data, both are easily solved via singular value decomposition
(SVD). The problem of subspace learning or PCA in the presence of outliers is
called robust subspace learning or robust PCA (RPCA). For long data sequences,
if one tries to use a single lower dimensional subspace to represent the data,
the required subspace dimension may end up being quite large. For such data, a
better model is to assume that it lies in a low-dimensional subspace that can
change over time, albeit gradually. The problem of tracking such data (and the
subspaces) while being robust to outliers is called robust subspace tracking
(RST). This article provides a magazine-style overview of the entire field of
robust subspace learning and tracking. In particular solutions for three
problems are discussed in detail: RPCA via sparse+low-rank matrix decomposition
(S+LR), RST via S+LR, and "robust subspace recovery (RSR)". RSR assumes that an
entire data vector is either an outlier or an inlier. The S+LR formulation
instead assumes that outliers occur on only a few data vector indices and hence
are well modeled as sparse corruptions.Comment: To appear, IEEE Signal Processing Magazine, July 201
- …