9,861 research outputs found
Robust Reduced-Rank Adaptive Processing Based on Parallel Subgradient Projection and Krylov Subspace Techniques
In this paper, we propose a novel reduced-rank adaptive filtering algorithm
by blending the idea of the Krylov subspace methods with the set-theoretic
adaptive filtering framework. Unlike the existing Krylov-subspace-based
reduced-rank methods, the proposed algorithm tracks the optimal point in the
sense of minimizing the \sinq{true} mean square error (MSE) in the Krylov
subspace, even when the estimated statistics become erroneous (e.g., due to
sudden changes of environments). Therefore, compared with those existing
methods, the proposed algorithm is more suited to adaptive filtering
applications. The algorithm is analyzed based on a modified version of the
adaptive projected subgradient method (APSM). Numerical examples demonstrate
that the proposed algorithm enjoys better tracking performance than the
existing methods for the interference suppression problem in code-division
multiple-access (CDMA) systems as well as for simple system identification
problems.Comment: 10 figures. In IEEE Transactions on Signal Processing, 201
Robust Subspace Learning: Robust PCA, Robust Subspace Tracking, and Robust Subspace Recovery
PCA is one of the most widely used dimension reduction techniques. A related
easier problem is "subspace learning" or "subspace estimation". Given
relatively clean data, both are easily solved via singular value decomposition
(SVD). The problem of subspace learning or PCA in the presence of outliers is
called robust subspace learning or robust PCA (RPCA). For long data sequences,
if one tries to use a single lower dimensional subspace to represent the data,
the required subspace dimension may end up being quite large. For such data, a
better model is to assume that it lies in a low-dimensional subspace that can
change over time, albeit gradually. The problem of tracking such data (and the
subspaces) while being robust to outliers is called robust subspace tracking
(RST). This article provides a magazine-style overview of the entire field of
robust subspace learning and tracking. In particular solutions for three
problems are discussed in detail: RPCA via sparse+low-rank matrix decomposition
(S+LR), RST via S+LR, and "robust subspace recovery (RSR)". RSR assumes that an
entire data vector is either an outlier or an inlier. The S+LR formulation
instead assumes that outliers occur on only a few data vector indices and hence
are well modeled as sparse corruptions.Comment: To appear, IEEE Signal Processing Magazine, July 201
Selective sampling importance resampling particle filter tracking with multibag subspace restoration
- …