5 research outputs found

    Subspace Detection of High-Dimensional Vectors using Compressing Sampling

    No full text
    <p>We consider the problem of detecting whether a high dimensional vector ∈ ℝ<sup>n</sup> lies in a r-dimensional subspace S, where r ≪ n, given few compressive measurements of the vector. This problem arises in several applications such as detecting anomalies, targets, interference and brain activations. In these applications, the object of interest is described by a large number of features and the ability to detect them using only linear combination of the features (without the need to measure, store or compute the entire feature vector) is desirable. We present a test statistic for subspace detection using compressive samples and demonstrate that the probability of error of the proposed detector decreases exponentially in the number of compressive samples, provided that the energy off the subspace scales as n. Using information-theoretic lower bounds, we demonstrate that no other detector can achieve the same probability of error for weaker signals. Simulation results also indicate that this scaling is near-optimal.</p

    Feature Selection For High-Dimensional Clustering

    No full text
    <p>We present a nonparametric method for selecting informative features in high-dimensional clustering problems. We start with a screening step that uses a test for multimodality. Then we apply kernel density estimation and mode clustering to the selected features. The output of the method consists of a list of relevant features, and cluster assignments. We provide explicit bounds on the error rate of the resulting clustering. In addition, we provide the first error bounds on mode based clustering.</p

    Density-Sensitive Semisupervised Inference

    No full text
    Semisupervised methods are techniques for using labeled data (X1; Y1),...,(Xn; Yn) together with unlabeled data Xn+1,...,XN to make predictions. These methods invoke some assumption that links the marginal distribution PX of X to the regression function f(x). For example, it is common to assume that f is very smooth over high density regions of PX. Many of the methods are ad-hoc and have been shown to work in specific examples but are lacking a theoretical foundation. We provide a minimax framework for analyzing semisupervised methods. In particular, we study methods based on metrics that are sensitive to the distribution PX. Our model includes a parameter α that controls the strength of the semisupervised assumption. We then use the data to adapt to α</p

    Minimax Theory for High-dimensional Gaussian Mixtures with Sparse Mean Separation

    No full text
    <p>While several papers have investigated computationally and statistically efficient methods for learning Gaussian mixtures, precise minimax bounds for their statistical performance as well as fundamental limits in high-dimensional settings are not well-understood. In this paper, we provide precise information theoretic bounds on the clustering accuracy and sample complexity of learning a mixture of two isotropic Gaussians in high dimensions under small mean separation. If there is a sparse subset of relevant dimensions that determine the mean separation, then the sample complexity only depends on the number of relevant dimensions and mean separation, and can be achieved by a simple computationally efficient procedure. Our results provide the first step of a theoretical basis for recent methods that combine feature selection and clustering.</p

    Subspace Learning from Extremely Compressed Measurements

    No full text
    <p>We consider learning the principal subspace of a large set of vectors from an extremely small number of compressive measurements of each vector. Our theoretical results show that even a constant number of measurements per column suffices to approximate the principal subspace to arbitrary precision, provided that the number of vectors is large. This result is achieved by a simple algorithm that computes the eigenvectors of an estimate of the covariance matrix. The main insight is to exploit an averaging effect that arises from applying a different random projection to each vector. We provide a number of simulations confirming our theoretical results</p
    corecore