7 research outputs found

    Sketching via hashing: from heavy hitters to compressed sensing to sparse fourier transform

    Get PDF
    Sketching via hashing is a popular and useful method for processing large data sets. Its basic idea is as follows. Suppose that we have a large multi-set of elements S=[formula], and we would like to identify the elements that occur “frequently" in S. The algorithm starts by selecting a hash function h that maps the elements into an array c[1…m]. The array entries are initialized to 0. Then, for each element a ∈ S, the algorithm increments c[h(a)]. At the end of the process, each array entry c[j] contains the count of all data elements a ∈ S mapped to j

    A Fast Hadamard Transform for Signals with Sub-linear Sparsity in the Transform Domain

    Get PDF
    In this paper, we design a new iterative low-complexity algorithm for computing the Walsh-Hadamard transform (WHT) of an N dimensional signal with a K-sparse WHT. We suppose that N is a power of two and K = O(N^α), scales sub-linearly in N for some α ∈ (0,1). Assuming a random support model for the nonzero transform-domain components, our algorithm reconstructs the WHT of the signal with a sample complexity O(K log_2(N/K)) and a computational complexity O(K log_2(K) log_2(N/K)). Moreover, the algorithm succeeds with a high probability approaching 1 for large dimension N. Our approach is mainly based on the subsampling (aliasing) property of the WHT, where by a carefully designed subsampling of the time-domain signal, a suitable aliasing pattern is induced in the transform domain. We treat the resulting aliasing patterns as parity-check constraints and represent them by a bipartite graph. We analyze the properties of the resulting bipartite graphs and borrow ideas from codes defined over sparse bipartite graphs to formulate the recovery of the nonzero spectral values as a peeling decoding algorithm for a specific sparse-graph code transmitted over a binary erasure channel (BEC). This enables us to use tools from coding theory (belief-propagation analysis) to characterize the asymptotic performance of our algorithm in the very sparse (α ∈ (0,1/3]) and the less sparse (α ∈ (1/3,1)) regime. Comprehensive simulation results are provided to assess the empirical performance of the proposed algorithm

    Compressed Sensing of Memoryless Sources:A Deterministic Hadamard Construction

    Get PDF
    Compressed sensing is a new trend in signal processing for efficient sampling and signal acquisition. The idea is that most real-world signals have a sparse representation in an appropriate basis and this can be exploited to capture the sparse signal by taking only a few linear projections. The recovery is possible by running appropriate low-complexity algorithms that exploit the sparsity (prior information) to reconstruct the signal from the linear projections (posterior information). The main benefit is that the required number of measurements is much smaller than the dimension of the signal. This results in a huge gain in sensor cost (in measurement devices) or a dramatic saving in data acquisition time. However, some difficulties naturally arise in applying the compressed sensing to real-world applications such as robustness issues in taking the linear projections and computational complexity of the recovery algorithm. In this thesis, we design structured matrices for compressed sensing. In particular, we claim that some of the practical difficulties can be reasonably solved by imposing some structure on the measurement matrices. The thesis evolves around the Hadamard matrices which are {+1,−1}\{+1,-1\}-valued matrices with many applications in signal processing, coding, optics and mathematics. As the title of the thesis implies, there are two main ingredients to this thesis. First, we use a memoryless assumption for the source, i.e., we assume that the nonzero components of the sparse signal are independently generated by a given probability distribution and their position is completely random. This allows us to use tools from probability, information theory and coding theory to rigorously assess the achievable performance. Second, using the mathematical properties of the Hadamard matrices, we design measurement matrices by selecting specific rows of a Hadamard matrix according to a deterministic criterion. We call the resulting matrices ``partial Hadamard matrices''. We design partial Hadamard matrices for three signal models: memoryless discrete signals and sparse signals with linear or sub-linear sparsity. A signal has linear sparsity if the number kk of its nonzero components is proportional to nn, the dimension of signal, whereas it has a sub-linear sparsity if kk scales like O(nα)O(n^\alpha) for some α∈(0,1)\alpha \in (0,1). We develop tools to rigorously analyze the performance of the proposed constructions by borrowing ideas from information theory and coding theory. We also extend our construction to distributed (multi-terminal) signals. Distributed compressed sensing is a ubiquitous problem in distributed data acquisition systems such as ad-hoc sensor networks. From both a theoretical and an engineering point of view, it is important to know how many measurement per dimension are necessary from different terminals in order to have a reliable estimate of the distributed data. We theoretically analyze this problem for a very simple setup where the components of the distributed signal are generated by a joint probability distribution which captures the spatial correlation among different terminals. We give an information-theoretic characterization of the measurements-rate region that results in a negligible recovery distortion. We also propose a low-complexity distributed message passing algorithm to achieve the theoretical limits
    corecore