10,721 research outputs found

    Sparse Modeling for Image and Vision Processing

    Get PDF
    In recent years, a large amount of multi-disciplinary research has been conducted on sparse models and their applications. In statistics and machine learning, the sparsity principle is used to perform model selection---that is, automatically selecting a simple model among a large collection of them. In signal processing, sparse coding consists of representing data with linear combinations of a few dictionary elements. Subsequently, the corresponding tools have been widely adopted by several scientific communities such as neuroscience, bioinformatics, or computer vision. The goal of this monograph is to offer a self-contained view of sparse modeling for visual recognition and image processing. More specifically, we focus on applications where the dictionary is learned and adapted to data, yielding a compact representation that has been successful in various contexts.Comment: 205 pages, to appear in Foundations and Trends in Computer Graphics and Visio

    Chirplet approximation of band-limited, real signals made easy

    Full text link
    In this paper we present algorithms for approximating real band-limited signals by multiple Gaussian Chirps. These algorithms do not rely on matching pursuit ideas. They are hierarchial and, at each stage, the number of terms in a given approximation depends only on the number of positive-valued maxima and negative-valued minima of a signed amplitude function characterizing part of the signal. Like the algorithms used in \cite{gre2} and unlike previous methods, our chirplet approximations require neither a complete dictionary of chirps nor complicated multi-dimensional searches to obtain suitable choices of chirp parameters

    Optimization with Sparsity-Inducing Penalties

    Get PDF
    Sparse estimation methods are aimed at using or obtaining parsimonious representations of data or models. They were first dedicated to linear variable selection but numerous extensions have now emerged such as structured sparsity or kernel selection. It turns out that many of the related estimation problems can be cast as convex optimization problems by regularizing the empirical risk with appropriate non-smooth norms. The goal of this paper is to present from a general perspective optimization tools and techniques dedicated to such sparsity-inducing penalties. We cover proximal methods, block-coordinate descent, reweighted ℓ2\ell_2-penalized techniques, working-set and homotopy methods, as well as non-convex formulations and extensions, and provide an extensive set of experiments to compare various algorithms from a computational point of view

    A Hierarchical Bayesian Model for Frame Representation

    Get PDF
    In many signal processing problems, it may be fruitful to represent the signal under study in a frame. If a probabilistic approach is adopted, it becomes then necessary to estimate the hyper-parameters characterizing the probability distribution of the frame coefficients. This problem is difficult since in general the frame synthesis operator is not bijective. Consequently, the frame coefficients are not directly observable. This paper introduces a hierarchical Bayesian model for frame representation. The posterior distribution of the frame coefficients and model hyper-parameters is derived. Hybrid Markov Chain Monte Carlo algorithms are subsequently proposed to sample from this posterior distribution. The generated samples are then exploited to estimate the hyper-parameters and the frame coefficients of the target signal. Validation experiments show that the proposed algorithms provide an accurate estimation of the frame coefficients and hyper-parameters. Application to practical problems of image denoising show the impact of the resulting Bayesian estimation on the recovered signal quality

    Highly computationally efficient state filter based on the delta operator

    Get PDF
    The Kalman filter is not suitable for the state estimation of linear systems with multistate delays, and the extended state vector Kalman filtering algorithm results in heavy computational burden because of the large dimension of the state estimation covariance matrix. Thus, in this paper, we develop a novel state estimation algorithm for enhancing the computational efficiency based on the delta operator. The computation analysis and the simulation example show the performance of the proposed algorithm
    • 

    corecore